You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hawq.apache.org by es...@apache.org on 2017/02/03 09:00:24 UTC
[22/50] [abbrv] incubator-hawq git commit: Revert "HAWQ-1248. Merge
Dockerfiles for HAWQ Dev into HAWQ code base."
Revert "HAWQ-1248. Merge Dockerfiles for HAWQ Dev into HAWQ code base."
This reverts commit 440ce595a2298ac9be16973f0c7c4c358ddb2cd0.
Reason for revert: RAT check fail with unapproved license in some of the files added
Project: http://git-wip-us.apache.org/repos/asf/incubator-hawq/repo
Commit: http://git-wip-us.apache.org/repos/asf/incubator-hawq/commit/1cb29096
Tree: http://git-wip-us.apache.org/repos/asf/incubator-hawq/tree/1cb29096
Diff: http://git-wip-us.apache.org/repos/asf/incubator-hawq/diff/1cb29096
Branch: refs/heads/2.1.0.0-incubating
Commit: 1cb29096ca89f4ed6f54ab795dfad518131969f2
Parents: 440ce59
Author: Ruilong Huo <rh...@pivotal.io>
Authored: Fri Jan 13 19:38:13 2017 +0800
Committer: Ruilong Huo <rh...@pivotal.io>
Committed: Sat Jan 14 11:40:11 2017 +0800
----------------------------------------------------------------------
contrib/hawq-docker/Makefile | 222 --------------
contrib/hawq-docker/README.md | 97 -------
.../centos6-docker/hawq-dev/Dockerfile | 123 --------
.../centos6-docker/hawq-test/Dockerfile | 40 ---
.../hawq-test/conf/capacity-scheduler.xml | 134 ---------
.../hawq-test/conf/configuration.xsl | 40 ---
.../hawq-test/conf/container-executor.cfg | 4 -
.../centos6-docker/hawq-test/conf/core-site.xml | 24 --
.../hawq-test/conf/hadoop-env.cmd | 92 ------
.../centos6-docker/hawq-test/conf/hadoop-env.sh | 110 -------
.../hawq-test/conf/hadoop-metrics.properties | 75 -----
.../hawq-test/conf/hadoop-metrics2.properties | 68 -----
.../hawq-test/conf/hadoop-policy.xml | 226 --------------
.../centos6-docker/hawq-test/conf/hdfs-site.xml | 100 -------
.../centos6-docker/hawq-test/conf/kms-acls.xml | 135 ---------
.../centos6-docker/hawq-test/conf/kms-env.sh | 55 ----
.../hawq-test/conf/kms-log4j.properties | 38 ---
.../centos6-docker/hawq-test/conf/kms-site.xml | 173 -----------
.../hawq-test/conf/log4j.properties | 291 -------------------
.../hawq-test/conf/mapred-env.cmd | 20 --
.../centos6-docker/hawq-test/conf/mapred-env.sh | 27 --
.../hawq-test/conf/mapred-queues.xml.template | 92 ------
.../hawq-test/conf/mapred-site.xml.template | 21 --
.../centos6-docker/hawq-test/conf/slaves | 1 -
.../hawq-test/conf/ssl-client.xml.example | 80 -----
.../hawq-test/conf/ssl-server.xml.example | 78 -----
.../centos6-docker/hawq-test/conf/yarn-env.cmd | 60 ----
.../centos6-docker/hawq-test/entrypoint.sh | 34 ---
.../centos6-docker/hawq-test/start-hdfs.sh | 39 ---
.../centos7-docker/hawq-dev/Dockerfile | 75 -----
.../centos7-docker/hawq-test/Dockerfile | 40 ---
.../hawq-test/conf/capacity-scheduler.xml | 134 ---------
.../hawq-test/conf/configuration.xsl | 40 ---
.../hawq-test/conf/container-executor.cfg | 4 -
.../centos7-docker/hawq-test/conf/core-site.xml | 24 --
.../hawq-test/conf/hadoop-env.cmd | 92 ------
.../centos7-docker/hawq-test/conf/hadoop-env.sh | 110 -------
.../hawq-test/conf/hadoop-metrics.properties | 75 -----
.../hawq-test/conf/hadoop-metrics2.properties | 68 -----
.../hawq-test/conf/hadoop-policy.xml | 226 --------------
.../centos7-docker/hawq-test/conf/hdfs-site.xml | 100 -------
.../centos7-docker/hawq-test/conf/kms-acls.xml | 135 ---------
.../centos7-docker/hawq-test/conf/kms-env.sh | 55 ----
.../hawq-test/conf/kms-log4j.properties | 38 ---
.../centos7-docker/hawq-test/conf/kms-site.xml | 173 -----------
.../hawq-test/conf/log4j.properties | 291 -------------------
.../hawq-test/conf/mapred-env.cmd | 20 --
.../centos7-docker/hawq-test/conf/mapred-env.sh | 27 --
.../hawq-test/conf/mapred-queues.xml.template | 92 ------
.../hawq-test/conf/mapred-site.xml.template | 21 --
.../centos7-docker/hawq-test/conf/slaves | 1 -
.../hawq-test/conf/ssl-client.xml.example | 80 -----
.../hawq-test/conf/ssl-server.xml.example | 78 -----
.../centos7-docker/hawq-test/conf/yarn-env.cmd | 60 ----
.../centos7-docker/hawq-test/entrypoint.sh | 33 ---
.../centos7-docker/hawq-test/start-hdfs.sh | 39 ---
56 files changed, 4630 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/1cb29096/contrib/hawq-docker/Makefile
----------------------------------------------------------------------
diff --git a/contrib/hawq-docker/Makefile b/contrib/hawq-docker/Makefile
deleted file mode 100644
index 120ebe2..0000000
--- a/contrib/hawq-docker/Makefile
+++ /dev/null
@@ -1,222 +0,0 @@
-#!/usr/bin/make all
-
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied. See the License for the
-# specific language governing permissions and limitations
-# under the License.
-
-THIS_MAKEFILE_PATH := $(abspath $(lastword $(MAKEFILE_LIST)))
-TOP_DIR := $(abspath $(dir ${THIS_MAKEFILE_PATH}))
-NDATANODES := 3
-CUR_DATANODE := 1
-OS_VERSION := centos7
-# Do not use underscore "_" in CLUSTER_ID
-CLUSTER_ID := $(OS_VERSION)
-# Monut this local directory to /data in data container and share with other containers
-LOCAL :=
-# networks used in docker
-NETWORK := $(CLUSTER_ID)_hawq_network
-
-all:
- @echo " Usage:"
- @echo " To setup a build and test environment: make run"
- @echo " To start all containers: make start"
- @echo " To stop all containers: make stop"
- @echo " To remove hdfs containers: make clean"
- @echo " To remove all containers: make distclean"
- @echo ""
- @echo " To build images locally: make build"
- @echo " To pull latest images: make pull"
-
-build:
- @make -f $(THIS_MAKEFILE_PATH) build-hawq-dev-$(OS_VERSION)
- @make -f $(THIS_MAKEFILE_PATH) build-hawq-test-$(OS_VERSION)
- @echo "Build Images Done!"
-
-build-hawq-dev-$(OS_VERSION): $(TOP_DIR)/$(OS_VERSION)-docker/hawq-dev/Dockerfile
- @echo build hawq-dev:$(OS_VERSION) image
- docker build -t hawq/hawq-dev:$(OS_VERSION) $(TOP_DIR)/$(OS_VERSION)-docker/hawq-dev/
-
-build-hawq-test-$(OS_VERSION): $(TOP_DIR)/$(OS_VERSION)-docker/hawq-test/Dockerfile
- @echo build hawq-test:$(OS_VERSION) image
- docker build -t hawq/hawq-test:$(OS_VERSION) $(TOP_DIR)/$(OS_VERSION)-docker/hawq-test/
-
-create-data-container:
- @echo create ${CLUSTER_ID}-data container
- @if [ ! -z "$(LOCAL)" -a ! -d "$(LOCAL)" ]; then \
- echo "LOCAL must be set to a directory!"; \
- exit 1; \
- fi
- @if [ -z "`docker ps -a --filter="name=${CLUSTER_ID}-data$$" | grep -v CONTAINER`" ]; then \
- if [ -z "$(LOCAL)" ]; then \
- docker create -v /data --name=${CLUSTER_ID}-data hawq/hawq-dev:$(OS_VERSION) /bin/true; \
- else \
- docker create -v $(LOCAL):/data --name=${CLUSTER_ID}-data hawq/hawq-dev:$(OS_VERSION) /bin/true; \
- fi \
- else \
- echo "${CLUSTER_ID}-data container already exist!"; \
- fi
-
-run:
- @if [ -z "`docker network ls 2>/dev/null`" ]; then \
- make -f $(THIS_MAKEFILE_PATH) NETWORK=default create-data-container && \
- make -f $(THIS_MAKEFILE_PATH) NETWORK=default run-hdfs; \
- else \
- if [ -z "`docker network ls 2>/dev/null | grep $(NETWORK)`" ]; then \
- echo create network $(NETWORK) && \
- docker network create --driver bridge $(NETWORK); \
- fi && \
- make -f $(THIS_MAKEFILE_PATH) create-data-container && \
- make -f $(THIS_MAKEFILE_PATH) run-hdfs; \
- fi
-
-run-hdfs:
- @make -f $(THIS_MAKEFILE_PATH) run-namenode-container
- @i=1; \
- while [ $$i -le $(NDATANODES) ] ; do \
- make -f $(THIS_MAKEFILE_PATH) CUR_DATANODE=$$i run-datanode-container; \
- i=$$((i+1)); \
- done
- @echo "HAWQ Environment Setup Done!"
- @echo 'run "docker exec -it ${CLUSTER_ID}-namenode bash" to attach to ${CLUSTER_ID}-namenode node'
-
-run-namenode-container:
- @echo "run ${CLUSTER_ID}-namenode container"
- @if [ -z "`docker ps -a --filter="name=${CLUSTER_ID}-namenode" | grep -v CONTAINER`" ]; then \
- docker run --privileged -itd --net=$(NETWORK) --hostname=${CLUSTER_ID}-namenode --name=${CLUSTER_ID}-namenode \
- --volumes-from ${CLUSTER_ID}-data hawq/hawq-test:$(OS_VERSION); \
- else \
- echo "${CLUSTER_ID}-namenode container already exist!"; \
- fi
-
-run-datanode-container:
- @echo "run ${CLUSTER_ID}-datanode$(CUR_DATANODE) container"
- @if [ -z "`docker ps -a --filter="name=${CLUSTER_ID}-datanode$(CUR_DATANODE)" | grep -v CONTAINER`" ]; then \
- docker run --privileged -itd --net=$(NETWORK) --hostname=${CLUSTER_ID}-datanode$(CUR_DATANODE) \
- --name=${CLUSTER_ID}-datanode$(CUR_DATANODE) -e NAMENODE=${CLUSTER_ID}-namenode \
- --volumes-from ${CLUSTER_ID}-data hawq/hawq-test:$(OS_VERSION); \
- else \
- echo "${CLUSTER_ID}-datanode$(CUR_DATANODE) container already exist!"; \
- fi
-
-start:
- @make -f $(THIS_MAKEFILE_PATH) start-hdfs
- @echo 'run "docker exec -it ${CLUSTER_ID}-namenode bash" to attach to ${CLUSTER_ID}-namenode node'
-
-start-hdfs:
- @make -f $(THIS_MAKEFILE_PATH) start-namenode-container
- @i=1;\
- while [ $$i -le $(NDATANODES) ] ; do \
- make -f $(THIS_MAKEFILE_PATH) CUR_DATANODE=$$i start-datanode-container; \
- i=$$((i+1)); \
- done
- @echo "Start All Containers Done!"
-
-start-namenode-container:
- @echo "start ${CLUSTER_ID}-namenode container"
- @if [ ! -z "`docker ps -a --filter="name=${CLUSTER_ID}-namenode" | grep -v CONTAINER`" ]; then \
- docker start ${CLUSTER_ID}-namenode; \
- else \
- echo "${CLUSTER_ID}-namenode container does not exist!" && exit 1; \
- fi
-
-start-datanode-container:
- @echo "start ${CLUSTER_ID}-datanode$(CUR_DATANODE) container"
- @if [ ! -z "`docker ps -a --filter="name=${CLUSTER_ID}-datanode$(CUR_DATANODE)" | grep -v CONTAINER`" ]; then \
- docker start ${CLUSTER_ID}-datanode$(CUR_DATANODE); \
- else \
- echo "${CLUSTER_ID}-datanode$(CUR_DATANODE) container does not exist!" && exit 1; \
- fi
-
-stop:
- @make -f $(THIS_MAKEFILE_PATH) stop-hdfs
-
-stop-hdfs:
- @make -f $(THIS_MAKEFILE_PATH) stop-namenode-container
- @i=1; \
- while [ $$i -le $(NDATANODES) ] ; do \
- make -f $(THIS_MAKEFILE_PATH) CUR_DATANODE=$$i stop-datanode-container; \
- i=$$((i+1)); \
- done
- @echo "Stop All Containers Done!"
-
-stop-namenode-container:
- @echo "stop ${CLUSTER_ID}-namenode container"
- @if [ ! -z "`docker ps -a --filter="name=${CLUSTER_ID}-namenode" | grep -v CONTAINER`" ]; then \
- docker stop -t 0 ${CLUSTER_ID}-namenode; \
- else \
- echo "${CLUSTER_ID}-namenode container does not exist!"; \
- fi
-
-stop-datanode-container:
- @echo "stop ${CLUSTER_ID}-datanode$(CUR_DATANODE) container"
- @if [ ! -z "`docker ps -a --filter="name=${CLUSTER_ID}-datanode$(CUR_DATANODE)" | grep -v CONTAINER`" ]; then \
- docker stop -t 0 ${CLUSTER_ID}-datanode$(CUR_DATANODE); \
- else \
- echo "${CLUSTER_ID}-datanode$(CUR_DATANODE) container does not exist!"; \
- fi
-
-remove-hdfs:
- @make -f $(THIS_MAKEFILE_PATH) remove-namenode-container
- @i=1; \
- while [ $$i -le $(NDATANODES) ] ; do \
- make -f $(THIS_MAKEFILE_PATH) CUR_DATANODE=$$i remove-datanode-container; \
- i=$$((i+1)); \
- done
- @echo "Remove HDFS Done!"
-
-remove-namenode-container:
- @echo "make ${CLUSTER_ID}-namenode container"
- @if [ ! -z "`docker ps -a --filter="name=${CLUSTER_ID}-namenode" | grep -v CONTAINER`" ]; then \
- docker rm -v ${CLUSTER_ID}-namenode; \
- else \
- echo "${CLUSTER_ID}-namenode container does not exist!"; \
- fi
-
-remove-datanode-container:
- @echo "make ${CLUSTER_ID}-datanode$(CUR_DATANODE) container"
- @if [ ! -z "`docker ps -a --filter="name=${CLUSTER_ID}-datanode$(CUR_DATANODE)" | grep -v CONTAINER`" ]; then \
- docker rm -v ${CLUSTER_ID}-datanode$(CUR_DATANODE); \
- else \
- echo "${CLUSTER_ID}-datanode$(CUR_DATANODE) container does not exist!"; \
- fi
-
-remove-data:
- @echo remove ${CLUSTER_ID}-data container
- @if [ ! -z "`docker ps -a --filter="name=${CLUSTER_ID}-data" | grep -v CONTAINER`" ]; then \
- docker rm -v ${CLUSTER_ID}-data; \
- else \
- echo "${CLUSTER_ID}-data container does not exist!"; \
- fi
-
-pull:
- @echo latest images
- docker pull hawq/hawq-dev:$(OS_VERSION)
- docker pull hawq/hawq-test:$(OS_VERSION)
-
-clean:
- @make -f $(THIS_MAKEFILE_PATH) stop 2>&1 >/dev/null || true
- @make -f $(THIS_MAKEFILE_PATH) remove-hdfs 2>&1 >/dev/null || true
- @echo "Clean Done!"
-
-distclean:
- @make -f $(THIS_MAKEFILE_PATH) stop 2>&1 >/dev/null || true
- @make -f $(THIS_MAKEFILE_PATH) remove-hdfs 2>&1 >/dev/null || true
- @make -f $(THIS_MAKEFILE_PATH) remove-data 2>&1 >/dev/null || true
- @if [ ! -z "`docker network ls 2>/dev/null | grep $(NETWORK)`" ]; then \
- echo remove network $(NETWORK); \
- docker network rm $(NETWORK) 2>&1 >/dev/null || true; \
- fi
- @echo "Distclean Done!"
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/1cb29096/contrib/hawq-docker/README.md
----------------------------------------------------------------------
diff --git a/contrib/hawq-docker/README.md b/contrib/hawq-docker/README.md
deleted file mode 100644
index 4adeaaf..0000000
--- a/contrib/hawq-docker/README.md
+++ /dev/null
@@ -1,97 +0,0 @@
-# hawq-docker
-
-hawq-docker is based on *wangzw's* repo *hawq-devel-env*. It is the docker images and scripts to help developers of Apache HAWQ to setup building and testing environment with docker.
-
-Both CentOS 7 and CentOS 6 are supported.
-Change variable **OS_VERSION** (:= centos7 OR centos6) in Makefile to switch between CentOS 7 and CentOS 6.
-
-Take CentOS 7 as an example below.
-
-# Install docker
-* following the instructions to install docker.
-https://docs.docker.com/
-
-# Setup build and test environment
-* clone hawq repository
-```
-git clone https://github.com/apache/incubator-hawq.git .
-cd incubator-hawq/contrib/hawq-docker
-```
-* Get the docker images
-```
- make pull (recommended)
-OR
- make build
-```
-(Command `make pull` is to pull docker images from Docker Hub, while command `make build` is to build docker images locally. In general, `make pull` is faster than `make build`.)
-* setup a 5 nodes virtual cluster for Apache HAWQ build and test.
-```
-make run
-```
-Now let's have a look about what we creted.
-```
-[root@localhost hawq-docker]# docker ps -a
-CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
-382b2b3360d1 hawq/hawq-test:centos7 "entrypoint.sh bash" 2 minutes ago Up 2 minutes centos7-datanode3
-86513c331d45 hawq/hawq-test:centos7 "entrypoint.sh bash" 2 minutes ago Up 2 minutes centos7-datanode2
-c0ab10e46e4a hawq/hawq-test:centos7 "entrypoint.sh bash" 2 minutes ago Up 2 minutes centos7-datanode1
-e27beea63953 hawq/hawq-test:centos7 "entrypoint.sh bash" 2 minutes ago Up 2 minutes centos7-namenode
-1f986959bd04 hawq/hawq-dev:centos7 "/bin/true" 2 minutes ago Created centos7-data
-```
-**centos7-data** is a data container and mounted to /data directory on all other containers to provide a shared storage for the cluster.
-
-# Build and Test Apache HAWQ
-* attach to namenode
-```
-docker exec -it centos7-namenode bash
-```
-* check if HDFS working well
-```
-sudo -u hdfs hdfs dfsadmin -report
-```
-* clone Apache HAWQ code to /data direcotry
-```
-git clone https://github.com/apache/incubator-hawq.git /data/hawq
-```
-* build Apache HAWQ
-```
-cd /data/hawq
-./configure --prefix=/data/hawq-dev
-make
-make install
-```
-(When you are using CentOS 6, run command `scl enable devtoolset-2 bash` before
-configuring hawq and run command `exit` after installing hawq.)
-* modify Apache HAWQ configuration
-```
-sed 's|localhost|centos7-namenode|g' -i /data/hawq-dev/etc/hawq-site.xml
-echo 'centos7-datanode1' > /data/hawq-dev/etc/slaves
-echo 'centos7-datanode2' >> /data/hawq-dev/etc/slaves
-echo 'centos7-datanode3' >> /data/hawq-dev/etc/slaves
-```
-* Initialize Apache HAWQ cluster
-```
-sudo -u hdfs hdfs dfs -chown gpadmin /
-source /data/hawq-dev/greenplum_path.sh
-hawq init cluster
-```
-Now you can connect to database with `psql` command.
-```
-[gpadmin@centos7-namenode data]$ psql -d postgres
-psql (8.2.15)
-Type "help" for help.
-
-postgres=#
-```
-# More command with this script
-```
- Usage:
- To setup a build and test environment: make run
- To start all containers: make start
- To stop all containers: make stop
- To remove hdfs containers: make clean
- To remove all containers: make distclean
- To build images locally: make build
- To pull latest images: make pull
-```
-
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/1cb29096/contrib/hawq-docker/centos6-docker/hawq-dev/Dockerfile
----------------------------------------------------------------------
diff --git a/contrib/hawq-docker/centos6-docker/hawq-dev/Dockerfile b/contrib/hawq-docker/centos6-docker/hawq-dev/Dockerfile
deleted file mode 100644
index 9fb8476..0000000
--- a/contrib/hawq-docker/centos6-docker/hawq-dev/Dockerfile
+++ /dev/null
@@ -1,123 +0,0 @@
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied. See the License for the
-# specific language governing permissions and limitations
-# under the License.
-
-FROM centos:6
-
-MAINTAINER Richard Guo <ri...@pivotal.io>
-
-# install all software we need
-RUN yum install -y epel-release && \
- yum makecache && \
- yum install -y man passwd sudo tar which git mlocate links make bzip2 \
- autoconf automake libtool m4 gcc gcc-c++ gdb flex cmake gperf indent \
- libuuid-devel krb5-devel libgsasl-devel expat-devel libxml2-devel \
- perl-ExtUtils-Embed pam-devel python-devel snappy-devel \
- libyaml-devel libevent-devel bzip2-devel openssl-devel \
- openldap-devel readline-devel net-snmp-devel apr-devel \
- libesmtp-devel xerces-c-devel python-pip json-c-devel \
- apache-ivy java-1.7.0-openjdk-devel wget \
- openssh-clients openssh-server perl-JSON && \
- yum clean all
-
-# update gcc
-RUN wget -O /etc/yum.repos.d/slc6-devtoolset.repo http://linuxsoft.cern.ch/cern/devtoolset/slc6-devtoolset.repo && \
- rpm --import http://ftp.scientificlinux.org/linux/scientific/5x/x86_64/RPM-GPG-KEYs/RPM-GPG-KEY-cern && \
- yum install -y devtoolset-2-gcc devtoolset-2-binutils devtoolset-2-gcc-c++ && \
- echo "source /opt/rh/devtoolset-2/enable" >> ~/.bashrc && \
- source ~/.bashrc
-
-# install libcurl 7.45.0
-RUN mkdir -p /tmp/build/ && \
- cd /tmp/build && curl -L "http://curl.haxx.se/download/curl-7.45.0.tar.bz2" -o curl-7.45.0.tar.bz2 && \
- tar -xjf curl-7.45.0.tar.bz2 && cd curl-7.45.0 && \
- ./configure --prefix=/usr && make && make install && \
- rm -rf /tmp/build && ldconfig
-
-# install maven
-RUN curl -L "http://repos.fedorapeople.org/repos/dchen/apache-maven/epel-apache-maven.repo" -o /etc/yum.repos.d/epel-apache-maven.repo && \
- yum install -y apache-maven && \
- yum clean all
-
-# OS requirements
-RUN echo "kernel.sem = 250 512000 100 2048" >> /etc/sysctl.conf
-
-# setup ssh server and keys for root
-RUN ssh-keygen -t rsa -N "" -f ~/.ssh/id_rsa && \
- cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys && \
- chmod 0600 ~/.ssh/authorized_keys
-
-# setup JAVA_HOME for all users
-RUN echo "#!/bin/sh" > /etc/profile.d/java.sh && \
- echo "export JAVA_HOME=/etc/alternatives/java_sdk" >> /etc/profile.d/java.sh && \
- chmod a+x /etc/profile.d/java.sh
-
-# install boost 1.59
- RUN mkdir -p /tmp/build && \
- cd /tmp/build && curl -L "http://downloads.sourceforge.net/project/boost/boost/1.59.0/boost_1_59_0.tar.bz2" -o boost_1_59_0.tar.bz2 && \
- tar -xjf boost_1_59_0.tar.bz2 && cd boost_1_59_0 && \
- ./bootstrap.sh && ./b2 --prefix=/usr -q && ./b2 --prefix=/usr -q install && \
- rm -rf /tmp/build
-
-# install bison 2.5.1
-RUN mkdir -p /tmp/build/ && \
- cd /tmp/build && curl -L "ftp://ftp.gnu.org/gnu/bison/bison-2.5.1.tar.gz" -o bison-2.5.1.tar.gz && \
- tar -xzf bison-2.5.1.tar.gz && cd bison-2.5.1 && \
- ./configure --prefix=/usr && make && make install && \
- rm -rf /tmp/build
-
-# install thrift 0.9.1
-RUN mkdir -p /tmp/build && \
- cd /tmp/build && curl -L "https://archive.apache.org/dist/thrift/0.9.1/thrift-0.9.1.tar.gz" -o thrift-0.9.1.tar.gz && \
- tar -xf thrift-0.9.1.tar.gz && cd thrift-0.9.1 && \
- ./configure --prefix=/usr --without-tests && \
- make && make install && \
- rm -rf /tmp/build
-
-# install protobuf 2.5.0
-RUN mkdir -p /tmp/build/ && \
- cd /tmp/build && curl -L "https://github.com/google/protobuf/releases/download/v2.5.0/protobuf-2.5.0.tar.bz2" -o protobuf-2.5.0.tar.bz2 && \
- tar -xjf protobuf-2.5.0.tar.bz2 && cd protobuf-2.5.0 && \
- ./configure --prefix=/usr && make && make install && ldconfig && \
- rm -rf /tmp/build
-
-# install python module
-RUN pip --retries=50 --timeout=300 install pycrypto
-
-# create user gpadmin since HAWQ cannot run under root
-RUN groupadd -g 1000 gpadmin && \
- useradd -u 1000 -g 1000 gpadmin && \
- echo "gpadmin ALL=(ALL) NOPASSWD: ALL" > /etc/sudoers.d/gpadmin
-
-# sudo should not require tty
-RUN sed -i -e 's|Defaults requiretty|#Defaults requiretty|' /etc/sudoers
-
-RUN echo "#!/bin/bash" > /etc/profile.d/user.sh && \
- echo "export USER=\`whoami\`" >> /etc/profile.d/user.sh && \
- chmod a+x /etc/profile.d/user.sh
-
-ENV BASEDIR /data
-RUN mkdir -p /data && chmod 777 /data
-
-USER gpadmin
-
-# setup ssh client keys for gpadmin
-RUN ssh-keygen -t rsa -N "" -f ~/.ssh/id_rsa && \
- cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys && \
- chmod 0600 ~/.ssh/authorized_keys
-
-WORKDIR /data
-
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/1cb29096/contrib/hawq-docker/centos6-docker/hawq-test/Dockerfile
----------------------------------------------------------------------
diff --git a/contrib/hawq-docker/centos6-docker/hawq-test/Dockerfile b/contrib/hawq-docker/centos6-docker/hawq-test/Dockerfile
deleted file mode 100644
index 94a04fe..0000000
--- a/contrib/hawq-docker/centos6-docker/hawq-test/Dockerfile
+++ /dev/null
@@ -1,40 +0,0 @@
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied. See the License for the
-# specific language governing permissions and limitations
-# under the License.
-
-FROM hawq/hawq-dev:centos6
-
-MAINTAINER Richard Guo <ri...@pivotal.io>
-
-USER root
-
-# install HDP 2.5.0
-RUN curl -L "http://public-repo-1.hortonworks.com/HDP/centos6/2.x/updates/2.5.0.0/hdp.repo" -o /etc/yum.repos.d/hdp.repo && \
- yum install -y hadoop hadoop-hdfs hadoop-libhdfs hadoop-yarn hadoop-mapreduce hadoop-client hdp-select && \
- yum clean all
-
-RUN ln -s /usr/hdp/current/hadoop-hdfs-namenode/../hadoop/sbin/hadoop-daemon.sh /usr/bin/hadoop-daemon.sh
-
-COPY conf/* /etc/hadoop/conf/
-
-COPY entrypoint.sh /usr/bin/entrypoint.sh
-COPY start-hdfs.sh /usr/bin/start-hdfs.sh
-
-USER gpadmin
-
-ENTRYPOINT ["entrypoint.sh"]
-CMD ["bash"]
-
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/1cb29096/contrib/hawq-docker/centos6-docker/hawq-test/conf/capacity-scheduler.xml
----------------------------------------------------------------------
diff --git a/contrib/hawq-docker/centos6-docker/hawq-test/conf/capacity-scheduler.xml b/contrib/hawq-docker/centos6-docker/hawq-test/conf/capacity-scheduler.xml
deleted file mode 100644
index 30f4eb9..0000000
--- a/contrib/hawq-docker/centos6-docker/hawq-test/conf/capacity-scheduler.xml
+++ /dev/null
@@ -1,134 +0,0 @@
-<!--
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License. See accompanying LICENSE file.
--->
-<configuration>
-
- <property>
- <name>yarn.scheduler.capacity.maximum-applications</name>
- <value>10000</value>
- <description>
- Maximum number of applications that can be pending and running.
- </description>
- </property>
-
- <property>
- <name>yarn.scheduler.capacity.maximum-am-resource-percent</name>
- <value>0.1</value>
- <description>
- Maximum percent of resources in the cluster which can be used to run
- application masters i.e. controls number of concurrent running
- applications.
- </description>
- </property>
-
- <property>
- <name>yarn.scheduler.capacity.resource-calculator</name>
- <value>org.apache.hadoop.yarn.util.resource.DefaultResourceCalculator</value>
- <description>
- The ResourceCalculator implementation to be used to compare
- Resources in the scheduler.
- The default i.e. DefaultResourceCalculator only uses Memory while
- DominantResourceCalculator uses dominant-resource to compare
- multi-dimensional resources such as Memory, CPU etc.
- </description>
- </property>
-
- <property>
- <name>yarn.scheduler.capacity.root.queues</name>
- <value>default</value>
- <description>
- The queues at the this level (root is the root queue).
- </description>
- </property>
-
- <property>
- <name>yarn.scheduler.capacity.root.default.capacity</name>
- <value>100</value>
- <description>Default queue target capacity.</description>
- </property>
-
- <property>
- <name>yarn.scheduler.capacity.root.default.user-limit-factor</name>
- <value>1</value>
- <description>
- Default queue user limit a percentage from 0.0 to 1.0.
- </description>
- </property>
-
- <property>
- <name>yarn.scheduler.capacity.root.default.maximum-capacity</name>
- <value>100</value>
- <description>
- The maximum capacity of the default queue.
- </description>
- </property>
-
- <property>
- <name>yarn.scheduler.capacity.root.default.state</name>
- <value>RUNNING</value>
- <description>
- The state of the default queue. State can be one of RUNNING or STOPPED.
- </description>
- </property>
-
- <property>
- <name>yarn.scheduler.capacity.root.default.acl_submit_applications</name>
- <value>*</value>
- <description>
- The ACL of who can submit jobs to the default queue.
- </description>
- </property>
-
- <property>
- <name>yarn.scheduler.capacity.root.default.acl_administer_queue</name>
- <value>*</value>
- <description>
- The ACL of who can administer jobs on the default queue.
- </description>
- </property>
-
- <property>
- <name>yarn.scheduler.capacity.node-locality-delay</name>
- <value>40</value>
- <description>
- Number of missed scheduling opportunities after which the CapacityScheduler
- attempts to schedule rack-local containers.
- Typically this should be set to number of nodes in the cluster, By default is setting
- approximately number of nodes in one rack which is 40.
- </description>
- </property>
-
- <property>
- <name>yarn.scheduler.capacity.queue-mappings</name>
- <value></value>
- <description>
- A list of mappings that will be used to assign jobs to queues
- The syntax for this list is [u|g]:[name]:[queue_name][,next mapping]*
- Typically this list will be used to map users to queues,
- for example, u:%user:%user maps all users to queues with the same name
- as the user.
- </description>
- </property>
-
- <property>
- <name>yarn.scheduler.capacity.queue-mappings-override.enable</name>
- <value>false</value>
- <description>
- If a queue mapping is present, will it override the value specified
- by the user? This can be used by administrators to place jobs in queues
- that are different than the one specified by the user.
- The default is false.
- </description>
- </property>
-
-</configuration>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/1cb29096/contrib/hawq-docker/centos6-docker/hawq-test/conf/configuration.xsl
----------------------------------------------------------------------
diff --git a/contrib/hawq-docker/centos6-docker/hawq-test/conf/configuration.xsl b/contrib/hawq-docker/centos6-docker/hawq-test/conf/configuration.xsl
deleted file mode 100644
index d50d80b..0000000
--- a/contrib/hawq-docker/centos6-docker/hawq-test/conf/configuration.xsl
+++ /dev/null
@@ -1,40 +0,0 @@
-<?xml version="1.0"?>
-<!--
- Licensed to the Apache Software Foundation (ASF) under one or more
- contributor license agreements. See the NOTICE file distributed with
- this work for additional information regarding copyright ownership.
- The ASF licenses this file to You under the Apache License, Version 2.0
- (the "License"); you may not use this file except in compliance with
- the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License.
--->
-<xsl:stylesheet xmlns:xsl="http://www.w3.org/1999/XSL/Transform" version="1.0">
-<xsl:output method="html"/>
-<xsl:template match="configuration">
-<html>
-<body>
-<table border="1">
-<tr>
- <td>name</td>
- <td>value</td>
- <td>description</td>
-</tr>
-<xsl:for-each select="property">
-<tr>
- <td><a name="{name}"><xsl:value-of select="name"/></a></td>
- <td><xsl:value-of select="value"/></td>
- <td><xsl:value-of select="description"/></td>
-</tr>
-</xsl:for-each>
-</table>
-</body>
-</html>
-</xsl:template>
-</xsl:stylesheet>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/1cb29096/contrib/hawq-docker/centos6-docker/hawq-test/conf/container-executor.cfg
----------------------------------------------------------------------
diff --git a/contrib/hawq-docker/centos6-docker/hawq-test/conf/container-executor.cfg b/contrib/hawq-docker/centos6-docker/hawq-test/conf/container-executor.cfg
deleted file mode 100644
index d68cee8..0000000
--- a/contrib/hawq-docker/centos6-docker/hawq-test/conf/container-executor.cfg
+++ /dev/null
@@ -1,4 +0,0 @@
-yarn.nodemanager.linux-container-executor.group=#configured value of yarn.nodemanager.linux-container-executor.group
-banned.users=#comma separated list of users who can not run applications
-min.user.id=1000#Prevent other super-users
-allowed.system.users=##comma separated list of system users who CAN run applications
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/1cb29096/contrib/hawq-docker/centos6-docker/hawq-test/conf/core-site.xml
----------------------------------------------------------------------
diff --git a/contrib/hawq-docker/centos6-docker/hawq-test/conf/core-site.xml b/contrib/hawq-docker/centos6-docker/hawq-test/conf/core-site.xml
deleted file mode 100644
index afc37fc..0000000
--- a/contrib/hawq-docker/centos6-docker/hawq-test/conf/core-site.xml
+++ /dev/null
@@ -1,24 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
- <property>
- <name>fs.defaultFS</name>
- <value>hdfs://${hdfs.namenode}:8020</value>
- </property>
-</configuration>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/1cb29096/contrib/hawq-docker/centos6-docker/hawq-test/conf/hadoop-env.cmd
----------------------------------------------------------------------
diff --git a/contrib/hawq-docker/centos6-docker/hawq-test/conf/hadoop-env.cmd b/contrib/hawq-docker/centos6-docker/hawq-test/conf/hadoop-env.cmd
deleted file mode 100644
index bb40ec9..0000000
--- a/contrib/hawq-docker/centos6-docker/hawq-test/conf/hadoop-env.cmd
+++ /dev/null
@@ -1,92 +0,0 @@
-@echo off
-@rem Licensed to the Apache Software Foundation (ASF) under one or more
-@rem contributor license agreements. See the NOTICE file distributed with
-@rem this work for additional information regarding copyright ownership.
-@rem The ASF licenses this file to You under the Apache License, Version 2.0
-@rem (the "License"); you may not use this file except in compliance with
-@rem the License. You may obtain a copy of the License at
-@rem
-@rem http://www.apache.org/licenses/LICENSE-2.0
-@rem
-@rem Unless required by applicable law or agreed to in writing, software
-@rem distributed under the License is distributed on an "AS IS" BASIS,
-@rem WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-@rem See the License for the specific language governing permissions and
-@rem limitations under the License.
-
-@rem Set Hadoop-specific environment variables here.
-
-@rem The only required environment variable is JAVA_HOME. All others are
-@rem optional. When running a distributed configuration it is best to
-@rem set JAVA_HOME in this file, so that it is correctly defined on
-@rem remote nodes.
-
-@rem The java implementation to use. Required.
-set JAVA_HOME=%JAVA_HOME%
-
-@rem The jsvc implementation to use. Jsvc is required to run secure datanodes.
-@rem set JSVC_HOME=%JSVC_HOME%
-
-@rem set HADOOP_CONF_DIR=
-
-@rem Extra Java CLASSPATH elements. Automatically insert capacity-scheduler.
-if exist %HADOOP_HOME%\contrib\capacity-scheduler (
- if not defined HADOOP_CLASSPATH (
- set HADOOP_CLASSPATH=%HADOOP_HOME%\contrib\capacity-scheduler\*.jar
- ) else (
- set HADOOP_CLASSPATH=%HADOOP_CLASSPATH%;%HADOOP_HOME%\contrib\capacity-scheduler\*.jar
- )
-)
-
-@rem If TEZ_CLASSPATH is defined in the env, that means that TEZ is enabled
-@rem append it to the HADOOP_CLASSPATH
-
-if defined TEZ_CLASSPATH (
- if not defined HADOOP_CLASSPATH (
- set HADOOP_CLASSPATH=%TEZ_CLASSPATH%
- ) else (
- set HADOOP_CLASSPATH=%HADOOP_CLASSPATH%;%TEZ_CLASSPATH%
- )
-)
-
-@rem The maximum amount of heap to use, in MB. Default is 1000.
-@rem set HADOOP_HEAPSIZE=
-@rem set HADOOP_NAMENODE_INIT_HEAPSIZE=""
-
-@rem Extra Java runtime options. Empty by default.
-@rem set HADOOP_OPTS=%HADOOP_OPTS% -Djava.net.preferIPv4Stack=true
-
-@rem Command specific options appended to HADOOP_OPTS when specified
-if not defined HADOOP_SECURITY_LOGGER (
- set HADOOP_SECURITY_LOGGER=INFO,RFAS
-)
-if not defined HDFS_AUDIT_LOGGER (
- set HDFS_AUDIT_LOGGER=INFO,NullAppender
-)
-
-set HADOOP_NAMENODE_OPTS=-Dhadoop.security.logger=%HADOOP_SECURITY_LOGGER% -Dhdfs.audit.logger=%HDFS_AUDIT_LOGGER% %HADOOP_NAMENODE_OPTS%
-set HADOOP_DATANODE_OPTS=-Dhadoop.security.logger=ERROR,RFAS %HADOOP_DATANODE_OPTS%
-set HADOOP_SECONDARYNAMENODE_OPTS=-Dhadoop.security.logger=%HADOOP_SECURITY_LOGGER% -Dhdfs.audit.logger=%HDFS_AUDIT_LOGGER% %HADOOP_SECONDARYNAMENODE_OPTS%
-
-@rem The following applies to multiple commands (fs, dfs, fsck, distcp etc)
-set HADOOP_CLIENT_OPTS=-Xmx512m %HADOOP_CLIENT_OPTS%
-@rem set HADOOP_JAVA_PLATFORM_OPTS="-XX:-UsePerfData %HADOOP_JAVA_PLATFORM_OPTS%"
-
-@rem On secure datanodes, user to run the datanode as after dropping privileges
-set HADOOP_SECURE_DN_USER=%HADOOP_SECURE_DN_USER%
-
-@rem Where log files are stored. %HADOOP_HOME%/logs by default.
-@rem set HADOOP_LOG_DIR=%HADOOP_LOG_DIR%\%USERNAME%
-
-@rem Where log files are stored in the secure data environment.
-set HADOOP_SECURE_DN_LOG_DIR=%HADOOP_LOG_DIR%\%HADOOP_HDFS_USER%
-
-@rem The directory where pid files are stored. /tmp by default.
-@rem NOTE: this should be set to a directory that can only be written to by
-@rem the user that will run the hadoop daemons. Otherwise there is the
-@rem potential for a symlink attack.
-set HADOOP_PID_DIR=%HADOOP_PID_DIR%
-set HADOOP_SECURE_DN_PID_DIR=%HADOOP_PID_DIR%
-
-@rem A string representing this instance of hadoop. %USERNAME% by default.
-set HADOOP_IDENT_STRING=%USERNAME%
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/1cb29096/contrib/hawq-docker/centos6-docker/hawq-test/conf/hadoop-env.sh
----------------------------------------------------------------------
diff --git a/contrib/hawq-docker/centos6-docker/hawq-test/conf/hadoop-env.sh b/contrib/hawq-docker/centos6-docker/hawq-test/conf/hadoop-env.sh
deleted file mode 100644
index 95511ed..0000000
--- a/contrib/hawq-docker/centos6-docker/hawq-test/conf/hadoop-env.sh
+++ /dev/null
@@ -1,110 +0,0 @@
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# Set Hadoop-specific environment variables here.
-
-# The only required environment variable is JAVA_HOME. All others are
-# optional. When running a distributed configuration it is best to
-# set JAVA_HOME in this file, so that it is correctly defined on
-# remote nodes.
-
-# The java implementation to use.
-export JAVA_HOME=/etc/alternatives/java_sdk
-
-# The jsvc implementation to use. Jsvc is required to run secure datanodes
-# that bind to privileged ports to provide authentication of data transfer
-# protocol. Jsvc is not required if SASL is configured for authentication of
-# data transfer protocol using non-privileged ports.
-#export JSVC_HOME=${JSVC_HOME}
-
-#export HADOOP_CONF_DIR=${HADOOP_CONF_DIR:-"/etc/hadoop"}
-
-# Extra Java CLASSPATH elements. Automatically insert capacity-scheduler.
-#for f in $HADOOP_HOME/contrib/capacity-scheduler/*.jar; do
-# if [ "$HADOOP_CLASSPATH" ]; then
-# export HADOOP_CLASSPATH=$HADOOP_CLASSPATH:$f
-# else
-# export HADOOP_CLASSPATH=$f
-# fi
-#done
-
-# The maximum amount of heap to use, in MB. Default is 1000.
-#export HADOOP_HEAPSIZE=
-#export HADOOP_NAMENODE_INIT_HEAPSIZE=""
-
-# Setup environment variable for docker image
-if [ -f /etc/profile.d/hadoop.sh ]; then
- . /etc/profile.d/hadoop.sh
-fi
-
-if [ -z "${NAMENODE}" ]; then
- echo "environment variable NAMENODE is not set!"
- exit 1
-fi
-
-# Extra Java runtime options. Empty by default.
-export HADOOP_OPTS="-Dhdfs.namenode=${NAMENODE}"
-#export HADOOP_OPTS="$HADOOP_OPTS -Djava.net.preferIPv4Stack=true"
-
-# Command specific options appended to HADOOP_OPTS when specified
-#export HADOOP_NAMENODE_OPTS="-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} $HADOOP_NAMENODE_OPTS"
-#export HADOOP_DATANODE_OPTS="-Dhadoop.security.logger=ERROR,RFAS $HADOOP_DATANODE_OPTS"
-
-#export HADOOP_SECONDARYNAMENODE_OPTS="-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} $HADOOP_SECONDARYNAMENODE_OPTS"
-
-#export HADOOP_NFS3_OPTS="$HADOOP_NFS3_OPTS"
-#export HADOOP_PORTMAP_OPTS="-Xmx512m $HADOOP_PORTMAP_OPTS"
-
-# The following applies to multiple commands (fs, dfs, fsck, distcp etc)
-#export HADOOP_CLIENT_OPTS="-Xmx512m $HADOOP_CLIENT_OPTS"
-#HADOOP_JAVA_PLATFORM_OPTS="-XX:-UsePerfData $HADOOP_JAVA_PLATFORM_OPTS"
-
-# On secure datanodes, user to run the datanode as after dropping privileges.
-# This **MUST** be uncommented to enable secure HDFS if using privileged ports
-# to provide authentication of data transfer protocol. This **MUST NOT** be
-# defined if SASL is configured for authentication of data transfer protocol
-# using non-privileged ports.
-#export HADOOP_SECURE_DN_USER=${HADOOP_SECURE_DN_USER}
-
-# Where log files are stored. $HADOOP_HOME/logs by default.
-export HADOOP_LOG_DIR=/var/log/hadoop
-export HADOOP_LOG_DIR=${HADOOP_LOG_DIR}/$USER
-
-# Where log files are stored in the secure data environment.
-#export HADOOP_SECURE_DN_LOG_DIR=${HADOOP_LOG_DIR}/${HADOOP_HDFS_USER}
-
-###
-# HDFS Mover specific parameters
-###
-# Specify the JVM options to be used when starting the HDFS Mover.
-# These options will be appended to the options specified as HADOOP_OPTS
-# and therefore may override any similar flags set in HADOOP_OPTS
-#
-# export HADOOP_MOVER_OPTS=""
-
-###
-# Advanced Users Only!
-###
-
-# The directory where pid files are stored. /tmp by default.
-# NOTE: this should be set to a directory that can only be written to by
-# the user that will run the hadoop daemons. Otherwise there is the
-# potential for a symlink attack.
-#export HADOOP_PID_DIR=${HADOOP_PID_DIR}
-#export HADOOP_SECURE_DN_PID_DIR=${HADOOP_PID_DIR}
-
-# A string representing this instance of hadoop. $USER by default.
-#export HADOOP_IDENT_STRING=$USER
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/1cb29096/contrib/hawq-docker/centos6-docker/hawq-test/conf/hadoop-metrics.properties
----------------------------------------------------------------------
diff --git a/contrib/hawq-docker/centos6-docker/hawq-test/conf/hadoop-metrics.properties b/contrib/hawq-docker/centos6-docker/hawq-test/conf/hadoop-metrics.properties
deleted file mode 100644
index c1b2eb7..0000000
--- a/contrib/hawq-docker/centos6-docker/hawq-test/conf/hadoop-metrics.properties
+++ /dev/null
@@ -1,75 +0,0 @@
-# Configuration of the "dfs" context for null
-dfs.class=org.apache.hadoop.metrics.spi.NullContext
-
-# Configuration of the "dfs" context for file
-#dfs.class=org.apache.hadoop.metrics.file.FileContext
-#dfs.period=10
-#dfs.fileName=/tmp/dfsmetrics.log
-
-# Configuration of the "dfs" context for ganglia
-# Pick one: Ganglia 3.0 (former) or Ganglia 3.1 (latter)
-# dfs.class=org.apache.hadoop.metrics.ganglia.GangliaContext
-# dfs.class=org.apache.hadoop.metrics.ganglia.GangliaContext31
-# dfs.period=10
-# dfs.servers=localhost:8649
-
-
-# Configuration of the "mapred" context for null
-mapred.class=org.apache.hadoop.metrics.spi.NullContext
-
-# Configuration of the "mapred" context for file
-#mapred.class=org.apache.hadoop.metrics.file.FileContext
-#mapred.period=10
-#mapred.fileName=/tmp/mrmetrics.log
-
-# Configuration of the "mapred" context for ganglia
-# Pick one: Ganglia 3.0 (former) or Ganglia 3.1 (latter)
-# mapred.class=org.apache.hadoop.metrics.ganglia.GangliaContext
-# mapred.class=org.apache.hadoop.metrics.ganglia.GangliaContext31
-# mapred.period=10
-# mapred.servers=localhost:8649
-
-
-# Configuration of the "jvm" context for null
-#jvm.class=org.apache.hadoop.metrics.spi.NullContext
-
-# Configuration of the "jvm" context for file
-#jvm.class=org.apache.hadoop.metrics.file.FileContext
-#jvm.period=10
-#jvm.fileName=/tmp/jvmmetrics.log
-
-# Configuration of the "jvm" context for ganglia
-# jvm.class=org.apache.hadoop.metrics.ganglia.GangliaContext
-# jvm.class=org.apache.hadoop.metrics.ganglia.GangliaContext31
-# jvm.period=10
-# jvm.servers=localhost:8649
-
-# Configuration of the "rpc" context for null
-rpc.class=org.apache.hadoop.metrics.spi.NullContext
-
-# Configuration of the "rpc" context for file
-#rpc.class=org.apache.hadoop.metrics.file.FileContext
-#rpc.period=10
-#rpc.fileName=/tmp/rpcmetrics.log
-
-# Configuration of the "rpc" context for ganglia
-# rpc.class=org.apache.hadoop.metrics.ganglia.GangliaContext
-# rpc.class=org.apache.hadoop.metrics.ganglia.GangliaContext31
-# rpc.period=10
-# rpc.servers=localhost:8649
-
-
-# Configuration of the "ugi" context for null
-ugi.class=org.apache.hadoop.metrics.spi.NullContext
-
-# Configuration of the "ugi" context for file
-#ugi.class=org.apache.hadoop.metrics.file.FileContext
-#ugi.period=10
-#ugi.fileName=/tmp/ugimetrics.log
-
-# Configuration of the "ugi" context for ganglia
-# ugi.class=org.apache.hadoop.metrics.ganglia.GangliaContext
-# ugi.class=org.apache.hadoop.metrics.ganglia.GangliaContext31
-# ugi.period=10
-# ugi.servers=localhost:8649
-
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/1cb29096/contrib/hawq-docker/centos6-docker/hawq-test/conf/hadoop-metrics2.properties
----------------------------------------------------------------------
diff --git a/contrib/hawq-docker/centos6-docker/hawq-test/conf/hadoop-metrics2.properties b/contrib/hawq-docker/centos6-docker/hawq-test/conf/hadoop-metrics2.properties
deleted file mode 100644
index 0c09228..0000000
--- a/contrib/hawq-docker/centos6-docker/hawq-test/conf/hadoop-metrics2.properties
+++ /dev/null
@@ -1,68 +0,0 @@
-# syntax: [prefix].[source|sink].[instance].[options]
-# See javadoc of package-info.java for org.apache.hadoop.metrics2 for details
-
-*.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink
-# default sampling period, in seconds
-*.period=10
-
-# The namenode-metrics.out will contain metrics from all context
-#namenode.sink.file.filename=namenode-metrics.out
-# Specifying a special sampling period for namenode:
-#namenode.sink.*.period=8
-
-#datanode.sink.file.filename=datanode-metrics.out
-
-#resourcemanager.sink.file.filename=resourcemanager-metrics.out
-
-#nodemanager.sink.file.filename=nodemanager-metrics.out
-
-#mrappmaster.sink.file.filename=mrappmaster-metrics.out
-
-#jobhistoryserver.sink.file.filename=jobhistoryserver-metrics.out
-
-# the following example split metrics of different
-# context to different sinks (in this case files)
-#nodemanager.sink.file_jvm.class=org.apache.hadoop.metrics2.sink.FileSink
-#nodemanager.sink.file_jvm.context=jvm
-#nodemanager.sink.file_jvm.filename=nodemanager-jvm-metrics.out
-#nodemanager.sink.file_mapred.class=org.apache.hadoop.metrics2.sink.FileSink
-#nodemanager.sink.file_mapred.context=mapred
-#nodemanager.sink.file_mapred.filename=nodemanager-mapred-metrics.out
-
-#
-# Below are for sending metrics to Ganglia
-#
-# for Ganglia 3.0 support
-# *.sink.ganglia.class=org.apache.hadoop.metrics2.sink.ganglia.GangliaSink30
-#
-# for Ganglia 3.1 support
-# *.sink.ganglia.class=org.apache.hadoop.metrics2.sink.ganglia.GangliaSink31
-
-# *.sink.ganglia.period=10
-
-# default for supportsparse is false
-# *.sink.ganglia.supportsparse=true
-
-#*.sink.ganglia.slope=jvm.metrics.gcCount=zero,jvm.metrics.memHeapUsedM=both
-#*.sink.ganglia.dmax=jvm.metrics.threadsBlocked=70,jvm.metrics.memHeapUsedM=40
-
-# Tag values to use for the ganglia prefix. If not defined no tags are used.
-# If '*' all tags are used. If specifiying multiple tags separate them with
-# commas. Note that the last segment of the property name is the context name.
-#
-#*.sink.ganglia.tagsForPrefix.jvm=ProcesName
-#*.sink.ganglia.tagsForPrefix.dfs=
-#*.sink.ganglia.tagsForPrefix.rpc=
-#*.sink.ganglia.tagsForPrefix.mapred=
-
-#namenode.sink.ganglia.servers=yourgangliahost_1:8649,yourgangliahost_2:8649
-
-#datanode.sink.ganglia.servers=yourgangliahost_1:8649,yourgangliahost_2:8649
-
-#resourcemanager.sink.ganglia.servers=yourgangliahost_1:8649,yourgangliahost_2:8649
-
-#nodemanager.sink.ganglia.servers=yourgangliahost_1:8649,yourgangliahost_2:8649
-
-#mrappmaster.sink.ganglia.servers=yourgangliahost_1:8649,yourgangliahost_2:8649
-
-#jobhistoryserver.sink.ganglia.servers=yourgangliahost_1:8649,yourgangliahost_2:8649
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/1cb29096/contrib/hawq-docker/centos6-docker/hawq-test/conf/hadoop-policy.xml
----------------------------------------------------------------------
diff --git a/contrib/hawq-docker/centos6-docker/hawq-test/conf/hadoop-policy.xml b/contrib/hawq-docker/centos6-docker/hawq-test/conf/hadoop-policy.xml
deleted file mode 100644
index 2bf5c02..0000000
--- a/contrib/hawq-docker/centos6-docker/hawq-test/conf/hadoop-policy.xml
+++ /dev/null
@@ -1,226 +0,0 @@
-<?xml version="1.0"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
-
- Licensed to the Apache Software Foundation (ASF) under one
- or more contributor license agreements. See the NOTICE file
- distributed with this work for additional information
- regarding copyright ownership. The ASF licenses this file
- to you under the Apache License, Version 2.0 (the
- "License"); you may not use this file except in compliance
- with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License.
-
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
- <property>
- <name>security.client.protocol.acl</name>
- <value>*</value>
- <description>ACL for ClientProtocol, which is used by user code
- via the DistributedFileSystem.
- The ACL is a comma-separated list of user and group names. The user and
- group list is separated by a blank. For e.g. "alice,bob users,wheel".
- A special value of "*" means all users are allowed.</description>
- </property>
-
- <property>
- <name>security.client.datanode.protocol.acl</name>
- <value>*</value>
- <description>ACL for ClientDatanodeProtocol, the client-to-datanode protocol
- for block recovery.
- The ACL is a comma-separated list of user and group names. The user and
- group list is separated by a blank. For e.g. "alice,bob users,wheel".
- A special value of "*" means all users are allowed.</description>
- </property>
-
- <property>
- <name>security.datanode.protocol.acl</name>
- <value>*</value>
- <description>ACL for DatanodeProtocol, which is used by datanodes to
- communicate with the namenode.
- The ACL is a comma-separated list of user and group names. The user and
- group list is separated by a blank. For e.g. "alice,bob users,wheel".
- A special value of "*" means all users are allowed.</description>
- </property>
-
- <property>
- <name>security.inter.datanode.protocol.acl</name>
- <value>*</value>
- <description>ACL for InterDatanodeProtocol, the inter-datanode protocol
- for updating generation timestamp.
- The ACL is a comma-separated list of user and group names. The user and
- group list is separated by a blank. For e.g. "alice,bob users,wheel".
- A special value of "*" means all users are allowed.</description>
- </property>
-
- <property>
- <name>security.namenode.protocol.acl</name>
- <value>*</value>
- <description>ACL for NamenodeProtocol, the protocol used by the secondary
- namenode to communicate with the namenode.
- The ACL is a comma-separated list of user and group names. The user and
- group list is separated by a blank. For e.g. "alice,bob users,wheel".
- A special value of "*" means all users are allowed.</description>
- </property>
-
- <property>
- <name>security.admin.operations.protocol.acl</name>
- <value>*</value>
- <description>ACL for AdminOperationsProtocol. Used for admin commands.
- The ACL is a comma-separated list of user and group names. The user and
- group list is separated by a blank. For e.g. "alice,bob users,wheel".
- A special value of "*" means all users are allowed.</description>
- </property>
-
- <property>
- <name>security.refresh.user.mappings.protocol.acl</name>
- <value>*</value>
- <description>ACL for RefreshUserMappingsProtocol. Used to refresh
- users mappings. The ACL is a comma-separated list of user and
- group names. The user and group list is separated by a blank. For
- e.g. "alice,bob users,wheel". A special value of "*" means all
- users are allowed.</description>
- </property>
-
- <property>
- <name>security.refresh.policy.protocol.acl</name>
- <value>*</value>
- <description>ACL for RefreshAuthorizationPolicyProtocol, used by the
- dfsadmin and mradmin commands to refresh the security policy in-effect.
- The ACL is a comma-separated list of user and group names. The user and
- group list is separated by a blank. For e.g. "alice,bob users,wheel".
- A special value of "*" means all users are allowed.</description>
- </property>
-
- <property>
- <name>security.ha.service.protocol.acl</name>
- <value>*</value>
- <description>ACL for HAService protocol used by HAAdmin to manage the
- active and stand-by states of namenode.</description>
- </property>
-
- <property>
- <name>security.zkfc.protocol.acl</name>
- <value>*</value>
- <description>ACL for access to the ZK Failover Controller
- </description>
- </property>
-
- <property>
- <name>security.qjournal.service.protocol.acl</name>
- <value>*</value>
- <description>ACL for QJournalProtocol, used by the NN to communicate with
- JNs when using the QuorumJournalManager for edit logs.</description>
- </property>
-
- <property>
- <name>security.mrhs.client.protocol.acl</name>
- <value>*</value>
- <description>ACL for HSClientProtocol, used by job clients to
- communciate with the MR History Server job status etc.
- The ACL is a comma-separated list of user and group names. The user and
- group list is separated by a blank. For e.g. "alice,bob users,wheel".
- A special value of "*" means all users are allowed.</description>
- </property>
-
- <!-- YARN Protocols -->
-
- <property>
- <name>security.resourcetracker.protocol.acl</name>
- <value>*</value>
- <description>ACL for ResourceTrackerProtocol, used by the
- ResourceManager and NodeManager to communicate with each other.
- The ACL is a comma-separated list of user and group names. The user and
- group list is separated by a blank. For e.g. "alice,bob users,wheel".
- A special value of "*" means all users are allowed.</description>
- </property>
-
- <property>
- <name>security.resourcemanager-administration.protocol.acl</name>
- <value>*</value>
- <description>ACL for ResourceManagerAdministrationProtocol, for admin commands.
- The ACL is a comma-separated list of user and group names. The user and
- group list is separated by a blank. For e.g. "alice,bob users,wheel".
- A special value of "*" means all users are allowed.</description>
- </property>
-
- <property>
- <name>security.applicationclient.protocol.acl</name>
- <value>*</value>
- <description>ACL for ApplicationClientProtocol, used by the ResourceManager
- and applications submission clients to communicate with each other.
- The ACL is a comma-separated list of user and group names. The user and
- group list is separated by a blank. For e.g. "alice,bob users,wheel".
- A special value of "*" means all users are allowed.</description>
- </property>
-
- <property>
- <name>security.applicationmaster.protocol.acl</name>
- <value>*</value>
- <description>ACL for ApplicationMasterProtocol, used by the ResourceManager
- and ApplicationMasters to communicate with each other.
- The ACL is a comma-separated list of user and group names. The user and
- group list is separated by a blank. For e.g. "alice,bob users,wheel".
- A special value of "*" means all users are allowed.</description>
- </property>
-
- <property>
- <name>security.containermanagement.protocol.acl</name>
- <value>*</value>
- <description>ACL for ContainerManagementProtocol protocol, used by the NodeManager
- and ApplicationMasters to communicate with each other.
- The ACL is a comma-separated list of user and group names. The user and
- group list is separated by a blank. For e.g. "alice,bob users,wheel".
- A special value of "*" means all users are allowed.</description>
- </property>
-
- <property>
- <name>security.resourcelocalizer.protocol.acl</name>
- <value>*</value>
- <description>ACL for ResourceLocalizer protocol, used by the NodeManager
- and ResourceLocalizer to communicate with each other.
- The ACL is a comma-separated list of user and group names. The user and
- group list is separated by a blank. For e.g. "alice,bob users,wheel".
- A special value of "*" means all users are allowed.</description>
- </property>
-
- <property>
- <name>security.job.task.protocol.acl</name>
- <value>*</value>
- <description>ACL for TaskUmbilicalProtocol, used by the map and reduce
- tasks to communicate with the parent tasktracker.
- The ACL is a comma-separated list of user and group names. The user and
- group list is separated by a blank. For e.g. "alice,bob users,wheel".
- A special value of "*" means all users are allowed.</description>
- </property>
-
- <property>
- <name>security.job.client.protocol.acl</name>
- <value>*</value>
- <description>ACL for MRClientProtocol, used by job clients to
- communciate with the MR ApplicationMaster to query job status etc.
- The ACL is a comma-separated list of user and group names. The user and
- group list is separated by a blank. For e.g. "alice,bob users,wheel".
- A special value of "*" means all users are allowed.</description>
- </property>
-
- <property>
- <name>security.applicationhistory.protocol.acl</name>
- <value>*</value>
- <description>ACL for ApplicationHistoryProtocol, used by the timeline
- server and the generic history service client to communicate with each other.
- The ACL is a comma-separated list of user and group names. The user and
- group list is separated by a blank. For e.g. "alice,bob users,wheel".
- A special value of "*" means all users are allowed.</description>
- </property>
-</configuration>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/1cb29096/contrib/hawq-docker/centos6-docker/hawq-test/conf/hdfs-site.xml
----------------------------------------------------------------------
diff --git a/contrib/hawq-docker/centos6-docker/hawq-test/conf/hdfs-site.xml b/contrib/hawq-docker/centos6-docker/hawq-test/conf/hdfs-site.xml
deleted file mode 100644
index f565658..0000000
--- a/contrib/hawq-docker/centos6-docker/hawq-test/conf/hdfs-site.xml
+++ /dev/null
@@ -1,100 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
- <property>
- <name>dfs.name.dir</name>
- <value>/tmp/hdfs/name</value>
- <final>true</final>
- </property>
-
- <property>
- <name>dfs.data.dir</name>
- <value>/tmp/hdfs/data</value>
- <final>true</final>
- </property>
-
- <property>
- <name>dfs.permissions</name>
- <value>true</value>
- </property>
-
- <property>
- <name>dfs.support.append</name>
- <value>true</value>
- </property>
-
- <property>
- <name>dfs.block.local-path-access.user</name>
- <value>${user.name}</value>
- </property>
-
- <property>
- <name>dfs.replication</name>
- <value>3</value>
- </property>
-
- <property>
- <name>dfs.datanode.socket.write.timeout</name>
- <value>0</value>
- <description>
- used for sockets to and from datanodes. It is 8 minutes by default. Some
- users set this to 0, effectively disabling the write timeout.
- </description>
- </property>
-
- <property>
- <name>dfs.webhdfs.enabled</name>
- <value>true</value>
- </property>
-
- <property>
- <name>dfs.allow.truncate</name>
- <value>true</value>
- </property>
-
- <property>
- <name>dfs.namenode.fs-limits.min-block-size</name>
- <value>1024</value>
- </property>
-
- <property>
- <name>dfs.client.read.shortcircuit</name>
- <value>true</value>
- </property>
-
- <property>
- <name>dfs.domain.socket.path</name>
- <value>/var/lib/hadoop-hdfs/dn_socket</value>
- </property>
-
- <property>
- <name>dfs.block.access.token.enable</name>
- <value>true</value>
- <description>
- If "true", access tokens are used as capabilities for accessing
- datanodes.
- If "false", no access tokens are checked on accessing datanodes.
- </description>
- </property>
-
- <property>
- <name>dfs.namenode.datanode.registration.ip-hostname-check</name>
- <value>false</value>
- </property>
-</configuration>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/1cb29096/contrib/hawq-docker/centos6-docker/hawq-test/conf/kms-acls.xml
----------------------------------------------------------------------
diff --git a/contrib/hawq-docker/centos6-docker/hawq-test/conf/kms-acls.xml b/contrib/hawq-docker/centos6-docker/hawq-test/conf/kms-acls.xml
deleted file mode 100644
index cba69f4..0000000
--- a/contrib/hawq-docker/centos6-docker/hawq-test/conf/kms-acls.xml
+++ /dev/null
@@ -1,135 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<!--
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License.
--->
-<configuration>
-
- <!-- This file is hot-reloaded when it changes -->
-
- <!-- KMS ACLs -->
-
- <property>
- <name>hadoop.kms.acl.CREATE</name>
- <value>*</value>
- <description>
- ACL for create-key operations.
- If the user is not in the GET ACL, the key material is not returned
- as part of the response.
- </description>
- </property>
-
- <property>
- <name>hadoop.kms.acl.DELETE</name>
- <value>*</value>
- <description>
- ACL for delete-key operations.
- </description>
- </property>
-
- <property>
- <name>hadoop.kms.acl.ROLLOVER</name>
- <value>*</value>
- <description>
- ACL for rollover-key operations.
- If the user is not in the GET ACL, the key material is not returned
- as part of the response.
- </description>
- </property>
-
- <property>
- <name>hadoop.kms.acl.GET</name>
- <value>*</value>
- <description>
- ACL for get-key-version and get-current-key operations.
- </description>
- </property>
-
- <property>
- <name>hadoop.kms.acl.GET_KEYS</name>
- <value>*</value>
- <description>
- ACL for get-keys operations.
- </description>
- </property>
-
- <property>
- <name>hadoop.kms.acl.GET_METADATA</name>
- <value>*</value>
- <description>
- ACL for get-key-metadata and get-keys-metadata operations.
- </description>
- </property>
-
- <property>
- <name>hadoop.kms.acl.SET_KEY_MATERIAL</name>
- <value>*</value>
- <description>
- Complementary ACL for CREATE and ROLLOVER operations to allow the client
- to provide the key material when creating or rolling a key.
- </description>
- </property>
-
- <property>
- <name>hadoop.kms.acl.GENERATE_EEK</name>
- <value>*</value>
- <description>
- ACL for generateEncryptedKey CryptoExtension operations.
- </description>
- </property>
-
- <property>
- <name>hadoop.kms.acl.DECRYPT_EEK</name>
- <value>*</value>
- <description>
- ACL for decryptEncryptedKey CryptoExtension operations.
- </description>
- </property>
-
- <property>
- <name>default.key.acl.MANAGEMENT</name>
- <value>*</value>
- <description>
- default ACL for MANAGEMENT operations for all key acls that are not
- explicitly defined.
- </description>
- </property>
-
- <property>
- <name>default.key.acl.GENERATE_EEK</name>
- <value>*</value>
- <description>
- default ACL for GENERATE_EEK operations for all key acls that are not
- explicitly defined.
- </description>
- </property>
-
- <property>
- <name>default.key.acl.DECRYPT_EEK</name>
- <value>*</value>
- <description>
- default ACL for DECRYPT_EEK operations for all key acls that are not
- explicitly defined.
- </description>
- </property>
-
- <property>
- <name>default.key.acl.READ</name>
- <value>*</value>
- <description>
- default ACL for READ operations for all key acls that are not
- explicitly defined.
- </description>
- </property>
-
-
-</configuration>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/1cb29096/contrib/hawq-docker/centos6-docker/hawq-test/conf/kms-env.sh
----------------------------------------------------------------------
diff --git a/contrib/hawq-docker/centos6-docker/hawq-test/conf/kms-env.sh b/contrib/hawq-docker/centos6-docker/hawq-test/conf/kms-env.sh
deleted file mode 100644
index 44dfe6a..0000000
--- a/contrib/hawq-docker/centos6-docker/hawq-test/conf/kms-env.sh
+++ /dev/null
@@ -1,55 +0,0 @@
-#!/bin/bash
-#
-# Licensed under the Apache License, Version 2.0 (the "License");
-# you may not use this file except in compliance with the License.
-# You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License. See accompanying LICENSE file.
-#
-
-# Set kms specific environment variables here.
-
-# Settings for the Embedded Tomcat that runs KMS
-# Java System properties for KMS should be specified in this variable
-#
-# export CATALINA_OPTS=
-
-# KMS logs directory
-#
-# export KMS_LOG=${KMS_HOME}/logs
-
-# KMS temporary directory
-#
-# export KMS_TEMP=${KMS_HOME}/temp
-
-# The HTTP port used by KMS
-#
-# export KMS_HTTP_PORT=16000
-
-# The Admin port used by KMS
-#
-# export KMS_ADMIN_PORT=`expr ${KMS_HTTP_PORT} + 1`
-
-# The maximum number of Tomcat handler threads
-#
-# export KMS_MAX_THREADS=1000
-
-# The location of the SSL keystore if using SSL
-#
-# export KMS_SSL_KEYSTORE_FILE=${HOME}/.keystore
-
-# The password of the SSL keystore if using SSL
-#
-# export KMS_SSL_KEYSTORE_PASS=password
-
-# The full path to any native libraries that need to be loaded
-# (For eg. location of natively compiled tomcat Apache portable
-# runtime (APR) libraries
-#
-# export JAVA_LIBRARY_PATH=${HOME}/lib/native
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/1cb29096/contrib/hawq-docker/centos6-docker/hawq-test/conf/kms-log4j.properties
----------------------------------------------------------------------
diff --git a/contrib/hawq-docker/centos6-docker/hawq-test/conf/kms-log4j.properties b/contrib/hawq-docker/centos6-docker/hawq-test/conf/kms-log4j.properties
deleted file mode 100644
index 8e6d909..0000000
--- a/contrib/hawq-docker/centos6-docker/hawq-test/conf/kms-log4j.properties
+++ /dev/null
@@ -1,38 +0,0 @@
-#
-# Licensed under the Apache License, Version 2.0 (the "License");
-# you may not use this file except in compliance with the License.
-# You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License. See accompanying LICENSE file.
-#
-
-# If the Java System property 'kms.log.dir' is not defined at KMS start up time
-# Setup sets its value to '${kms.home}/logs'
-
-log4j.appender.kms=org.apache.log4j.DailyRollingFileAppender
-log4j.appender.kms.DatePattern='.'yyyy-MM-dd
-log4j.appender.kms.File=${kms.log.dir}/kms.log
-log4j.appender.kms.Append=true
-log4j.appender.kms.layout=org.apache.log4j.PatternLayout
-log4j.appender.kms.layout.ConversionPattern=%d{ISO8601} %-5p %c{1} - %m%n
-
-log4j.appender.kms-audit=org.apache.log4j.DailyRollingFileAppender
-log4j.appender.kms-audit.DatePattern='.'yyyy-MM-dd
-log4j.appender.kms-audit.File=${kms.log.dir}/kms-audit.log
-log4j.appender.kms-audit.Append=true
-log4j.appender.kms-audit.layout=org.apache.log4j.PatternLayout
-log4j.appender.kms-audit.layout.ConversionPattern=%d{ISO8601} %m%n
-
-log4j.logger.kms-audit=INFO, kms-audit
-log4j.additivity.kms-audit=false
-
-log4j.rootLogger=ALL, kms
-log4j.logger.org.apache.hadoop.conf=ERROR
-log4j.logger.org.apache.hadoop=INFO
-log4j.logger.com.sun.jersey.server.wadl.generators.WadlGeneratorJAXBGrammarGenerator=OFF
\ No newline at end of file
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/1cb29096/contrib/hawq-docker/centos6-docker/hawq-test/conf/kms-site.xml
----------------------------------------------------------------------
diff --git a/contrib/hawq-docker/centos6-docker/hawq-test/conf/kms-site.xml b/contrib/hawq-docker/centos6-docker/hawq-test/conf/kms-site.xml
deleted file mode 100644
index a810ca4..0000000
--- a/contrib/hawq-docker/centos6-docker/hawq-test/conf/kms-site.xml
+++ /dev/null
@@ -1,173 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<!--
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License.
--->
-<configuration>
-
- <!-- KMS Backend KeyProvider -->
-
- <property>
- <name>hadoop.kms.key.provider.uri</name>
- <value>jceks://file@/${user.home}/kms.keystore</value>
- <description>
- URI of the backing KeyProvider for the KMS.
- </description>
- </property>
-
- <property>
- <name>hadoop.security.keystore.JavaKeyStoreProvider.password</name>
- <value>none</value>
- <description>
- If using the JavaKeyStoreProvider, the password for the keystore file.
- </description>
- </property>
-
- <!-- KMS Cache -->
-
- <property>
- <name>hadoop.kms.cache.enable</name>
- <value>true</value>
- <description>
- Whether the KMS will act as a cache for the backing KeyProvider.
- When the cache is enabled, operations like getKeyVersion, getMetadata,
- and getCurrentKey will sometimes return cached data without consulting
- the backing KeyProvider. Cached values are flushed when keys are deleted
- or modified.
- </description>
- </property>
-
- <property>
- <name>hadoop.kms.cache.timeout.ms</name>
- <value>600000</value>
- <description>
- Expiry time for the KMS key version and key metadata cache, in
- milliseconds. This affects getKeyVersion and getMetadata.
- </description>
- </property>
-
- <property>
- <name>hadoop.kms.current.key.cache.timeout.ms</name>
- <value>30000</value>
- <description>
- Expiry time for the KMS current key cache, in milliseconds. This
- affects getCurrentKey operations.
- </description>
- </property>
-
- <!-- KMS Audit -->
-
- <property>
- <name>hadoop.kms.audit.aggregation.window.ms</name>
- <value>10000</value>
- <description>
- Duplicate audit log events within the aggregation window (specified in
- ms) are quashed to reduce log traffic. A single message for aggregated
- events is printed at the end of the window, along with a count of the
- number of aggregated events.
- </description>
- </property>
-
- <!-- KMS Security -->
-
- <property>
- <name>hadoop.kms.authentication.type</name>
- <value>simple</value>
- <description>
- Authentication type for the KMS. Can be either "simple"
- or "kerberos".
- </description>
- </property>
-
- <property>
- <name>hadoop.kms.authentication.kerberos.keytab</name>
- <value>${user.home}/kms.keytab</value>
- <description>
- Path to the keytab with credentials for the configured Kerberos principal.
- </description>
- </property>
-
- <property>
- <name>hadoop.kms.authentication.kerberos.principal</name>
- <value>HTTP/localhost</value>
- <description>
- The Kerberos principal to use for the HTTP endpoint.
- The principal must start with 'HTTP/' as per the Kerberos HTTP SPNEGO specification.
- </description>
- </property>
-
- <property>
- <name>hadoop.kms.authentication.kerberos.name.rules</name>
- <value>DEFAULT</value>
- <description>
- Rules used to resolve Kerberos principal names.
- </description>
- </property>
-
- <!-- Authentication cookie signature source -->
-
- <property>
- <name>hadoop.kms.authentication.signer.secret.provider</name>
- <value>random</value>
- <description>
- Indicates how the secret to sign the authentication cookies will be
- stored. Options are 'random' (default), 'string' and 'zookeeper'.
- If using a setup with multiple KMS instances, 'zookeeper' should be used.
- </description>
- </property>
-
- <!-- Configuration for 'zookeeper' authentication cookie signature source -->
-
- <property>
- <name>hadoop.kms.authentication.signer.secret.provider.zookeeper.path</name>
- <value>/hadoop-kms/hadoop-auth-signature-secret</value>
- <description>
- The Zookeeper ZNode path where the KMS instances will store and retrieve
- the secret from.
- </description>
- </property>
-
- <property>
- <name>hadoop.kms.authentication.signer.secret.provider.zookeeper.connection.string</name>
- <value>#HOSTNAME#:#PORT#,...</value>
- <description>
- The Zookeeper connection string, a list of hostnames and port comma
- separated.
- </description>
- </property>
-
- <property>
- <name>hadoop.kms.authentication.signer.secret.provider.zookeeper.auth.type</name>
- <value>kerberos</value>
- <description>
- The Zookeeper authentication type, 'none' or 'sasl' (Kerberos).
- </description>
- </property>
-
- <property>
- <name>hadoop.kms.authentication.signer.secret.provider.zookeeper.kerberos.keytab</name>
- <value>/etc/hadoop/conf/kms.keytab</value>
- <description>
- The absolute path for the Kerberos keytab with the credentials to
- connect to Zookeeper.
- </description>
- </property>
-
- <property>
- <name>hadoop.kms.authentication.signer.secret.provider.zookeeper.kerberos.principal</name>
- <value>kms/#HOSTNAME#</value>
- <description>
- The Kerberos service principal used to connect to Zookeeper.
- </description>
- </property>
-
-</configuration>