You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by dm...@apache.org on 2015/01/12 12:53:08 UTC
[13/21] ambari git commit: AMBARI-9068. Remove HDP 1.3 stack defn
from Ambari. (dlysnichenko)
http://git-wip-us.apache.org/repos/asf/ambari/blob/23b7c110/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapred-env.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapred-env.xml b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapred-env.xml
deleted file mode 100644
index e29aab4..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapred-env.xml
+++ /dev/null
@@ -1,81 +0,0 @@
-<?xml version="1.0"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
--->
-
-<configuration>
- <property>
- <name>jtnode_opt_newsize</name>
- <value>200</value>
- <description>Mem New Size.</description>
- </property>
- <property>
- <name>jtnode_opt_maxnewsize</name>
- <value>200</value>
- <description>Max New size.</description>
- </property>
- <property>
- <name>jtnode_heapsize</name>
- <value>1024</value>
- <description>Maximum Java heap size for JobTracker in MB (Java option -Xmx)</description>
- </property>
- <property>
- <name>lzo_enabled</name>
- <value>true</value>
- <description>LZO compression enabled</description>
- </property>
- <property>
- <name>snappy_enabled</name>
- <value>true</value>
- <description>LZO compression enabled</description>
- </property>
- <property>
- <name>rca_enabled</name>
- <value>true</value>
- <description>Enable Job Diagnostics.</description>
- </property>
- <property>
- <name>mapred_user</name>
- <value>mapred</value>
- <property-type>USER</property-type>
- <description>MapReduce User.</description>
- </property>
- <property>
- <name>rca_properties</name>
- <value>
-ambari.jobhistory.database={ambari_db_rca_url}
-ambari.jobhistory.driver={ambari_db_rca_driver}
-ambari.jobhistory.user={ambari_db_rca_username}
-ambari.jobhistory.password={ambari_db_rca_password}
-ambari.jobhistory.logger=${{hadoop.root.logger}}
-
-log4j.appender.JHA=org.apache.ambari.log4j.hadoop.mapreduce.jobhistory.JobHistoryAppender
-log4j.appender.JHA.database={ambari_db_rca_url}
-log4j.appender.JHA.driver={ambari_db_rca_driver}
-log4j.appender.JHA.user={ambari_db_rca_username}
-log4j.appender.JHA.password={ambari_db_rca_password}
-
-log4j.logger.org.apache.hadoop.mapred.JobHistory$JobHistoryLogger=DEBUG,JHA
-log4j.additivity.org.apache.hadoop.mapred.JobHistory$JobHistoryLogger=true
-
- </value>
- </property>
-
-</configuration>
http://git-wip-us.apache.org/repos/asf/ambari/blob/23b7c110/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapred-queue-acls.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapred-queue-acls.xml b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapred-queue-acls.xml
deleted file mode 100644
index f1f0ec5..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapred-queue-acls.xml
+++ /dev/null
@@ -1,52 +0,0 @@
-<?xml version="1.0"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-
-<!--
- Licensed to the Apache Software Foundation (ASF) under one or more
- contributor license agreements. See the NOTICE file distributed with
- this work for additional information regarding copyright ownership.
- The ASF licenses this file to You under the Apache License, Version 2.0
- (the "License"); you may not use this file except in compliance with
- the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License.
--->
-
-<!-- mapred-queue-acls.xml -->
-<configuration supports_final="true">
-
-
-<!-- queue default -->
-
- <property>
- <name>mapred.queue.default.acl-submit-job</name>
- <value>*</value>
- <description> Comma separated list of user and group names that are allowed
- to submit jobs to the 'default' queue. The user list and the group list
- are separated by a blank. For e.g. alice,bob group1,group2.
- If set to the special value '*', it means all users are allowed to
- submit jobs.
- </description>
- </property>
-
- <property>
- <name>mapred.queue.default.acl-administer-jobs</name>
- <value>*</value>
- <description> Comma separated list of user and group names that are allowed
- to delete jobs or modify job's priority for jobs not owned by the current
- user in the 'default' queue. The user list and the group list
- are separated by a blank. For e.g. alice,bob group1,group2.
- If set to the special value '*', it means all users are allowed to do
- this operation.
- </description>
- </property>
-
- <!-- END ACLs -->
-
-</configuration>
http://git-wip-us.apache.org/repos/asf/ambari/blob/23b7c110/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapred-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapred-site.xml b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapred-site.xml
deleted file mode 100644
index 7cc0dbf..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapred-site.xml
+++ /dev/null
@@ -1,554 +0,0 @@
-<?xml version="1.0"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-
-<!--
- Licensed to the Apache Software Foundation (ASF) under one or more
- contributor license agreements. See the NOTICE file distributed with
- this work for additional information regarding copyright ownership.
- The ASF licenses this file to You under the Apache License, Version 2.0
- (the "License"); you may not use this file except in compliance with
- the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration supports_final="true" xmlns:xi="http://www.w3.org/2001/XInclude">
-
- <!-- i/o properties -->
-
- <property>
- <name>io.sort.mb</name>
- <value>200</value>
- <description>
- The total amount of Map-side buffer memory to use while sorting files
- </description>
- </property>
-
- <property>
- <name>io.sort.record.percent</name>
- <value>.2</value>
- <description>The percentage of io.sort.mb dedicated to tracking record boundaries. Let this value be r, io.sort.mb be x.
- The maximum number of records collected before the collection thread must block is equal to (r * x) / 4
- </description>
- </property>
-
- <property>
- <name>io.sort.spill.percent</name>
- <value>0.9</value>
- <description>Percentage of sort buffer used for record collection</description>
- </property>
-
- <property>
- <name>io.sort.factor</name>
- <value>100</value>
- <description>The number of streams to merge at once while sorting files. This determines the number of open file handles.
- </description>
- </property>
-
- <!-- map/reduce properties -->
-
- <property>
- <name>mapred.tasktracker.tasks.sleeptime-before-sigkill</name>
- <value>250</value>
- <description>Normally, this is the amount of time before killing
- processes, and the recommended-default is 5.000 seconds - a value of
- 5000 here. In this case, we are using it solely to blast tasks before
- killing them, and killing them very quickly (1/4 second) to guarantee
- that we do not leave VMs around for later jobs.
- </description>
- </property>
-
- <property>
- <name>mapred.job.tracker.handler.count</name>
- <value>50</value>
- <description>
- The number of server threads for the JobTracker. This should be roughly
- 4% of the number of tasktracker nodes.
- </description>
- </property>
-
- <property>
- <name>mapred.system.dir</name>
- <value>/mapred/system</value>
- <description>Path on the HDFS where where the MapReduce framework stores system files</description>
- <final>true</final>
- </property>
-
- <property>
- <name>mapred.job.tracker</name>
- <!-- cluster variant -->
- <value>localhost:50300</value>
- <description>JobTracker address</description>
- <final>true</final>
- </property>
-
- <property>
- <name>mapred.job.tracker.http.address</name>
- <!-- cluster variant -->
- <value>localhost:50030</value>
- <description>JobTracker host and http port address</description>
- <final>true</final>
- </property>
-
- <property>
- <!-- cluster specific -->
- <name>mapred.local.dir</name>
- <value>/hadoop/mapred</value>
- <description>The local directory where MapReduce stores intermediate data files. May be a comma-separated list of
- directories on different devices in order to spread disk i/o. Directories that do not exist are ignored.
- </description>
- <final>true</final>
- </property>
-
- <property>
- <name>mapreduce.cluster.administrators</name>
- <value> hadoop</value>
- <description>Cluster administrators. Irrespective of the job ACLs configured, cluster administrators always have
- access to view and modify a job.
- </description>
- </property>
-
- <property>
- <name>mapred.reduce.parallel.copies</name>
- <value>30</value>
- <description>The default number of parallel transfers run by reduce
- during the copy(shuffle) phase.
- </description>
- </property>
-
- <property>
- <name>mapred.tasktracker.map.tasks.maximum</name>
- <value>4</value>
- <description>The maximum number of map tasks that will be run simultaneously by a task tracker.</description>
- </property>
-
- <property>
- <name>mapred.tasktracker.reduce.tasks.maximum</name>
- <value>2</value>
- <description>The maximum number of reduce tasks that will be run simultaneously by a task tracker.</description>
- </property>
-
- <property>
- <name>tasktracker.http.threads</name>
- <value>50</value>
- <description>The number of worker threads that for the http server. This is used for map output fetching.</description>
- </property>
-
- <property>
- <name>mapred.map.tasks.speculative.execution</name>
- <value>false</value>
- <description>If true, then multiple instances of some map tasks
- may be executed in parallel.</description>
- </property>
-
- <property>
- <name>mapred.reduce.tasks.speculative.execution</name>
- <value>false</value>
- <description>If true, then multiple instances of some reduce tasks
- may be executed in parallel.</description>
- </property>
-
- <property>
- <name>mapred.reduce.slowstart.completed.maps</name>
- <value>0.05</value>
- <description>Fraction of the number of maps in the job which should be complete before reduces are scheduled for the job.</description>
- </property>
-
- <property>
- <name>mapred.inmem.merge.threshold</name>
- <value>1000</value>
- <description>The threshold, in terms of the number of files
- for the in-memory merge process. When we accumulate threshold number of files
- we initiate the in-memory merge and spill to disk. A value of 0 or less than
- 0 indicates we want to DON'T have any threshold and instead depend only on
- the ramfs's memory consumption to trigger the merge.
- </description>
- </property>
-
- <property>
- <name>mapred.job.shuffle.merge.percent</name>
- <value>0.66</value>
- <description>The usage threshold at which an in-memory merge will be
- initiated, expressed as a percentage of the total memory allocated to
- storing in-memory map outputs, as defined by
- mapred.job.shuffle.input.buffer.percent.
- </description>
- </property>
-
- <property>
- <name>mapred.job.shuffle.input.buffer.percent</name>
- <value>0.7</value>
- <description>The percentage of memory to be allocated from the maximum heap
- size to storing map outputs during the shuffle.
- </description>
- </property>
-
- <property>
- <name>mapred.map.output.compression.codec</name>
- <value>org.apache.hadoop.io.compress.SnappyCodec</value>
- <description>If the map outputs are compressed, how should they be
- compressed
- </description>
- </property>
-
- <property>
- <name>mapred.output.compression.type</name>
- <value>BLOCK</value>
- <description>If the job outputs are to compressed as SequenceFiles, how should
- they be compressed? Should be one of NONE, RECORD or BLOCK.
- </description>
- </property>
-
-
- <property>
- <name>mapred.jobtracker.completeuserjobs.maximum</name>
- <value>0</value>
- <description>The maximum number of complete jobs per user to keep around before delegating them to the job history.</description>
- </property>
-
- <property>
- <name>mapred.jobtracker.taskScheduler</name>
- <value>org.apache.hadoop.mapred.CapacityTaskScheduler</value>
- <description>The class responsible for scheduling the tasks.</description>
- </property>
-
- <property>
- <name>mapred.jobtracker.restart.recover</name>
- <value>false</value>
- <description>"true" to enable (job) recovery upon restart,
- "false" to start afresh
- </description>
- </property>
-
- <property>
- <name>mapred.job.reduce.input.buffer.percent</name>
- <value>0.0</value>
- <description>The percentage of memory- relative to the maximum heap size- to
- retain map outputs during the reduce. When the shuffle is concluded, any
- remaining map outputs in memory must consume less than this threshold before
- the reduce can begin.
- </description>
- </property>
-
- <property>
- <name>mapreduce.reduce.input.limit</name>
- <value>10737418240</value>
- <description>The limit on the input size of the reduce. (This value
- is 10 Gb.) If the estimated input size of the reduce is greater than
- this value, job is failed. A value of -1 means that there is no limit
- set. </description>
- </property>
-
-
- <!-- copied from kryptonite configuration -->
-
- <property>
- <name>mapred.task.timeout</name>
- <value>600000</value>
- <description>The number of milliseconds before a task will be
- terminated if it neither reads an input, writes an output, nor
- updates its status string.
- </description>
- </property>
-
- <property>
- <name>jetty.connector</name>
- <value>org.mortbay.jetty.nio.SelectChannelConnector</value>
- <description>The connector to be used by Jetty server.</description>
- </property>
-
- <property>
- <name>mapred.task.tracker.task-controller</name>
- <value>org.apache.hadoop.mapred.DefaultTaskController</value>
- <description>
- TaskController which is used to launch and manage task execution.
- </description>
- </property>
-
- <property>
- <name>mapred.child.root.logger</name>
- <value>INFO,TLA</value>
- <description>Logger configuration for the TaskTracker child processes</description>
- </property>
-
- <property>
- <name>mapred.child.java.opts</name>
- <value>-server -Xmx768m -Djava.net.preferIPv4Stack=true</value>
- <description>Java options for the TaskTracker child processes</description>
- </property>
-
- <property>
- <name>mapred.cluster.map.memory.mb</name>
- <value>1536</value>
- <description>
- The virtual memory size of a single Map slot in the MapReduce framework
- </description>
- </property>
-
- <property>
- <name>mapred.cluster.reduce.memory.mb</name>
- <value>2048</value>
- <description>
- The virtual memory size of a single Reduce slot in the MapReduce framework
- </description>
- </property>
-
- <property>
- <name>mapred.job.map.memory.mb</name>
- <value>1536</value>
- <description>
- Virtual memory for single Map task
- </description>
- </property>
-
- <property>
- <name>mapred.job.reduce.memory.mb</name>
- <value>2048</value>
- <description>
- Virtual memory for single Reduce task
- </description>
- </property>
-
- <property>
- <name>mapred.cluster.max.map.memory.mb</name>
- <value>6144</value>
- <description>
- Upper limit on virtual memory size for a single Map task of any MapReduce job
- </description>
- </property>
-
- <property>
- <name>mapred.cluster.max.reduce.memory.mb</name>
- <value>4096</value>
- <description>
- Upper limit on virtual memory size for a single Reduce task of any MapReduce job
- </description>
- </property>
-
- <property>
- <name>mapred.hosts</name>
- <value>/etc/hadoop/conf/mapred.include</value>
- <description>
- Names a file that contains the list of nodes that may
- connect to the jobtracker. If the value is empty, all hosts are
- permitted.
- </description>
- </property>
-
- <property>
- <name>mapred.hosts.exclude</name>
- <value>/etc/hadoop/conf/mapred.exclude</value>
- <description>
- Names a file that contains the list of hosts that
- should be excluded by the jobtracker. If the value is empty, no
- hosts are excluded.
- </description>
- </property>
-
- <property>
- <name>mapred.max.tracker.blacklists</name>
- <value>16</value>
- <description>
- if node is reported blacklisted by 16 successful jobs within timeout-window, it will be graylisted
- </description>
- </property>
-
- <property>
- <name>mapred.healthChecker.script.path</name>
- <value>/etc/hadoop/conf/health_check</value>
- <description>
- Directory path to view job status
- </description>
- </property>
-
- <property>
- <name>mapred.healthChecker.interval</name>
- <value>135000</value>
- <description>Frequency of the node health script to be run, in milliseconds</description>
- </property>
-
- <property>
- <name>mapred.healthChecker.script.timeout</name>
- <value>60000</value>
- <description>Time after node health script should be killed if unresponsive and considered that the script has failed.</description>
- </property>
-
- <property>
- <name>mapred.job.tracker.persist.jobstatus.active</name>
- <value>false</value>
- <description>Indicates if persistency of job status information is
- active or not.
- </description>
- </property>
-
- <property>
- <name>mapred.job.tracker.persist.jobstatus.hours</name>
- <value>1</value>
- <description>The number of hours job status information is persisted in DFS.
- The job status information will be available after it drops of the memory
- queue and between jobtracker restarts. With a zero value the job status
- information is not persisted at all in DFS.
- </description>
- </property>
-
- <property>
- <name>mapred.job.tracker.persist.jobstatus.dir</name>
- <value>/mapred/jobstatus</value>
- <description>The directory where the job status information is persisted
- in a file system to be available after it drops of the memory queue and
- between jobtracker restarts.
- </description>
- </property>
-
- <property>
- <name>mapred.jobtracker.retirejob.check</name>
- <value>10000</value>
- <description>Interval for the check for jobs to be retired. </description>
- </property>
-
- <property>
- <name>mapred.jobtracker.retirejob.interval</name>
- <value>21600000</value>
- <description> Completed Job retirement interval. </description>
- </property>
-
- <property>
- <name>mapred.job.tracker.history.completed.location</name>
- <value>/mapred/history/done</value>
- <description>The completed job history files are stored at this single well known location.</description>
- </property>
-
- <property>
- <name>mapred.jobtracker.maxtasks.per.job</name>
- <value>-1</value>
- <final>true</final>
- <description>The maximum number of tasks for a single job.
- A value of -1 indicates that there is no maximum. </description>
- </property>
-
- <property>
- <name>mapreduce.fileoutputcommitter.marksuccessfuljobs</name>
- <value>false</value>
- <description>Enable this flag to create _SUCCESS file for successful job. </description>
- </property>
-
- <property>
- <name>mapred.userlog.retain.hours</name>
- <value>24</value>
- <description>
- The maximum time, in hours, for which the user-logs are to be retained after the job completion.
- </description>
- </property>
-
- <property>
- <name>mapred.job.reuse.jvm.num.tasks</name>
- <value>1</value>
- <description>
- How many tasks to run per jvm. If set to -1, there is no limit
- </description>
- <final>true</final>
- </property>
-
- <property>
- <name>hadoop.job.history.user.location</name>
- <value>none</value>
- <description> Location to store the history files of a particular job. If set to none, then the job histories are
- not collected to anywhere outside the master node.
- </description>
- <final>true</final>
- </property>
-
-
-
- <property>
- <name>mapreduce.jobtracker.staging.root.dir</name>
- <value>/user</value>
- <description>The Path prefix for where the staging directories should be placed. The next level is always the user's
- name. It is a path in the default file system.</description>
- </property>
-
- <property>
- <name>mapreduce.tasktracker.group</name>
- <value>hadoop</value>
- <description>The group that the task controller uses for accessing the task controller. The mapred user must be a member and users should *not* be members.</description>
-
- </property>
-
- <property>
- <name>mapreduce.jobtracker.split.metainfo.maxsize</name>
- <value>50000000</value>
- <final>true</final>
- <description>If the size of the split metainfo file is larger than this, the JobTracker will fail the job during
- initialize.
- </description>
- </property>
- <property>
- <name>mapreduce.history.server.embedded</name>
- <value>false</value>
- <description>Should job history server be embedded within Job tracker
- process</description>
- <final>true</final>
- </property>
-
- <property>
- <name>mapreduce.history.server.http.address</name>
- <!-- cluster variant -->
- <value>localhost:51111</value>
- <description>Http address of the history server</description>
- <final>true</final>
- </property>
-
-
- <property>
- <name>mapred.jobtracker.blacklist.fault-timeout-window</name>
- <value>180</value>
- <description>
- 3-hour sliding window (value is in minutes)
- </description>
- </property>
-
- <property>
- <name>mapred.jobtracker.blacklist.fault-bucket-width</name>
- <value>15</value>
- <description>
- 15-minute bucket size (value is in minutes)
- </description>
- </property>
-
- <property>
- <name>mapred.queue.names</name>
- <value>default</value>
- <description> Comma separated list of queues configured for this jobtracker.</description>
- </property>
-
- <property>
- <name>mapreduce.jobhistory.intermediate-done-dir</name>
- <value>/mr-history/tmp</value>
- <description>
- Directory where history files are written by MapReduce jobs.
- </description>
- </property>
-
- <property>
- <name>mapreduce.jobhistory.done-dir</name>
- <value>/mr-history/done</value>
- <description>
- Directory where history files are managed by the MR JobHistory Server.
- </description>
- </property>
-
- <property>
- <name>mapreduce.jobhistory.webapp.address</name>
- <value>localhost:19888</value>
- <description>Enter your JobHistoryServer hostname.</description>
- </property>
-
-</configuration>
http://git-wip-us.apache.org/repos/asf/ambari/blob/23b7c110/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapreduce-log4j.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapreduce-log4j.xml b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapreduce-log4j.xml
deleted file mode 100644
index b446400..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapreduce-log4j.xml
+++ /dev/null
@@ -1,54 +0,0 @@
-<?xml version="1.0"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
--->
-
-<configuration supports_final="false">
-
- <property>
- <name>content</name>
- <description>Custom log4j.properties</description>
- <value>
-#Relative to Mapred LOG DIR Prefix
-mapred.log.dir=.
-#
-# Job Summary Appender
-#
-# Use following logger to send summary to separate file defined by
-# hadoop.mapreduce.jobsummary.log.file rolled daily:
-# hadoop.mapreduce.jobsummary.logger=INFO,JSA
-#
-hadoop.mapreduce.jobsummary.logger=${hadoop.root.logger}
-hadoop.mapreduce.jobsummary.log.file=hadoop-mapreduce.jobsummary.log
-log4j.appender.JSA=org.apache.log4j.DailyRollingFileAppender
-
-log4j.appender.JSA.File=${mapred.log.dir}/${hadoop.mapreduce.jobsummary.log.file}
-
-log4j.appender.JSA.layout=org.apache.log4j.PatternLayout
-log4j.appender.JSA.layout.ConversionPattern=%d{yy/MM/dd HH:mm:ss} %p %c{2}: %m%n
-log4j.appender.JSA.DatePattern=.yyyy-MM-dd
-log4j.appender.JSA.layout=org.apache.log4j.PatternLayout
-log4j.logger.org.apache.hadoop.mapred.JobInProgress$JobSummary=${hadoop.mapreduce.jobsummary.logger}
-log4j.additivity.org.apache.hadoop.mapred.JobInProgress$JobSummary=false
-
- </value>
- </property>
-
-</configuration>
http://git-wip-us.apache.org/repos/asf/ambari/blob/23b7c110/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/metainfo.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/metainfo.xml b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/metainfo.xml
deleted file mode 100644
index 875b858..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/metainfo.xml
+++ /dev/null
@@ -1,140 +0,0 @@
-<?xml version="1.0"?>
-<!--
- Licensed to the Apache Software Foundation (ASF) under one or more
- contributor license agreements. See the NOTICE file distributed with
- this work for additional information regarding copyright ownership.
- The ASF licenses this file to You under the Apache License, Version 2.0
- (the "License"); you may not use this file except in compliance with
- the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License.
--->
-
-<metainfo>
- <schemaVersion>2.0</schemaVersion>
- <services>
- <service>
- <name>MAPREDUCE</name>
- <displayName>MapReduce</displayName>
- <comment>Apache Hadoop Distributed Processing Framework</comment>
- <version>1.2.0.1.3.3.0</version>
- <components>
- <component>
- <name>JOBTRACKER</name>
- <displayName>JobTracker</displayName>
- <category>MASTER</category>
- <cardinality>1</cardinality>
- <commandScript>
- <script>scripts/jobtracker.py</script>
- <scriptType>PYTHON</scriptType>
- <timeout>600</timeout>
- </commandScript>
- <customCommands>
- <customCommand>
- <name>DECOMMISSION</name>
- <commandScript>
- <script>scripts/jobtracker.py</script>
- <scriptType>PYTHON</scriptType>
- <timeout>600</timeout>
- </commandScript>
- </customCommand>
- </customCommands>
- </component>
-
- <component>
- <name>TASKTRACKER</name>
- <displayName>TaskTracker</displayName>
- <category>SLAVE</category>
- <cardinality>1+</cardinality>
- <commandScript>
- <script>scripts/tasktracker.py</script>
- <scriptType>PYTHON</scriptType>
- <timeout>600</timeout>
- </commandScript>
- </component>
-
- <component>
- <name>MAPREDUCE_CLIENT</name>
- <displayName>MapReduce Client</displayName>
- <category>CLIENT</category>
- <cardinality>1+</cardinality>
- <commandScript>
- <script>scripts/client.py</script>
- <scriptType>PYTHON</scriptType>
- <timeout>600</timeout>
- </commandScript>
- <configFiles>
- <configFile>
- <type>xml</type>
- <fileName>core-site.xml</fileName>
- <dictionaryName>core-site</dictionaryName>
- </configFile>
- <configFile>
- <type>xml</type>
- <fileName>mapred-site.xml</fileName>
- <dictionaryName>mapred-site</dictionaryName>
- </configFile>
- <configFile>
- <type>env</type>
- <fileName>log4j.properties</fileName>
- <dictionaryName>hdfs-log4j,mapreduce-log4j</dictionaryName>
- </configFile>
- </configFiles>
- </component>
-
- <component>
- <name>HISTORYSERVER</name>
- <displayName>History Server</displayName>
- <category>MASTER</category>
- <cardinality>1</cardinality>
- <auto-deploy>
- <enabled>true</enabled>
- <co-locate>MAPREDUCE/JOBTRACKER</co-locate>
- </auto-deploy>
- <commandScript>
- <script>scripts/historyserver.py</script>
- <scriptType>PYTHON</scriptType>
- <timeout>600</timeout>
- </commandScript>
- </component>
- </components>
-
- <osSpecifics>
- <osSpecific>
- <osFamily>any</osFamily>
- <packages>
- <package>
- <name>hadoop</name>
- </package>
- </packages>
- </osSpecific>
- </osSpecifics>
-
- <commandScript>
- <script>scripts/service_check.py</script>
- <scriptType>PYTHON</scriptType>
- <timeout>300</timeout>
- </commandScript>
-
- <requiredServices>
- <service>HDFS</service>
- </requiredServices>
-
- <configuration-dependencies>
- <config-type>capacity-scheduler</config-type>
- <config-type>core-site</config-type>
- <config-type>mapred-site</config-type>
- <config-type>mapred-env</config-type>
- <config-type>mapred-queue-acls</config-type>
- <config-type>mapreduce-log4j</config-type>
- </configuration-dependencies>
- </service>
-
- </services>
-</metainfo>
http://git-wip-us.apache.org/repos/asf/ambari/blob/23b7c110/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/alerts/alert_mapreduce_directory_space.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/alerts/alert_mapreduce_directory_space.py b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/alerts/alert_mapreduce_directory_space.py
deleted file mode 100644
index d6e5d75..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/alerts/alert_mapreduce_directory_space.py
+++ /dev/null
@@ -1,93 +0,0 @@
-#!/usr/bin/env python
-
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-"""
-
-import collections
-import os
-import platform
-
-RESULT_CODE_OK = 'OK'
-RESULT_CODE_CRITICAL = 'CRITICAL'
-RESULT_CODE_UNKNOWN = 'UNKNOWN'
-
-MAPREDUCE_LOCAL_DIR_KEY = '{{mapred-site/mapred.local.dir}}'
-
-def get_tokens():
- """
- Returns a tuple of tokens in the format {{site/property}} that will be used
- to build the dictionary passed into execute
- """
- return (MAPREDUCE_LOCAL_DIR_KEY,)
-
-
-def execute(parameters=None, host_name=None):
- """
- Returns a tuple containing the result code and a pre-formatted result label
-
- Keyword arguments:
- parameters (dictionary): a mapping of parameter key to value
- host_name (string): the name of this host where the alert is running
- """
- if parameters is None:
- return (('UNKNOWN', ['There were no parameters supplied to the script.']))
-
- if set([MAPREDUCE_LOCAL_DIR_KEY]).issubset(parameters):
- mapreduce_local_directories = parameters[MAPREDUCE_LOCAL_DIR_KEY]
- else:
- return (('UNKNOWN', ['The MapReduce Local Directory is required.']))
-
- directory_list = mapreduce_local_directories.split(",")
- for directory in directory_list:
- disk_usage = None
- try:
- disk_usage = _get_disk_usage(directory)
- except NotImplementedError, platform_error:
- return (RESULT_CODE_UNKNOWN, [str(platform_error)])
-
- if disk_usage is None or disk_usage.total == 0:
- return (RESULT_CODE_UNKNOWN, ['Unable to determine the disk usage.'])
-
- percent = disk_usage.used / float(disk_usage.total) * 100
-
- if percent > 85:
- message = 'The disk usage of {0} is {1:d}%'.format(directory,percent)
- return (RESULT_CODE_CRITICAL, [message])
-
- return (RESULT_CODE_OK, ["All MapReduce local directories have sufficient space."])
-
-
-def _get_disk_usage(path):
- """
- returns a named tuple that contains the total, used, and free disk space
- in bytes
- """
- used = 0
- total = 0
- free = 0
-
- if 'statvfs' in dir(os):
- disk_stats = os.statvfs(path)
- free = disk_stats.f_bavail * disk_stats.f_frsize
- total = disk_stats.f_blocks * disk_stats.f_frsize
- used = (disk_stats.f_blocks - disk_stats.f_bfree) * disk_stats.f_frsize
- else:
- raise NotImplementedError("{0} is not a supported platform for this alert".format(platform.platform()))
-
- DiskInfo = collections.namedtuple('DiskInfo', 'total used free')
- return DiskInfo(total=total, used=used, free=free)
http://git-wip-us.apache.org/repos/asf/ambari/blob/23b7c110/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/client.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/client.py b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/client.py
deleted file mode 100644
index 4ba1031..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/client.py
+++ /dev/null
@@ -1,49 +0,0 @@
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-Ambari Agent
-
-"""
-import sys
-from resource_management import *
-
-from mapreduce import mapreduce
-from service import service
-
-class Client(Script):
-
- def install(self, env):
- self.install_packages(env)
- self.configure(env)
-
- def configure(self, env):
- import params
- env.set_params(params)
- mapreduce()
-
- def status(self, env):
- raise ClientComponentHasNoStatus()
-
- def generate_configs_get_template_file_content(self, filename, dicts):
- import params
- content = super(Client,self).generate_configs_get_template_file_content(filename, dicts)
- if filename == 'log4j.properties':
- content += params.rca_properties
- return content
-
-if __name__ == "__main__":
- Client().execute()
http://git-wip-us.apache.org/repos/asf/ambari/blob/23b7c110/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/historyserver.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/historyserver.py b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/historyserver.py
deleted file mode 100644
index bccd0c6..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/historyserver.py
+++ /dev/null
@@ -1,58 +0,0 @@
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-Ambari Agent
-
-"""
-import sys
-from resource_management import *
-
-from mapreduce import mapreduce
-from service import service
-
-class Historyserver(Script):
- def install(self, env):
- self.install_packages(env)
-
- def configure(self, env):
- import params
- env.set_params(params)
- mapreduce(name="historyserver")
-
- def start(self, env):
- import params
- env.set_params(params)
- self.configure(env)
- service('historyserver',
- action='start'
- )
-
- def stop(self, env):
- import params
- env.set_params(params)
-
- service('historyserver',
- action='stop'
- )
-
- def status(self, env):
- import status_params
- env.set_params(status_params)
- check_process_status(status_params.historyserver_pid_file)
-
-if __name__ == "__main__":
- Historyserver().execute()
http://git-wip-us.apache.org/repos/asf/ambari/blob/23b7c110/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/jobtracker.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/jobtracker.py b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/jobtracker.py
deleted file mode 100644
index 2ca04fe..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/jobtracker.py
+++ /dev/null
@@ -1,87 +0,0 @@
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-Ambari Agent
-
-"""
-
-import sys
-from resource_management import *
-
-from mapreduce import mapreduce
-from service import service
-
-class Jobtracker(Script):
- def install(self, env):
- self.install_packages(env)
-
- def configure(self, env):
- import params
- env.set_params(params)
- mapreduce(name="jobtracker")
-
- def start(self, env):
- import params
- env.set_params(params)
- self.configure(env) # FOR SECURITY
- service('jobtracker',
- action='start'
- )
-
- def stop(self, env):
- import params
- env.set_params(params)
-
- service('jobtracker',
- action='stop'
- )
-
- def status(self, env):
- import status_params
- env.set_params(status_params)
- check_process_status(status_params.jobtracker_pid_file)
- pass
-
- def decommission(self, env):
- import params
-
- env.set_params(params)
-
- mapred_user = params.mapred_user
- conf_dir = params.conf_dir
- user_group = params.user_group
-
- if params.security_enabled:
- Execute(params.kinit_cmd,
- user = params.mapred_user)
-
- File(params.exclude_file_path,
- content=Template("exclude_hosts_list.j2"),
- owner=mapred_user,
- group=user_group
- )
-
- if params.update_exclude_file_only == False:
- ExecuteHadoop('mradmin -refreshNodes',
- user=mapred_user,
- conf_dir=conf_dir,
- kinit_override=True)
- pass
- pass
-
-if __name__ == "__main__":
- Jobtracker().execute()
http://git-wip-us.apache.org/repos/asf/ambari/blob/23b7c110/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/mapreduce.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/mapreduce.py b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/mapreduce.py
deleted file mode 100644
index 2ab22b8..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/mapreduce.py
+++ /dev/null
@@ -1,178 +0,0 @@
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-Ambari Agent
-
-"""
-
-from resource_management import *
-import sys
-import os
-
-
-def mapreduce(name=None):
- import params
-
-
- if name in ["jobtracker","historyserver"]:
- params.HdfsDirectory("/mapred",
- action="create_delayed",
- owner=params.mapred_user
- )
- params.HdfsDirectory("/mapred/system",
- action="create_delayed",
- owner=params.mapred_user
- )
- params.HdfsDirectory("/mapred/history",
- action="create_delayed",
- owner=params.mapred_user
- )
- params.HdfsDirectory(params.mapreduce_jobhistory_intermediate_done_dir,
- action="create_delayed",
- owner=params.mapred_user,
- group=params.user_group,
- mode=0777
- )
- params.HdfsDirectory(params.mapreduce_jobhistory_done_dir,
- action="create_delayed",
- owner=params.mapred_user,
- group=params.user_group,
- mode=0777
- )
- params.HdfsDirectory(None, action="create")
-
- Directory(params.mapred_pid_dir,
- owner=params.mapred_user,
- group=params.user_group,
- recursive=True
- )
-
- mapred_log_dir = os.path.join(params.mapred_log_dir_prefix, params.mapred_user)
- mapred_userlogs_dir = os.path.join(mapred_log_dir, "userlogs")
-
- Directory(mapred_log_dir,
- recursive=True,
- owner=params.mapred_user,
- group=params.user_group
- )
- Directory(mapred_userlogs_dir,
- recursive=True,
- mode = 01777
- )
- if name == 'jobtracker':
- File(os.path.join(mapred_log_dir, 'hadoop-mapreduce.jobsummary.log'),
- owner=params.mapred_user,
- group=params.user_group,
- mode=0664
- )
-
- Directory(params.mapred_local_dir.split(','),
- owner=params.mapred_user,
- mode=0755,
- recursive=True,
- ignore_failures=True
- )
-
- File(params.exclude_file_path,
- owner=params.mapred_user,
- group=params.user_group,
- )
-
- File(params.mapred_hosts_file_path,
- owner=params.mapred_user,
- group=params.user_group,
- )
-
- if params.security_enabled:
- tc_mode = 0644
- tc_owner = "root"
- else:
- tc_mode = None
- tc_owner = params.hdfs_user
-
- if params.security_enabled:
- File(os.path.join(params.hadoop_bin, "task-controller"),
- owner="root",
- group=params.mapred_tt_group,
- mode=06050
- )
- File(os.path.join(params.hadoop_conf_dir, 'taskcontroller.cfg'),
- owner = tc_owner,
- mode = tc_mode,
- group = params.mapred_tt_group,
- content=Template("taskcontroller.cfg.j2")
- )
- else:
- File(os.path.join(params.hadoop_conf_dir, 'taskcontroller.cfg'),
- owner=tc_owner,
- content=Template("taskcontroller.cfg.j2")
- )
-
- if "capacity-scheduler" in params.config['configurations']:
- XmlConfig("capacity-scheduler.xml",
- conf_dir=params.hadoop_conf_dir,
- configurations=params.config['configurations'][
- 'capacity-scheduler'],
- configuration_attributes=params.config['configuration_attributes']['capacity-scheduler'],
- owner=params.hdfs_user,
- group=params.user_group
- )
-
- if "mapred-queue-acls" in params.config['configurations']:
- XmlConfig("mapred-queue-acls.xml",
- conf_dir=params.hadoop_conf_dir,
- configurations=params.config['configurations'][
- 'mapred-queue-acls'],
- configuration_attributes=params.config['configuration_attributes']['mapred-queue-acls'],
- owner=params.mapred_user,
- group=params.user_group
- )
- elif os.path.exists(
- os.path.join(params.hadoop_conf_dir, "mapred-queue-acls.xml")):
- File(os.path.join(params.hadoop_conf_dir, "mapred-queue-acls.xml"),
- owner=params.mapred_user,
- group=params.user_group
- )
-
- if "mapred-site" in params.config['configurations']:
- XmlConfig("mapred-site.xml",
- conf_dir=params.hadoop_conf_dir,
- configurations=params.config['configurations']['mapred-site'],
- configuration_attributes=params.config['configuration_attributes']['mapred-site'],
- owner=params.mapred_user,
- group=params.user_group
- )
-
- if os.path.exists(os.path.join(params.hadoop_conf_dir, 'fair-scheduler.xml')):
- File(os.path.join(params.hadoop_conf_dir, 'fair-scheduler.xml'),
- owner=params.mapred_user,
- group=params.user_group
- )
-
- if os.path.exists(
- os.path.join(params.hadoop_conf_dir, 'ssl-client.xml.example')):
- File(os.path.join(params.hadoop_conf_dir, 'ssl-client.xml.example'),
- owner=params.mapred_user,
- group=params.user_group
- )
-
- if os.path.exists(
- os.path.join(params.hadoop_conf_dir, 'ssl-server.xml.example')):
- File(os.path.join(params.hadoop_conf_dir, 'ssl-server.xml.example'),
- owner=params.mapred_user,
- group=params.user_group
- )
http://git-wip-us.apache.org/repos/asf/ambari/blob/23b7c110/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/params.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/params.py b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/params.py
deleted file mode 100644
index 06ebc39..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/params.py
+++ /dev/null
@@ -1,100 +0,0 @@
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-Ambari Agent
-
-"""
-from resource_management import *
-import status_params
-
-# server configurations
-config = Script.get_config()
-
-conf_dir = "/etc/hadoop/conf"
-
-mapred_user = status_params.mapred_user
-pid_dir_prefix = status_params.pid_dir_prefix
-mapred_pid_dir = status_params.mapred_pid_dir
-
-historyserver_pid_file = status_params.historyserver_pid_file
-jobtracker_pid_file = status_params.jobtracker_pid_file
-tasktracker_pid_file = status_params.tasktracker_pid_file
-
-hadoop_libexec_dir = '/usr/lib/hadoop/libexec'
-hadoop_bin = "/usr/lib/hadoop/bin"
-user_group = config['configurations']['cluster-env']['user_group']
-hdfs_log_dir_prefix = config['configurations']['hadoop-env']['hdfs_log_dir_prefix']
-mapred_log_dir_prefix = hdfs_log_dir_prefix
-mapred_local_dir = config['configurations']['mapred-site']['mapred.local.dir']
-update_exclude_file_only = default("/commandParams/update_exclude_file_only",False)
-
-hadoop_jar_location = "/usr/lib/hadoop/"
-smokeuser = config['configurations']['cluster-env']['smokeuser']
-security_enabled = config['configurations']['cluster-env']['security_enabled']
-smoke_user_keytab = config['configurations']['cluster-env']['smokeuser_keytab']
-kinit_path_local = functions.get_kinit_path(["/usr/bin", "/usr/kerberos/bin", "/usr/sbin"])
-
-#exclude file
-mr_exclude_hosts = default("/clusterHostInfo/decom_tt_hosts", [])
-exclude_file_path = config['configurations']['mapred-site']['mapred.hosts.exclude']
-mapred_hosts_file_path = config['configurations']['mapred-site']['mapred.hosts']
-
-#hdfs directories
-mapreduce_jobhistory_intermediate_done_dir = default('/configurations/mapred-site/mapreduce.jobhistory.intermediate-done-dir', '/mr-history/tmp')
-mapreduce_jobhistory_done_dir = config['configurations']['mapred-site']['mapred.job.tracker.history.completed.location']
-#for create_hdfs_directory
-hostname = config["hostname"]
-hadoop_conf_dir = "/etc/hadoop/conf"
-hadoop_pid_dir_prefix = config['configurations']['hadoop-env']['hadoop_pid_dir_prefix']
-
-if security_enabled:
- jt_principal_name = config['configurations']['mapred-site']['mapreduce.jobtracker.kerberos.principal']
- jt_principal_name = jt_principal_name.replace('_HOST',hostname.lower())
- jt_user_keytab = config['configurations']['mapred-site']['mapreduce.jobtracker.keytab.file']
- kinit_cmd = format("{kinit_path_local} -kt {jt_user_keytab} {jt_principal_name}")
-
-hdfs_user_keytab = config['configurations']['hadoop-env']['hdfs_user_keytab']
-hdfs_user = config['configurations']['hadoop-env']['hdfs_user']
-hdfs_principal_name = config['configurations']['hadoop-env']['hdfs_principal_name']
-import functools
-#create partial functions with common arguments for every HdfsDirectory call
-#to create hdfs directory we need to call params.HdfsDirectory in code
-HdfsDirectory = functools.partial(
- HdfsDirectory,
- conf_dir=hadoop_conf_dir,
- hdfs_user=hdfs_user,
- security_enabled = security_enabled,
- keytab = hdfs_user_keytab,
- kinit_path_local = kinit_path_local
-)
-
-mapred_tt_group = default("/configurations/mapred-site/mapreduce.tasktracker.group", user_group)
-
-slave_hosts = default("/clusterHostInfo/slave_hosts", [])
-
-rca_enabled = False
-if 'mapred-env' in config['configurations']:
- rca_enabled = config['configurations']['mapred-env']['rca_enabled']
-
-ambari_db_rca_url = config['hostLevelParams']['ambari_db_rca_url']
-ambari_db_rca_driver = config['hostLevelParams']['ambari_db_rca_driver']
-ambari_db_rca_username = config['hostLevelParams']['ambari_db_rca_username']
-ambari_db_rca_password = config['hostLevelParams']['ambari_db_rca_password']
-
-rca_properties = ''
-if rca_enabled and 'rca_properties' in config['configurations']['mapred-env']:
- rca_properties = format(config['configurations']['mapred-env']['rca_properties'])
\ No newline at end of file
http://git-wip-us.apache.org/repos/asf/ambari/blob/23b7c110/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/service.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/service.py b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/service.py
deleted file mode 100644
index 1244472..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/service.py
+++ /dev/null
@@ -1,56 +0,0 @@
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-Ambari Agent
-
-"""
-
-from resource_management import *
-
-
-def service(
- name,
- action='start'):
-
- import params
-
- pid_file = format("{mapred_pid_dir}/hadoop-{mapred_user}-{name}.pid")
- hadoop_daemon = format("export HADOOP_LIBEXEC_DIR={hadoop_libexec_dir} && {hadoop_bin}/hadoop-daemon.sh")
- cmd = format("{hadoop_daemon} --config {conf_dir}")
-
- if action == 'start':
- daemon_cmd = format("{cmd} start {name}")
- no_op = format("ls {pid_file} >/dev/null 2>&1 && ps -p `cat {pid_file}` >/dev/null 2>&1")
- Execute(daemon_cmd,
- user=params.mapred_user,
- not_if=no_op
- )
-
- Execute(no_op,
- user=params.mapred_user,
- not_if=no_op,
- initial_wait=5
- )
- elif action == 'stop':
- daemon_cmd = format("{cmd} stop {name}")
-
- Execute(daemon_cmd,
- user=params.mapred_user
- )
- File(pid_file,
- action = "delete",
- )
http://git-wip-us.apache.org/repos/asf/ambari/blob/23b7c110/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/service_check.py b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/service_check.py
deleted file mode 100644
index c32e1e1..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/service_check.py
+++ /dev/null
@@ -1,87 +0,0 @@
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-Ambari Agent
-
-"""
-
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-Ambari Agent
-
-"""
-
-from resource_management import *
-
-class ServiceCheck(Script):
- def service_check(self, env):
- import params
- env.set_params(params)
-
- jar_location = params.hadoop_jar_location
- input_file = 'mapredsmokeinput'
- output_file = "mapredsmokeoutput"
-
- cleanup_cmd = format("dfs -rmr {output_file} {input_file}")
- create_file_cmd = format("{cleanup_cmd} ; hadoop dfs -put /etc/passwd {input_file}")
- test_cmd = format("fs -test -e {output_file}")
- run_wordcount_job = format("jar {jar_location}/hadoop-examples.jar wordcount {input_file} {output_file}")
-
- if params.security_enabled:
- kinit_cmd = format("{kinit_path_local} -kt {smoke_user_keytab} {smokeuser};")
-
- Execute(kinit_cmd,
- user=params.smokeuser
- )
-
- ExecuteHadoop(create_file_cmd,
- tries=1,
- try_sleep=5,
- user=params.smokeuser,
- conf_dir=params.conf_dir
- )
-
- ExecuteHadoop(run_wordcount_job,
- tries=1,
- try_sleep=5,
- user=params.smokeuser,
- conf_dir=params.conf_dir,
- logoutput=True
- )
-
- ExecuteHadoop(test_cmd,
- user=params.smokeuser,
- conf_dir=params.conf_dir
- )
-
-if __name__ == "__main__":
- ServiceCheck().execute()
http://git-wip-us.apache.org/repos/asf/ambari/blob/23b7c110/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/status_params.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/status_params.py b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/status_params.py
deleted file mode 100644
index 11986b0..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/status_params.py
+++ /dev/null
@@ -1,32 +0,0 @@
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-Ambari Agent
-
-"""
-
-from resource_management import *
-
-config = Script.get_config()
-
-mapred_user = config['configurations']['mapred-env']['mapred_user']
-pid_dir_prefix = config['configurations']['hadoop-env']['hadoop_pid_dir_prefix']
-mapred_pid_dir = format("{pid_dir_prefix}/{mapred_user}")
-
-jobtracker_pid_file = format("{mapred_pid_dir}/hadoop-{mapred_user}-jobtracker.pid")
-tasktracker_pid_file = format("{mapred_pid_dir}/hadoop-{mapred_user}-tasktracker.pid")
-historyserver_pid_file = format("{mapred_pid_dir}/hadoop-{mapred_user}-historyserver.pid")
http://git-wip-us.apache.org/repos/asf/ambari/blob/23b7c110/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/tasktracker.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/tasktracker.py b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/tasktracker.py
deleted file mode 100644
index c405b94..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/scripts/tasktracker.py
+++ /dev/null
@@ -1,101 +0,0 @@
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-Ambari Agent
-
-"""
-
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-Ambari Agent
-
-"""
-
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-Ambari Agent
-
-"""
-import sys
-from resource_management import *
-
-from mapreduce import mapreduce
-from service import service
-
-class Tasktracker(Script):
- def install(self, env):
- self.install_packages(env)
- self.configure(env)
-
- def configure(self, env):
- import params
- env.set_params(params)
- mapreduce()
-
- def start(self, env):
- import params
- env.set_params(params)
- self.configure(env) # FOR SECURITY
- service('tasktracker',
- action='start'
- )
-
- def stop(self, env):
- import params
- env.set_params(params)
-
- service('tasktracker',
- action='stop'
- )
-
- def status(self, env):
- import status_params
- env.set_params(status_params)
- check_process_status(status_params.tasktracker_pid_file)
-
-if __name__ == "__main__":
- Tasktracker().execute()
http://git-wip-us.apache.org/repos/asf/ambari/blob/23b7c110/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/templates/exclude_hosts_list.j2
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/templates/exclude_hosts_list.j2 b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/templates/exclude_hosts_list.j2
deleted file mode 100644
index 6e99050..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/templates/exclude_hosts_list.j2
+++ /dev/null
@@ -1,22 +0,0 @@
-{#
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-#}
-
-
-{% for host in mr_exclude_hosts %}
-{{host}}
-{% endfor %}
\ No newline at end of file
http://git-wip-us.apache.org/repos/asf/ambari/blob/23b7c110/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/templates/taskcontroller.cfg.j2
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/templates/taskcontroller.cfg.j2 b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/templates/taskcontroller.cfg.j2
deleted file mode 100644
index 3d5f4f2..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/package/templates/taskcontroller.cfg.j2
+++ /dev/null
@@ -1,38 +0,0 @@
-{#
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-#}
-
-#/*
-# * Licensed to the Apache Software Foundation (ASF) under one
-# * or more contributor license agreements. See the NOTICE file
-# * distributed with this work for additional information
-# * regarding copyright ownership. The ASF licenses this file
-# * to you under the Apache License, Version 2.0 (the
-# * "License"); you may not use this file except in compliance
-# * with the License. You may obtain a copy of the License at
-# *
-# * http://www.apache.org/licenses/LICENSE-2.0
-# *
-# * Unless required by applicable law or agreed to in writing, software
-# * distributed under the License is distributed on an "AS IS" BASIS,
-# * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# * See the License for the specific language governing permissions and
-# * limitations under the License.
-# */
-mapred.local.dir={{mapred_local_dir}}
-mapreduce.tasktracker.group={{mapred_tt_group}}
-hadoop.log.dir={{hdfs_log_dir_prefix}}/{{mapred_user}}
http://git-wip-us.apache.org/repos/asf/ambari/blob/23b7c110/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/OOZIE/alerts.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/OOZIE/alerts.json b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/OOZIE/alerts.json
deleted file mode 100644
index caca810..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/OOZIE/alerts.json
+++ /dev/null
@@ -1,42 +0,0 @@
-{
- "OOZIE": {
- "service": [],
- "OOZIE_SERVER": [
- {
- "name": "oozie_server_webui",
- "label": "Oozie Server Web UI",
- "description": "This host-level alert is triggered if the Oozie server Web UI is unreachable.",
- "interval": 1,
- "scope": "ANY",
- "source": {
- "type": "WEB",
- "uri": {
- "http": "{{oozie-site/oozie.base.url}}/oozie"
- },
- "reporting": {
- "ok": {
- "text": "HTTP {0} response in {2:.3f} seconds"
- },
- "warning":{
- "text": "HTTP {0} response in {2:.3f} seconds"
- },
- "critical": {
- "text": "Connection failed to {1}"
- }
- }
- }
- },
- {
- "name": "oozie_server_status",
- "label": "Oozie Server Status",
- "description": "This host-level alert is triggered if the Oozie server cannot be determined to be up and responding to client requests.",
- "interval": 1,
- "scope": "ANY",
- "source": {
- "type": "SCRIPT",
- "path": "HDP/1.3.2/services/OOZIE/package/alerts/alert_check_oozie_server.py"
- }
- }
- ]
- }
-}
\ No newline at end of file
http://git-wip-us.apache.org/repos/asf/ambari/blob/23b7c110/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/OOZIE/configuration/oozie-env.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/OOZIE/configuration/oozie-env.xml b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/OOZIE/configuration/oozie-env.xml
deleted file mode 100644
index 37ae8dd..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/OOZIE/configuration/oozie-env.xml
+++ /dev/null
@@ -1,121 +0,0 @@
-<?xml version="1.0"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
--->
-
-<configuration>
- <property>
- <name>oozie_user</name>
- <value>oozie</value>
- <property-type>USER</property-type>
- <description>Oozie User.</description>
- </property>
- <property>
- <name>oozie_database</name>
- <value>New Derby Database</value>
- <description>Oozie Server Database.</description>
- </property>
- <property>
- <name>oozie_derby_database</name>
- <value>Derby</value>
- <description>Oozie Derby Database.</description>
- </property>
- <property>
- <name>oozie_data_dir</name>
- <value>/hadoop/oozie/data</value>
- <description>Data directory in which the Oozie DB exists</description>
- </property>
- <property>
- <name>oozie_log_dir</name>
- <value>/var/log/oozie</value>
- <description>Directory for oozie logs</description>
- </property>
- <property>
- <name>oozie_pid_dir</name>
- <value>/var/run/oozie</value>
- <description>Directory in which the pid files for oozie reside.</description>
- </property>
- <property>
- <name>oozie_admin_port</name>
- <value>11001</value>
- <description>The admin port Oozie server runs.</description>
- </property>
-
- <!-- oozie-env.sh -->
- <property>
- <name>content</name>
- <description>This is the jinja template for oozie-env.sh file</description>
- <value>
-#!/bin/bash
-
-#Set JAVA HOME
-export JAVA_HOME={{java_home}}
-export JRE_HOME={{java_home}}
-
-# Set Oozie specific environment variables here.
-
-# Settings for the Embedded Tomcat that runs Oozie
-# Java System properties for Oozie should be specified in this variable
-#
-# export CATALINA_OPTS=
-
-# Oozie configuration file to load from Oozie configuration directory
-#
-# export OOZIE_CONFIG_FILE=oozie-site.xml
-
-# Oozie logs directory
-#
-export OOZIE_LOG={{oozie_log_dir}}
-
-# Oozie pid directory
-#
-export CATALINA_PID={{pid_file}}
-
-#Location of the data for oozie
-export OOZIE_DATA={{oozie_data_dir}}
-
-# Oozie Log4J configuration file to load from Oozie configuration directory
-#
-# export OOZIE_LOG4J_FILE=oozie-log4j.properties
-
-# Reload interval of the Log4J configuration file, in seconds
-#
-# export OOZIE_LOG4J_RELOAD=10
-
-# The port Oozie server runs
-#
-export OOZIE_HTTP_PORT={{oozie_server_port}}
-
-# The admin port Oozie server runs
-#
-export OOZIE_ADMIN_PORT={{oozie_server_admin_port}}
-
-# The host name Oozie server runs on
-#
-# export OOZIE_HTTP_HOSTNAME=`hostname -f`
-
-# The base URL for callback URLs to Oozie
-#
-# export OOZIE_BASE_URL="http://${OOZIE_HTTP_HOSTNAME}:${OOZIE_HTTP_PORT}/oozie"
-export JAVA_LIBRARY_PATH=/usr/lib/hadoop/lib/native/Linux-amd64-64
- </value>
- </property>
-
-</configuration>
http://git-wip-us.apache.org/repos/asf/ambari/blob/23b7c110/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/OOZIE/configuration/oozie-log4j.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/OOZIE/configuration/oozie-log4j.xml b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/OOZIE/configuration/oozie-log4j.xml
deleted file mode 100644
index 7f7158f..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/OOZIE/configuration/oozie-log4j.xml
+++ /dev/null
@@ -1,97 +0,0 @@
-<?xml version="1.0"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
--->
-
-<configuration supports_final="false">
-
- <property>
- <name>content</name>
- <description>Custom log4j.properties</description>
- <value>
-#
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License. See accompanying LICENSE file.
-#
-
-# If the Java System property 'oozie.log.dir' is not defined at Oozie start up time
-# XLogService sets its value to '${oozie.home}/logs'
-
-log4j.appender.oozie=org.apache.log4j.DailyRollingFileAppender
-log4j.appender.oozie.DatePattern='.'yyyy-MM-dd-HH
-log4j.appender.oozie.File=${oozie.log.dir}/oozie.log
-log4j.appender.oozie.Append=true
-log4j.appender.oozie.layout=org.apache.log4j.PatternLayout
-log4j.appender.oozie.layout.ConversionPattern=%d{ISO8601} %5p %c{1}:%L - SERVER[${oozie.instance.id}] %m%n
-
-log4j.appender.oozieops=org.apache.log4j.DailyRollingFileAppender
-log4j.appender.oozieops.DatePattern='.'yyyy-MM-dd
-log4j.appender.oozieops.File=${oozie.log.dir}/oozie-ops.log
-log4j.appender.oozieops.Append=true
-log4j.appender.oozieops.layout=org.apache.log4j.PatternLayout
-log4j.appender.oozieops.layout.ConversionPattern=%d{ISO8601} %5p %c{1}:%L - %m%n
-
-log4j.appender.oozieinstrumentation=org.apache.log4j.DailyRollingFileAppender
-log4j.appender.oozieinstrumentation.DatePattern='.'yyyy-MM-dd
-log4j.appender.oozieinstrumentation.File=${oozie.log.dir}/oozie-instrumentation.log
-log4j.appender.oozieinstrumentation.Append=true
-log4j.appender.oozieinstrumentation.layout=org.apache.log4j.PatternLayout
-log4j.appender.oozieinstrumentation.layout.ConversionPattern=%d{ISO8601} %5p %c{1}:%L - %m%n
-
-log4j.appender.oozieaudit=org.apache.log4j.DailyRollingFileAppender
-log4j.appender.oozieaudit.DatePattern='.'yyyy-MM-dd
-log4j.appender.oozieaudit.File=${oozie.log.dir}/oozie-audit.log
-log4j.appender.oozieaudit.Append=true
-log4j.appender.oozieaudit.layout=org.apache.log4j.PatternLayout
-log4j.appender.oozieaudit.layout.ConversionPattern=%d{ISO8601} %5p %c{1}:%L - %m%n
-
-log4j.appender.openjpa=org.apache.log4j.DailyRollingFileAppender
-log4j.appender.openjpa.DatePattern='.'yyyy-MM-dd
-log4j.appender.openjpa.File=${oozie.log.dir}/oozie-jpa.log
-log4j.appender.openjpa.Append=true
-log4j.appender.openjpa.layout=org.apache.log4j.PatternLayout
-log4j.appender.openjpa.layout.ConversionPattern=%d{ISO8601} %5p %c{1}:%L - %m%n
-
-log4j.logger.openjpa=INFO, openjpa
-log4j.logger.oozieops=INFO, oozieops
-log4j.logger.oozieinstrumentation=ALL, oozieinstrumentation
-log4j.logger.oozieaudit=ALL, oozieaudit
-log4j.logger.org.apache.oozie=INFO, oozie
-log4j.logger.org.apache.hadoop=WARN, oozie
-log4j.logger.org.mortbay=WARN, oozie
-log4j.logger.org.hsqldb=WARN, oozie
-log4j.logger.org.apache.hadoop.security.authentication.server=INFO, oozie
- </value>
- </property>
-
-</configuration>