You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@bigtop.apache.org by rv...@apache.org on 2012/02/07 23:21:49 UTC
svn commit: r1241663 - in /incubator/bigtop/branches/RCs:
bigtop-packages/src/common/hbase/ bigtop-packages/src/deb/hbase/
bigtop-packages/src/rpm/hbase/SPECS/
bigtop-test-framework/src/main/groovy/org/apache/bigtop/itest/pmanager/
bigtop-tests/test-ar...
Author: rvs
Date: Tue Feb 7 22:21:49 2012
New Revision: 1241663
URL: http://svn.apache.org/viewvc?rev=1241663&view=rev
Log:
Merge branch 'trunk' into RCs
Modified:
incubator/bigtop/branches/RCs/bigtop-packages/src/common/hbase/install_hbase.sh
incubator/bigtop/branches/RCs/bigtop-packages/src/deb/hbase/rules
incubator/bigtop/branches/RCs/bigtop-packages/src/rpm/hbase/SPECS/hbase.spec
incubator/bigtop/branches/RCs/bigtop-test-framework/src/main/groovy/org/apache/bigtop/itest/pmanager/PackageManager.groovy
incubator/bigtop/branches/RCs/bigtop-tests/test-artifacts/package/src/main/resources/package_data_apt.xml
incubator/bigtop/branches/RCs/bigtop-tests/test-artifacts/package/src/main/resources/package_data_urpmi.xml
incubator/bigtop/branches/RCs/bigtop-tests/test-artifacts/package/src/main/resources/package_data_yum.xml
incubator/bigtop/branches/RCs/bigtop-tests/test-artifacts/package/src/main/resources/package_data_zypper.xml
Modified: incubator/bigtop/branches/RCs/bigtop-packages/src/common/hbase/install_hbase.sh
URL: http://svn.apache.org/viewvc/incubator/bigtop/branches/RCs/bigtop-packages/src/common/hbase/install_hbase.sh?rev=1241663&r1=1241662&r2=1241663&view=diff
==============================================================================
--- incubator/bigtop/branches/RCs/bigtop-packages/src/common/hbase/install_hbase.sh (original)
+++ incubator/bigtop/branches/RCs/bigtop-packages/src/common/hbase/install_hbase.sh Tue Feb 7 22:21:49 2012
@@ -131,10 +131,9 @@ elif [ -e /usr/lib/bigtop-utils/bigtop-d
. /usr/lib/bigtop-utils/bigtop-detect-javahome
fi
-export ZOOKEEPER_CONF=\${ZOOKEEPER_CONF:-/etc/zookeeper}
export HADOOP_CONF=\${HADOOP_CONF:-/etc/hadoop/conf}
export ZOOKEEPER_HOME=\${ZOOKEEPER_HOME:-/usr/lib/zookeeper}
-export HBASE_CLASSPATH=\$ZOOKEEPER_CONF:\$HADOOP_CONF:\$HADOOP_HOME/*:\$HADOOP_HOME/lib/*:\$ZOOKEEPER_HOME/*:\$ZOOKEEPER_HOME/lib/*:\$HBASE_CLASSPATH
+export HBASE_CLASSPATH=\$HADOOP_CONF:\$HADOOP_HOME/*:\$HADOOP_HOME/lib/*:\$ZOOKEEPER_HOME/*:\$ZOOKEEPER_HOME/lib/*:\$HBASE_CLASSPATH
exec /usr/lib/hbase/bin/hbase "\$@"
EOF
Modified: incubator/bigtop/branches/RCs/bigtop-packages/src/deb/hbase/rules
URL: http://svn.apache.org/viewvc/incubator/bigtop/branches/RCs/bigtop-packages/src/deb/hbase/rules?rev=1241663&r1=1241662&r2=1241663&view=diff
==============================================================================
--- incubator/bigtop/branches/RCs/bigtop-packages/src/deb/hbase/rules (original)
+++ incubator/bigtop/branches/RCs/bigtop-packages/src/deb/hbase/rules Tue Feb 7 22:21:49 2012
@@ -58,6 +58,8 @@ install: build
cp debian/hbase.nofiles.conf debian/tmp/etc/security/limits.d/${hbase_pkg_name}.nofiles.conf
# Symlink in the dependency jars from their packages. Both of these packages
# provide an unversioned symlink foo.jar -> foo-0.1.2.jar.
+ rm -f debian/tmp/usr/lib/hadoop-*
+ rm -f debian/tmp/usr/lib/zookeeper-*
ln -f -s ${hbase_jar_deps} debian/tmp/usr/lib/${hbase_pkg_name}/lib/
ln -s /var/log/${hbase_pkg_name} debian/tmp/usr/lib/${hbase_pkg_name}/logs
ln -s /var/run/${hbase_pkg_name} debian/tmp/usr/lib/${hbase_pkg_name}/pids
Modified: incubator/bigtop/branches/RCs/bigtop-packages/src/rpm/hbase/SPECS/hbase.spec
URL: http://svn.apache.org/viewvc/incubator/bigtop/branches/RCs/bigtop-packages/src/rpm/hbase/SPECS/hbase.spec?rev=1241663&r1=1241662&r2=1241663&view=diff
==============================================================================
--- incubator/bigtop/branches/RCs/bigtop-packages/src/rpm/hbase/SPECS/hbase.spec (original)
+++ incubator/bigtop/branches/RCs/bigtop-packages/src/rpm/hbase/SPECS/hbase.spec Tue Feb 7 22:21:49 2012
@@ -231,6 +231,8 @@ done
%__install -d -m 0755 $RPM_BUILD_ROOT/usr/bin
# Pull zookeeper and hadoop from their packages
+rm -f $RPM_BUILD_ROOT/%{lib_hbase}/hadoop-*
+rm -f $RPM_BUILD_ROOT/%{lib_hbase}/zookeeper-*
ln -f -s %{hbase_jar_deps} $RPM_BUILD_ROOT/%{lib_hbase}
%pre
Modified: incubator/bigtop/branches/RCs/bigtop-test-framework/src/main/groovy/org/apache/bigtop/itest/pmanager/PackageManager.groovy
URL: http://svn.apache.org/viewvc/incubator/bigtop/branches/RCs/bigtop-test-framework/src/main/groovy/org/apache/bigtop/itest/pmanager/PackageManager.groovy?rev=1241663&r1=1241662&r2=1241663&view=diff
==============================================================================
--- incubator/bigtop/branches/RCs/bigtop-test-framework/src/main/groovy/org/apache/bigtop/itest/pmanager/PackageManager.groovy (original)
+++ incubator/bigtop/branches/RCs/bigtop-test-framework/src/main/groovy/org/apache/bigtop/itest/pmanager/PackageManager.groovy Tue Feb 7 22:21:49 2012
@@ -209,7 +209,7 @@ public abstract class PackageManager {
switch (linux_flavor ?: OS.linux_flavor) {
case ~/(?is).*(ubuntu|debian).*/:
return new AptCmdLinePackageManager();
- case ~/(?is).*(redhat|centos|rhel).*/:
+ case ~/(?is).*(redhat|centos|rhel|fedora).*/:
return new YumCmdLinePackageManager();
case ~/(?is).*(suse|sles|sled).*/:
return new ZypperCmdLinePackageManager();
Modified: incubator/bigtop/branches/RCs/bigtop-tests/test-artifacts/package/src/main/resources/package_data_apt.xml
URL: http://svn.apache.org/viewvc/incubator/bigtop/branches/RCs/bigtop-tests/test-artifacts/package/src/main/resources/package_data_apt.xml?rev=1241663&r1=1241662&r2=1241663&view=diff
==============================================================================
--- incubator/bigtop/branches/RCs/bigtop-tests/test-artifacts/package/src/main/resources/package_data_apt.xml (original)
+++ incubator/bigtop/branches/RCs/bigtop-tests/test-artifacts/package/src/main/resources/package_data_apt.xml Tue Feb 7 22:21:49 2012
@@ -65,7 +65,7 @@
* Smart defaults for services. You can get a properly configured system
running quickly, while still being able to override settings as needed.
</description>
- <url>http://incubator.apache.org/whirr</url>
+ <url>http://whirr.apache.org/</url>
</metadata>
<deps>
<sun-java6-jre/>
@@ -1139,36 +1139,6 @@
</flume-conf>
</alternatives>
</flume>
-<flume-master>
- <metadata>
- <summary>central administration point for the flume data collection system</summary>
- <description>The Flume master daemon is the central administration and data path control
- point for flume nodes.</description>
- <url>http://www.cloudera.com</url>
- </metadata>
- <deps>
- <flume>/self</flume>
- </deps>
- <services>
- <flume-master>
- <runlevel>2</runlevel><runlevel>3</runlevel><runlevel>4</runlevel><runlevel>5</runlevel>
- <oninstall>start</oninstall>
- <configured>true</configured>
- </flume-master>
- </services>
- <content>
- <file name="/." owners="0" perm="drwxr-xr-x" user="root" group="root"/>
- <file name="/etc" owners="119" perm="drwxr-xr-x" user="root" group="root"/>
- <file name="/etc/init.d" owners="32" perm="drwxr-xr-x" user="root" group="root"/>
- <config name="/etc/init.d/flume-master" owners="1" perm="-rwxr-xr-x" user="root" group="root"/>
- <file name="/usr" owners="418" perm="drwxr-xr-x" user="root" group="root"/>
- <file name="/usr/share" owners="418" perm="drwxr-xr-x" user="root" group="root"/>
- <file name="/usr/share/doc" owners="418" perm="drwxr-xr-x" user="root" group="root"/>
- <file name="/usr/share/doc/flume-master" owners="1" perm="drwxr-xr-x" user="root" group="root"/>
- <file name="/usr/share/doc/flume-master/copyright" owners="1" perm="-rw-r--r--" user="root" group="root"/>
- <file name="/usr/share/doc/flume-master/changelog.Debian.gz" owners="1" perm="-rw-r--r--" user="root" group="root"/>
- </content>
-</flume-master>
<flume-node>
<metadata>
<summary>core element of Flume's data path that collects and delivers data</summary>
@@ -1446,7 +1416,7 @@
In case of workflow job failure, the workflow job can be rerun skipping
previously completed actions, the workflow application can be patched before
being rerun.</description>
- <url>http://archive.cloudera.com/cdh/3/oozie</url>
+ <url>http://incubator.apache.org/oozie/</url>
</metadata>
<deps>
<oozie-client>/self</oozie-client>
@@ -1575,7 +1545,7 @@
user can deploy workflows and perform other administrative and
monitoring tasks such as start, stop, kill, resume workflows
and coordinator jobs.</description>
- <url>http://archive.cloudera.com/cdh/3/oozie</url>
+ <url>http://incubator.apache.org/oozie/</url>
</metadata>
<content>
<file name="/." owners="0" perm="drwxr-xr-x" user="root" group="root"/>
@@ -2390,7 +2360,7 @@
<summary>A high-performance coordination service for distributed applications.</summary>
<description>ZooKeeper is a centralized service for maintaining configuration information, naming, providing distributed synchronization, and providing group services. All of these kinds of services are used in some form or another by distributed applications. Each time they are implemented there is a lot of work that goes into fixing the bugs and race conditions that are inevitable. Because of the difficulty of implementing these kinds of services, applications initially usually skimp on them ,which make them brittle in the presence of change and difficult to manage. Even when done correctly, different implementations of these services lead to management complexity when the applications are deployed.
</description>
- <url>http://hadoop.apache.org/zookeeper/</url>
+ <url>http://zookeeper.apache.org/</url>
<!-- group>misc</group -->
</metadata>
<deps>
@@ -2787,7 +2757,7 @@
<!-- license>APL2</license -->
<!-- arch>universal</arch -->
<summary>This runs the zookeeper server on startup.</summary>
- <url>http://hadoop.apache.org/zookeeper/</url>
+ <url>http://zookeeper.apache.org/</url>
<!-- vendor>(none)</vendor -->
<!-- group>Development/Libraries</group -->
<!-- depends><dep>adduser</dep><dep>sun-java6-jre</dep><dep>sun-java6-bin</dep></depends -->
@@ -2842,7 +2812,7 @@
automatically, allowing the user to focus on semantics rather than efficiency.
* Extensibility
Users can create their own functions to do special-purpose processing.</description>
- <url>http://hadoop.apache.org/pig/</url>
+ <url>http://pig.apache.org/</url>
</metadata>
<deps>
<sun-java6-jre/>
@@ -6787,7 +6757,7 @@
plug in their custom mappers and reducers to do more sophisticated
analysis which may not be supported by the built-in capabilities of
the language.</description>
- <url>http://hadoop.apache.org/hive/</url>
+ <url>http://hive.apache.org/</url>
</metadata>
<deps>
<sun-java6-jre/>
@@ -7099,7 +7069,7 @@
<description>Use it when you need random, realtime read/write access to your Big Data.
This project's goal is the hosting of very large tables -- billions of rows
X millions of columns -- atop clusters of commodity hardware.</description>
- <url>http://hadoop.apache.org/hbase/</url>
+ <url>http://hbase.apache.org/</url>
<!-- group>misc</group -->
</metadata>
<deps>
@@ -7290,7 +7260,7 @@
<metadata>
<summary>Documentation for HBase</summary>
<description>This package contains the HBase manual and JavaDoc.</description>
- <url>http://hadoop.apache.org/hbase/</url>
+ <url>http://hbase.apache.org/</url>
<!-- group>misc</group -->
</metadata>
<content>
@@ -9065,7 +9035,7 @@
<metadata>
<summary>HMaster is the "master server" for a HBase</summary>
<description>There is only one HMaster for a single HBase deployment.</description>
- <url>http://hadoop.apache.org/hbase/</url>
+ <url>http://hbase.apache.org/</url>
</metadata>
<deps>
<hbase>/self</hbase>
@@ -9098,7 +9068,7 @@
<summary>HRegionServer makes a set of HRegions available to clients</summary>
<description>It checks in with the HMaster. There are many HRegionServers in a single
HBase deployment.</description>
- <url>http://hadoop.apache.org/hbase/</url>
+ <url>http://hbase.apache.org/</url>
</metadata>
<deps>
<hbase>/self</hbase>
@@ -9131,7 +9101,7 @@
<summary>Provides an HBase Thrift service</summary>
<description>This package provides a Thrift service interface to the HBase distributed
database.</description>
- <url>http://hadoop.apache.org/hbase/</url>
+ <url>http://hbase.apache.org/</url>
</metadata>
<deps>
<hbase>/self</hbase>
Modified: incubator/bigtop/branches/RCs/bigtop-tests/test-artifacts/package/src/main/resources/package_data_urpmi.xml
URL: http://svn.apache.org/viewvc/incubator/bigtop/branches/RCs/bigtop-tests/test-artifacts/package/src/main/resources/package_data_urpmi.xml?rev=1241663&r1=1241662&r2=1241663&view=diff
==============================================================================
--- incubator/bigtop/branches/RCs/bigtop-tests/test-artifacts/package/src/main/resources/package_data_urpmi.xml (original)
+++ incubator/bigtop/branches/RCs/bigtop-tests/test-artifacts/package/src/main/resources/package_data_urpmi.xml Tue Feb 7 22:21:49 2012
@@ -63,7 +63,7 @@ also on potential use cases. Come to the
service.
* Smart defaults for services. You can get a properly configured system
running quickly, while still being able to override settings as needed.</description>
- <url>http://incubator.apache.org/whirr</url>
+ <url>http://whirr.apache.org/</url>
</metadata>
<deps>
<tag name="/bin/bash"/>
@@ -804,7 +804,7 @@ also on potential use cases. Come to the
mechanisms and many failover and recovery mechanisms. The system is centrally
managed and allows for intelligent dynamic management. It uses a simple
extensible data model that allows for online analytic applications.</description>
- <url>https://github.com/cloudera/flume</url>
+ <url>http://incubator.apache.org/projects/flume.html</url>
</metadata>
<deps>
<tag name="/bin/sh"/>
@@ -1117,44 +1117,11 @@ also on potential use cases. Come to the
</flume-conf>
</alternatives>
</flume>
-<flume-master>
- <metadata>
- <summary>The flume master daemon is the central administration and data path control point for flume nodes.</summary>
- <description>Flume is a reliable, scalable, and manageable distributed data collection application for collecting data such as logs and delivering it to data stores such as Hadoop's HDFS. It can efficiently collect, aggregate, and move large amounts of log data. It has a simple, but flexible, architecture based on streaming data flows. It is robust and fault tolerant with tunable reliability mechanisms and many failover and recovery mechanisms. The system is centrally managed and allows for intelligent dynamic management. It uses a simple extensible data model that allows for online analytic applications.</description>
- <url>https://github.com/cloudera/flume</url>
- </metadata>
- <deps>
- <tag name="/bin/bash"/>
- <tag name="/bin/sh"/>
- <tag name="/sbin/chkconfig"/>
- <tag name="/sbin/service"/>
- <tag name="/usr/bin/env"/>
- <tag name="/usr/sbin/useradd"/>
- <flume>/self</flume>
- <jre>>=1.6</jre>
- <redhat-lsb/>
- <sh-utils/>
- <textutils/>
- </deps>
- <services>
- <flume-master>
- <runlevel>2</runlevel><runlevel>3</runlevel><runlevel>4</runlevel><runlevel>5</runlevel>
- <oninstall>stop</oninstall>
- <configured>true</configured>
- </flume-master>
- </services>
- <content>
- <file name="/etc/rc.d/init.d/flume-master" owners="1" perm="-rwxr-xr-x" user="root" group="root"/>
- <file name="/usr/lib/flume" owners="2" perm="drwxr-xr-x" user="root" group="root"/>
- <file name="/usr/lib/flume/bin" owners="2" perm="drwxr-xr-x" user="root" group="root"/>
- <file name="/usr/lib/flume/bin/flume-daemon.sh" owners="2" perm="-rwxr-xr-x" user="root" group="root"/>
- </content>
-</flume-master>
<flume-node>
<metadata>
<summary>The flume node daemon is a core element of flume's data path and is responsible for generating, processing, and delivering data.</summary>
<description>Flume is a reliable, scalable, and manageable distributed data collection application for collecting data such as logs and delivering it to data stores such as Hadoop's HDFS. It can efficiently collect, aggregate, and move large amounts of log data. It has a simple, but flexible, architecture based on streaming data flows. It is robust and fault tolerant with tunable reliability mechanisms and many failover and recovery mechanisms. The system is centrally managed and allows for intelligent dynamic management. It uses a simple extensible data model that allows for online analytic applications.</description>
- <url>https://github.com/cloudera/flume</url>
+ <url>http://incubator.apache.org/projects/flume.html</url>
</metadata>
<deps>
<tag name="/bin/bash"/>
@@ -1418,7 +1385,7 @@ also on potential use cases. Come to the
In case of workflow job failure, the workflow job can be rerun skipping
previously completed actions, the workflow application can be patched before
being rerun.</description>
- <url>http://www.cloudera.com</url>
+ <url>http://incubator.apache.org/oozie/</url>
</metadata>
<deps>
<tag name="/bin/bash"/>
@@ -1542,7 +1509,7 @@ also on potential use cases. Come to the
you can also change the status of the entire system, get vesion
information. This client utility also allows you to validate
any worflows before they are deployed to the Oozie server.</description>
- <url>http://www.cloudera.com</url>
+ <url>http://incubator.apache.org/oozie/</url>
</metadata>
<deps>
<tag name="/bin/bash"/>
@@ -2356,7 +2323,7 @@ into fixing the bugs and race conditions
difficulty of implementing these kinds of services, applications initially
usually skimp on them ,which make them brittle in the presence of change and
difficult to manage. Even when done correctly, different implementations of these services lead to management complexity when the applications are deployed.</description>
- <url>http://hadoop.apache.org/zookeeper/</url>
+ <url>http://zookeeper.apache.org/</url>
<!-- group>misc</group -->
</metadata>
<deps>
@@ -2740,7 +2707,7 @@ difficult to manage. Even when done corr
<!-- arch>universal</arch -->
<summary>The Hadoop Zookeeper server</summary>
<description>This package starts the zookeeper server on startup</description>
- <url>http://hadoop.apache.org/zookeeper/</url>
+ <url>http://zookeeper.apache.org/</url>
<!-- vendor>(none)</vendor -->
<!-- group>Development/Libraries</group -->
<!-- depends><dep>adduser</dep><dep>sun-java6-jre</dep><dep>sun-java6-bin</dep></depends -->
@@ -2788,7 +2755,7 @@ difficult to manage. Even when done corr
automatically, allowing the user to focus on semantics rather than efficiency.
* Extensibility
Users can create their own functions to do special-purpose processing.</description>
- <url>http://hadoop.apache.org/pig/</url>
+ <url>http://pig.apache.org/</url>
</metadata>
<deps>
<tag name="/bin/sh"/>
@@ -6717,7 +6684,7 @@ difficult to manage. Even when done corr
plug in their custom mappers and reducers to do more sophisticated
analysis which may not be supported by the built-in capabilities of
the language.</description>
- <url>http://hadoop.apache.org/hive/</url>
+ <url>http://hive.apache.org/</url>
</metadata>
<deps>
<tag name="/bin/sh"/>
@@ -7157,7 +7124,7 @@ difficult to manage. Even when done corr
* Cascading source and sink modules
* Extensible jruby-based (JIRB) shell
* Support for exporting metrics via the Hadoop metrics subsystem to files or Ganglia; or via JMX</description>
- <url>http://hadoop.apache.org/hbase/</url>
+ <url>http://hbase.apache.org/</url>
<!-- group>misc</group -->
</metadata>
<deps>
@@ -7336,7 +7303,7 @@ difficult to manage. Even when done corr
<metadata>
<summary>Hbase Documentation</summary>
<description>Documentation for Hbase</description>
- <url>http://hadoop.apache.org/hbase/</url>
+ <url>http://hbase.apache.org/</url>
<!-- group>misc</group -->
</metadata>
<content>
@@ -9105,7 +9072,7 @@ difficult to manage. Even when done corr
<metadata>
<summary>The Hadoop HBase master Server.</summary>
<description>HMaster is the "master server" for a HBase. There is only one HMaster for a single HBase deployment.</description>
- <url>http://hadoop.apache.org/hbase/</url>
+ <url>http://hbase.apache.org/</url>
</metadata>
<deps>
<tag name="/bin/bash"/>
@@ -9128,7 +9095,7 @@ difficult to manage. Even when done corr
<metadata>
<summary>The Hadoop HBase RegionServer server.</summary>
<description>HRegionServer makes a set of HRegions available to clients. It checks in with the HMaster. There are many HRegionServers in a single HBase deployment.</description>
- <url>http://hadoop.apache.org/hbase/</url>
+ <url>http://hbase.apache.org/</url>
</metadata>
<deps>
<tag name="/bin/bash"/>
@@ -9151,7 +9118,7 @@ difficult to manage. Even when done corr
<metadata>
<summary>The Hadoop HBase Thrift Interface</summary>
<description>ThriftServer - this class starts up a Thrift server which implements the Hbase API specified in the Hbase.thrift IDL file. "Thrift is a software framework for scalable cross-language services development. It combines a powerful software stack with a code generation engine to build services that work efficiently and seamlessly between C++, Java, Python, PHP, and Ruby. Thrift was developed at Facebook, and we are now releasing it as open source." For additional information, see http://developers.facebook.com/thrift/. Facebook has announced their intent to migrate Thrift into Apache Incubator.</description>
- <url>http://hadoop.apache.org/hbase/</url>
+ <url>http://hbase.apache.org/</url>
</metadata>
<deps>
<tag name="/bin/bash"/>
Modified: incubator/bigtop/branches/RCs/bigtop-tests/test-artifacts/package/src/main/resources/package_data_yum.xml
URL: http://svn.apache.org/viewvc/incubator/bigtop/branches/RCs/bigtop-tests/test-artifacts/package/src/main/resources/package_data_yum.xml?rev=1241663&r1=1241662&r2=1241663&view=diff
==============================================================================
--- incubator/bigtop/branches/RCs/bigtop-tests/test-artifacts/package/src/main/resources/package_data_yum.xml (original)
+++ incubator/bigtop/branches/RCs/bigtop-tests/test-artifacts/package/src/main/resources/package_data_yum.xml Tue Feb 7 22:21:49 2012
@@ -156,7 +156,7 @@ also on potential use cases. Come to the
service.
* Smart defaults for services. You can get a properly configured system
running quickly, while still being able to override settings as needed.</description>
- <url>http://incubator.apache.org/whirr</url>
+ <url>http://whirr.apache.org/</url>
</metadata>
<deps>
<tag name="/bin/bash"/>
@@ -1060,7 +1060,7 @@ also on potential use cases. Come to the
mechanisms and many failover and recovery mechanisms. The system is centrally
managed and allows for intelligent dynamic management. It uses a simple
extensible data model that allows for online analytic applications.</description>
- <url>https://github.com/cloudera/flume</url>
+ <url>http://incubator.apache.org/projects/flume.html</url>
</metadata>
<deps>
<tag name="/bin/sh"/>
@@ -1369,46 +1369,11 @@ also on potential use cases. Come to the
</flume-conf>
</alternatives>
</flume>
- <flume-master>
- <metadata>
- <summary>The flume master daemon is the central administration and data path control point for flume nodes.</summary>
- <description>Flume is a reliable, scalable, and manageable distributed data collection application for collecting data such as logs and delivering it to data stores such as Hadoop's HDFS. It can efficiently collect, aggregate, and move large amounts of log data. It has a simple, but flexible, architecture based on streaming data flows. It is robust and fault tolerant with tunable reliability mechanisms and many failover and recovery mechanisms. The system is centrally managed and allows for intelligent dynamic management. It uses a simple extensible data model that allows for online analytic applications.</description>
- <url>https://github.com/cloudera/flume</url>
- </metadata>
- <deps>
- <tag name="/bin/bash"/>
- <tag name="/bin/sh"/>
- <tag name="/sbin/chkconfig"/>
- <tag name="/sbin/service"/>
- <tag name="/usr/bin/env"/>
- <tag name="/usr/sbin/useradd"/>
- <flume>/self</flume>
- <jre>>=1.6</jre>
- <redhat-lsb/>
- <sh-utils/>
- <textutils/>
- </deps>
- <services>
- <flume-master>
- <runlevel>2</runlevel>
- <runlevel>3</runlevel>
- <runlevel>4</runlevel>
- <runlevel>5</runlevel>
- <oninstall>stop</oninstall>
- <configured>true</configured>
- </flume-master>
- </services>
- <content>
- <file name="/etc/rc.d/init.d/flume-master" perm="-rwxr-xr-x" group="root" owners="-1" user="root"/>
- <file name="/usr/lib/flume" perm="drwxr-xr-x" group="root" owners="-1" user="root"/>
- <file name="/usr/lib/flume/bin" perm="drwxr-xr-x" group="root" owners="-1" user="root"/>
- </content>
- </flume-master>
<flume-node>
<metadata>
<summary>The flume node daemon is a core element of flume's data path and is responsible for generating, processing, and delivering data.</summary>
<description>Flume is a reliable, scalable, and manageable distributed data collection application for collecting data such as logs and delivering it to data stores such as Hadoop's HDFS. It can efficiently collect, aggregate, and move large amounts of log data. It has a simple, but flexible, architecture based on streaming data flows. It is robust and fault tolerant with tunable reliability mechanisms and many failover and recovery mechanisms. The system is centrally managed and allows for intelligent dynamic management. It uses a simple extensible data model that allows for online analytic applications.</description>
- <url>https://github.com/cloudera/flume</url>
+ <url>http://incubator.apache.org/projects/flume.html</url>
</metadata>
<deps>
<tag name="/bin/bash"/>
@@ -1685,7 +1650,7 @@ also on potential use cases. Come to the
In case of workflow job failure, the workflow job can be rerun skipping
previously completed actions, the workflow application can be patched before
being rerun.</description>
- <url>http://www.cloudera.com</url>
+ <url>http://incubator.apache.org/oozie/</url>
</metadata>
<deps>
<tag name="/bin/bash"/>
@@ -1815,7 +1780,7 @@ also on potential use cases. Come to the
you can also change the status of the entire system, get vesion
information. This client utility also allows you to validate
any worflows before they are deployed to the Oozie server.</description>
- <url>http://www.cloudera.com</url>
+ <url>http://incubator.apache.org/oozie/</url>
</metadata>
<deps>
<tag name="/bin/bash"/>
@@ -4171,7 +4136,7 @@ into fixing the bugs and race conditions
difficulty of implementing these kinds of services, applications initially
usually skimp on them ,which make them brittle in the presence of change and
difficult to manage. Even when done correctly, different implementations of these services lead to management complexity when the applications are deployed.</description>
- <url>http://hadoop.apache.org/zookeeper/</url>
+ <url>http://zookeeper.apache.org/</url>
</metadata>
<deps>
<tag name="/bin/sh"/>
@@ -4550,7 +4515,7 @@ difficult to manage. Even when done corr
<metadata>
<summary>The Hadoop Zookeeper server</summary>
<description>This package starts the zookeeper server on startup</description>
- <url>http://hadoop.apache.org/zookeeper/</url>
+ <url>http://zookeeper.apache.org/</url>
</metadata>
<deps>
<tag name="/bin/bash"/>
@@ -4595,7 +4560,7 @@ difficult to manage. Even when done corr
automatically, allowing the user to focus on semantics rather than efficiency.
* Extensibility
Users can create their own functions to do special-purpose processing.</description>
- <url>http://hadoop.apache.org/pig/</url>
+ <url>http://pig.apache.org/</url>
</metadata>
<deps>
<tag name="/bin/sh"/>
@@ -8471,7 +8436,7 @@ difficult to manage. Even when done corr
plug in their custom mappers and reducers to do more sophisticated
analysis which may not be supported by the built-in capabilities of
the language.</description>
- <url>http://hadoop.apache.org/hive/</url>
+ <url>http://hive.apache.org/</url>
</metadata>
<deps>
<tag name="/bin/sh"/>
@@ -8927,7 +8892,7 @@ difficult to manage. Even when done corr
* Cascading source and sink modules
* Extensible jruby-based (JIRB) shell
* Support for exporting metrics via the Hadoop metrics subsystem to files or Ganglia; or via JMX</description>
- <url>http://hadoop.apache.org/hbase/</url>
+ <url>http://hbase.apache.org/</url>
</metadata>
<deps>
<tag name="/bin/sh"/>
@@ -9114,7 +9079,7 @@ difficult to manage. Even when done corr
<metadata>
<summary>Hbase Documentation</summary>
<description>Documentation for Hbase</description>
- <url>http://hadoop.apache.org/hbase/</url>
+ <url>http://hbase.apache.org/</url>
</metadata>
<content>
<file name="/usr/share/doc/hbase-0.90.4.16" perm="drwxr-xr-x" group="root" owners="-1" user="root"/>
@@ -11842,7 +11807,7 @@ difficult to manage. Even when done corr
<metadata>
<summary>The Hadoop HBase master Server.</summary>
<description>HMaster is the "master server" for a HBase. There is only one HMaster for a single HBase deployment.</description>
- <url>http://hadoop.apache.org/hbase/</url>
+ <url>http://hbase.apache.org/</url>
</metadata>
<deps>
<tag name="/bin/bash"/>
@@ -11868,7 +11833,7 @@ difficult to manage. Even when done corr
<metadata>
<summary>The Hadoop HBase RegionServer server.</summary>
<description>HRegionServer makes a set of HRegions available to clients. It checks in with the HMaster. There are many HRegionServers in a single HBase deployment.</description>
- <url>http://hadoop.apache.org/hbase/</url>
+ <url>http://hbase.apache.org/</url>
</metadata>
<deps>
<tag name="/bin/bash"/>
@@ -11894,7 +11859,7 @@ difficult to manage. Even when done corr
<metadata>
<summary>The Hadoop HBase Thrift Interface</summary>
<description>ThriftServer - this class starts up a Thrift server which implements the Hbase API specified in the Hbase.thrift IDL file. "Thrift is a software framework for scalable cross-language services development. It combines a powerful software stack with a code generation engine to build services that work efficiently and seamlessly between C++, Java, Python, PHP, and Ruby. Thrift was developed at Facebook, and we are now releasing it as open source." For additional information, see http://developers.facebook.com/thrift/. Facebook has announced their intent to migrate Thrift into Apache Incubator.</description>
- <url>http://hadoop.apache.org/hbase/</url>
+ <url>http://hbase.apache.org/</url>
</metadata>
<deps>
<tag name="/bin/bash"/>
Modified: incubator/bigtop/branches/RCs/bigtop-tests/test-artifacts/package/src/main/resources/package_data_zypper.xml
URL: http://svn.apache.org/viewvc/incubator/bigtop/branches/RCs/bigtop-tests/test-artifacts/package/src/main/resources/package_data_zypper.xml?rev=1241663&r1=1241662&r2=1241663&view=diff
==============================================================================
--- incubator/bigtop/branches/RCs/bigtop-tests/test-artifacts/package/src/main/resources/package_data_zypper.xml (original)
+++ incubator/bigtop/branches/RCs/bigtop-tests/test-artifacts/package/src/main/resources/package_data_zypper.xml Tue Feb 7 22:21:49 2012
@@ -63,7 +63,7 @@ also on potential use cases. Come to the
service.
* Smart defaults for services. You can get a properly configured system
running quickly, while still being able to override settings as needed.</description>
- <url>http://incubator.apache.org/whirr</url>
+ <url>http://whirr.apache.org/</url>
</metadata>
<deps>
<jre>>=1.6</jre>
@@ -804,7 +804,7 @@ also on potential use cases. Come to the
mechanisms and many failover and recovery mechanisms. The system is centrally
managed and allows for intelligent dynamic management. It uses a simple
extensible data model that allows for online analytic applications.</description>
- <url>https://github.com/cloudera/flume</url>
+ <url>http://incubator.apache.org/projects/flume.html</url>
</metadata>
<deps>
<sh-utils/>
@@ -1117,44 +1117,11 @@ also on potential use cases. Come to the
</flume-conf>
</alternatives>
</flume>
-<flume-master>
- <metadata>
- <summary>The flume master daemon is the central administration and data path control point for flume nodes.</summary>
- <description>Flume is a reliable, scalable, and manageable distributed data collection application for collecting data such as logs and delivering it to data stores such as Hadoop's HDFS. It can efficiently collect, aggregate, and move large amounts of log data. It has a simple, but flexible, architecture based on streaming data flows. It is robust and fault tolerant with tunable reliability mechanisms and many failover and recovery mechanisms. The system is centrally managed and allows for intelligent dynamic management. It uses a simple extensible data model that allows for online analytic applications.</description>
- <url>https://github.com/cloudera/flume</url>
- </metadata>
- <deps>
- <flume>/self</flume>
- <sh-utils/>
- <textutils/>
- <tag name="/usr/sbin/useradd"/>
- <tag name="/sbin/chkconfig"/>
- <tag name="/sbin/service"/>
- <jre>>=1.6</jre>
- <insserv/>
- <tag name="/bin/sh"/>
- <tag name="/bin/bash"/>
- <tag name="/usr/bin/env"/>
- </deps>
- <services>
- <flume-master>
- <!-- BUG https://jira.cloudera.com/browse/KITCHEN-1095 --><runlevel>3</runlevel><runlevel>4</runlevel><runlevel>5</runlevel>
- <oninstall>stop</oninstall>
- <configured>true</configured>
- </flume-master>
- </services>
- <content>
- <file name="/etc/rc.d/flume-master" owners="1" perm="-rwxr-xr-x" user="root" group="root"/>
- <file name="/usr/lib/flume" owners="2" perm="drwxr-xr-x" user="root" group="root"/>
- <file name="/usr/lib/flume/bin" owners="2" perm="drwxr-xr-x" user="root" group="root"/>
- <file name="/usr/lib/flume/bin/flume-daemon.sh" owners="2" perm="-rwxr-xr-x" user="root" group="root"/>
- </content>
-</flume-master>
<flume-node>
<metadata>
<summary>The flume node daemon is a core element of flume's data path and is responsible for generating, processing, and delivering data.</summary>
<description>Flume is a reliable, scalable, and manageable distributed data collection application for collecting data such as logs and delivering it to data stores such as Hadoop's HDFS. It can efficiently collect, aggregate, and move large amounts of log data. It has a simple, but flexible, architecture based on streaming data flows. It is robust and fault tolerant with tunable reliability mechanisms and many failover and recovery mechanisms. The system is centrally managed and allows for intelligent dynamic management. It uses a simple extensible data model that allows for online analytic applications.</description>
- <url>https://github.com/cloudera/flume</url>
+ <url>http://incubator.apache.org/projects/flume.html</url>
</metadata>
<deps>
<flume>/self</flume>
@@ -1418,7 +1385,7 @@ also on potential use cases. Come to the
In case of workflow job failure, the workflow job can be rerun skipping
previously completed actions, the workflow application can be patched before
being rerun.</description>
- <url>http://www.cloudera.com</url>
+ <url>http://incubator.apache.org/oozie/</url>
</metadata>
<deps>
<tag name="/usr/sbin/groupadd"/>
@@ -1543,7 +1510,7 @@ also on potential use cases. Come to the
you can also change the status of the entire system, get vesion
information. This client utility also allows you to validate
any worflows before they are deployed to the Oozie server.</description>
- <url>http://www.cloudera.com</url>
+ <url>http://incubator.apache.org/oozie/</url>
</metadata>
<deps>
<tag name="/bin/bash"/>
@@ -2356,7 +2323,7 @@ into fixing the bugs and race conditions
difficulty of implementing these kinds of services, applications initially
usually skimp on them ,which make them brittle in the presence of change and
difficult to manage. Even when done correctly, different implementations of these services lead to management complexity when the applications are deployed.</description>
- <url>http://hadoop.apache.org/zookeeper/</url>
+ <url>http://zookeeper.apache.org/</url>
<!-- group>misc</group -->
</metadata>
<deps>
@@ -2728,9 +2695,9 @@ difficult to manage. Even when done corr
<alternatives>
<zookeeper-conf>
<status>auto</status>
- <link>/etc/zookeeper</link>
- <value>/etc/zookeeper.dist</value>
- <alt>/etc/zookeeper.dist</alt>
+ <link>/etc/zookeeper/conf</link>
+ <value>/etc/zookeeper/conf.dist</value>
+ <alt>/etc/zookeeper/conf.dist</alt>
</zookeeper-conf>
</alternatives>
</zookeeper>
@@ -2740,7 +2707,7 @@ difficult to manage. Even when done corr
<!-- arch>universal</arch -->
<summary>The Hadoop Zookeeper server</summary>
<description>This package starts the zookeeper server on startup</description>
- <url>http://hadoop.apache.org/zookeeper/</url>
+ <url>http://zookeeper.apache.org/</url>
<!-- vendor>(none)</vendor -->
<!-- group>Development/Libraries</group -->
<!-- depends><dep>adduser</dep><dep>sun-java6-jre</dep><dep>sun-java6-bin</dep></depends -->
@@ -2788,7 +2755,7 @@ difficult to manage. Even when done corr
automatically, allowing the user to focus on semantics rather than efficiency.
* Extensibility
Users can create their own functions to do special-purpose processing.</description>
- <url>http://hadoop.apache.org/pig/</url>
+ <url>http://pig.apache.org/</url>
</metadata>
<deps>
<hadoop/>
@@ -6696,12 +6663,12 @@ difficult to manage. Even when done corr
<doc name="/usr/share/man/man1/pig.1.gz" owners="1" perm="-rw-r--r--" user="root" group="root"/>
</content>
<alternatives>
- <pig> <!-- BUG: https://issues.cloudera.org/browse/DISTRO-223 -->
+ <pig-conf>
<status>auto</status>
<link>/etc/pig/conf</link>
<value>/etc/pig/conf.dist</value>
<alt>/etc/pig/conf.dist</alt>
- </pig>
+ </pig-conf>
</alternatives>
</pig>
<hive>
@@ -6717,7 +6684,7 @@ difficult to manage. Even when done corr
plug in their custom mappers and reducers to do more sophisticated
analysis which may not be supported by the built-in capabilities of
the language.</description>
- <url>http://hadoop.apache.org/hive/</url>
+ <url>http://hive.apache.org/</url>
</metadata>
<deps>
<hadoop>>=0.20.1</hadoop>
@@ -7077,7 +7044,7 @@ difficult to manage. Even when done corr
* Cascading source and sink modules
* Extensible jruby-based (JIRB) shell
* Support for exporting metrics via the Hadoop metrics subsystem to files or Ganglia; or via JMX</description>
- <url>http://hadoop.apache.org/hbase/</url>
+ <url>http://hbase.apache.org/</url>
<!-- group>misc</group -->
</metadata>
<deps>
@@ -7256,7 +7223,7 @@ difficult to manage. Even when done corr
<metadata>
<summary>Hbase Documentation</summary>
<description>Documentation for Hbase</description>
- <url>http://hadoop.apache.org/hbase/</url>
+ <url>http://hbase.apache.org/</url>
<!-- group>misc</group -->
</metadata>
<content>
@@ -9025,7 +8992,7 @@ difficult to manage. Even when done corr
<metadata>
<summary>The Hadoop HBase master Server.</summary>
<description>HMaster is the "master server" for a HBase. There is only one HMaster for a single HBase deployment.</description>
- <url>http://hadoop.apache.org/hbase/</url>
+ <url>http://hbase.apache.org/</url>
</metadata>
<deps>
<hbase>/self</hbase>
@@ -9048,7 +9015,7 @@ difficult to manage. Even when done corr
<metadata>
<summary>The Hadoop HBase RegionServer server.</summary>
<description>HRegionServer makes a set of HRegions available to clients. It checks in with the HMaster. There are many HRegionServers in a single HBase deployment.</description>
- <url>http://hadoop.apache.org/hbase/</url>
+ <url>http://hbase.apache.org/</url>
</metadata>
<deps>
<hbase>/self</hbase>
@@ -9071,7 +9038,7 @@ difficult to manage. Even when done corr
<metadata>
<summary>The Hadoop HBase Thrift Interface</summary>
<description>ThriftServer - this class starts up a Thrift server which implements the Hbase API specified in the Hbase.thrift IDL file. "Thrift is a software framework for scalable cross-language services development. It combines a powerful software stack with a code generation engine to build services that work efficiently and seamlessly between C++, Java, Python, PHP, and Ruby. Thrift was developed at Facebook, and we are now releasing it as open source." For additional information, see http://developers.facebook.com/thrift/. Facebook has announced their intent to migrate Thrift into Apache Incubator.</description>
- <url>http://hadoop.apache.org/hbase/</url>
+ <url>http://hbase.apache.org/</url>
</metadata>
<deps>
<hbase>/self</hbase>