You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by cu...@apache.org on 2007/01/31 23:34:48 UTC
svn commit: r502024 - in /lucene/hadoop/trunk: CHANGES.txt build.xml
src/contrib/ec2/README.txt src/contrib/ec2/bin/hadoop-ec2-env.sh
src/contrib/ec2/bin/hadoop-ec2-env.sh.template
Author: cutting
Date: Wed Jan 31 14:34:48 2007
New Revision: 502024
URL: http://svn.apache.org/viewvc?view=rev&rev=502024
Log:
HADOOP-962. In contrib/ec2: make scripts executable, etc. Contributed by Tom White.
Added:
lucene/hadoop/trunk/src/contrib/ec2/README.txt
lucene/hadoop/trunk/src/contrib/ec2/bin/hadoop-ec2-env.sh.template
Removed:
lucene/hadoop/trunk/src/contrib/ec2/bin/hadoop-ec2-env.sh
Modified:
lucene/hadoop/trunk/CHANGES.txt
lucene/hadoop/trunk/build.xml
Modified: lucene/hadoop/trunk/CHANGES.txt
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/CHANGES.txt?view=diff&rev=502024&r1=502023&r2=502024
==============================================================================
--- lucene/hadoop/trunk/CHANGES.txt (original)
+++ lucene/hadoop/trunk/CHANGES.txt Wed Jan 31 14:34:48 2007
@@ -118,6 +118,10 @@
permitting it to take advantage of native compression facilities.
(Sanjay Dahiya via cutting)
+37. HADOOP-962. In contrib/ec2: make scripts executable in tar file;
+ add a README; make the environment file use a template.
+ (Tom White via cutting)
+
Release 0.10.1 - 2007-01-10
Modified: lucene/hadoop/trunk/build.xml
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/build.xml?view=diff&rev=502024&r1=502023&r2=502024
==============================================================================
--- lucene/hadoop/trunk/build.xml (original)
+++ lucene/hadoop/trunk/build.xml Wed Jan 31 14:34:48 2007
@@ -20,6 +20,7 @@
<property name="lib.dir" value="${basedir}/lib"/>
<property name="conf.dir" value="${basedir}/conf"/>
<property name="docs.dir" value="${basedir}/docs"/>
+ <property name="contrib.dir" value="${basedir}/src/contrib"/>
<property name="docs.src" value="${basedir}/src/web"/>
<property name="libhdfs.src" value="${basedir}/src/c++/libhdfs"/>
@@ -111,6 +112,7 @@
<mkdir dir="${test.build.testjar}"/>
<touch datetime="01/25/1971 2:00 pm">
<fileset dir="${conf.dir}" includes="**/*.template"/>
+ <fileset dir="${contrib.dir}" includes="**/*.template"/>
</touch>
<!-- copy all of the jsp and static files -->
@@ -125,6 +127,11 @@
<mapper type="glob" from="*.template" to="*"/>
</copy>
+ <copy todir="${contrib.dir}" verbose="true">
+ <fileset dir="${contrib.dir}" includes="**/*.template"/>
+ <mapper type="glob" from="*.template" to="*"/>
+ </copy>
+
<exec executable="sh">
<arg line="src/saveVersion.sh ${version}"/>
</exec>
@@ -266,7 +273,7 @@
<target name="compile-contrib" depends="compile-core">
<subant target="deploy">
- <fileset file="${basedir}/src/contrib/build.xml"/>
+ <fileset file="${contrib.dir}/build.xml"/>
</subant>
</target>
@@ -449,7 +456,7 @@
<target name="test-contrib" depends="compile-core, compile-core-test">
<subant target="test">
- <fileset file="${basedir}/src/contrib/build.xml"/>
+ <fileset file="${contrib.dir}/build.xml"/>
</subant>
</target>
@@ -561,10 +568,6 @@
<fileset dir="${conf.dir}" excludes="**/*.template"/>
</copy>
- <chmod perm="ugo+x" type="file">
- <fileset dir="${dist.dir}/bin"/>
- </chmod>
-
<copy todir="${dist.dir}/docs">
<fileset dir="${build.docs}"/>
</copy>
@@ -576,11 +579,17 @@
</copy>
<copy todir="${dist.dir}/src" includeEmptyDirs="true">
- <fileset dir="src"/>
+ <fileset dir="src" excludes="**/*.template"/>
</copy>
-
+
<copy todir="${dist.dir}/" file="build.xml"/>
+ <chmod perm="ugo+x" type="file">
+ <fileset dir="${dist.dir}/bin"/>
+ <fileset dir="${dist.dir}/src/contrib/ec2/bin"/>
+ <fileset dir="${dist.dir}/src/contrib/ec2/bin/image"/>
+ </chmod>
+
</target>
<!-- ================================================================== -->
@@ -590,11 +599,15 @@
<tar compression="gzip" longfile="gnu"
destfile="${build.dir}/${final.name}.tar.gz">
<tarfileset dir="${build.dir}" mode="664">
- <exclude name="${final.name}/bin/*" />
+ <exclude name="${final.name}/bin/*" />
+ <exclude name="${final.name}/src/contrib/ec2/bin/*" />
+ <exclude name="${final.name}/src/contrib/ec2/bin/image/*" />
<include name="${final.name}/**" />
</tarfileset>
<tarfileset dir="${build.dir}" mode="755">
<include name="${final.name}/bin/*" />
+ <include name="${final.name}/src/contrib/ec2/bin/*" />
+ <include name="${final.name}/src/contrib/ec2/bin/image/*" />
</tarfileset>
</tar>
</target>
Added: lucene/hadoop/trunk/src/contrib/ec2/README.txt
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/ec2/README.txt?view=auto&rev=502024
==============================================================================
--- lucene/hadoop/trunk/src/contrib/ec2/README.txt (added)
+++ lucene/hadoop/trunk/src/contrib/ec2/README.txt Wed Jan 31 14:34:48 2007
@@ -0,0 +1,9 @@
+Hadoop EC2
+
+This collection of scripts allows you to run Hadoop clusters on Amazon.com's Elastic Compute Cloud (EC2) service described at:
+
+ http://aws.amazon.com/ec2
+
+For full instructions, please visit the Hadoop wiki at:
+
+ http://wiki.apache.org/lucene-hadoop/AmazonEC2#AutomatedScripts
\ No newline at end of file
Added: lucene/hadoop/trunk/src/contrib/ec2/bin/hadoop-ec2-env.sh.template
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/ec2/bin/hadoop-ec2-env.sh.template?view=auto&rev=502024
==============================================================================
--- lucene/hadoop/trunk/src/contrib/ec2/bin/hadoop-ec2-env.sh.template (added)
+++ lucene/hadoop/trunk/src/contrib/ec2/bin/hadoop-ec2-env.sh.template Wed Jan 31 14:34:48 2007
@@ -0,0 +1,40 @@
+# Set environment variables for running Hadoop on Amazon EC2 here. All are required.
+
+# Your Amazon Account Number.
+AWS_ACCOUNT_ID=
+
+# Your Amazon AWS access key.
+AWS_ACCESS_KEY_ID=
+
+# Your Amazon AWS secret access key.
+AWS_SECRET_ACCESS_KEY=
+
+# The Amazon S3 bucket where the Hadoop AMI you create will be stored.
+S3_BUCKET=
+
+# SSH options used when connecting to EC2 instances.
+# Change the -i option to be the absolute path to your keypair that you set up in the Amazon Getting Started guide.
+SSH_OPTS='-i /home/<yourname>/id_rsa-gsg-keypair -o StrictHostKeyChecking=no'
+
+# Location of EC2 keys.
+# The default setting is probably OK if you set up EC2 following the Amazon Getting Started guide.
+EC2_KEYDIR=~/.ec2
+
+# The download URL for the Sun JDK. Visit http://java.sun.com/javase/downloads/index_jdk5.jsp and get the URL for the "Linux self-extracting file".
+JAVA_BINARY_URL=''
+
+# The version number of the installed JDK.
+JAVA_VERSION=1.5.0_11
+
+# The EC2 group to run your cluster in.
+GROUP=hadoop-cluster-group
+
+# The version of Hadoop to install.
+HADOOP_VERSION=0.10.1
+
+# The hostname of the master node in the cluster. You need to be able to set the DNS for this host to point to the master's IP address.
+# See http://www.dyndns.com/services/dns/dyndns/, for example.
+MASTER_HOST=
+
+# The number of nodes in your cluster.
+NO_INSTANCES=2