You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by cu...@apache.org on 2007/01/31 23:34:48 UTC

svn commit: r502024 - in /lucene/hadoop/trunk: CHANGES.txt build.xml src/contrib/ec2/README.txt src/contrib/ec2/bin/hadoop-ec2-env.sh src/contrib/ec2/bin/hadoop-ec2-env.sh.template

Author: cutting
Date: Wed Jan 31 14:34:48 2007
New Revision: 502024

URL: http://svn.apache.org/viewvc?view=rev&rev=502024
Log:
HADOOP-962.  In contrib/ec2: make scripts executable, etc.  Contributed by Tom White.

Added:
    lucene/hadoop/trunk/src/contrib/ec2/README.txt
    lucene/hadoop/trunk/src/contrib/ec2/bin/hadoop-ec2-env.sh.template
Removed:
    lucene/hadoop/trunk/src/contrib/ec2/bin/hadoop-ec2-env.sh
Modified:
    lucene/hadoop/trunk/CHANGES.txt
    lucene/hadoop/trunk/build.xml

Modified: lucene/hadoop/trunk/CHANGES.txt
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/CHANGES.txt?view=diff&rev=502024&r1=502023&r2=502024
==============================================================================
--- lucene/hadoop/trunk/CHANGES.txt (original)
+++ lucene/hadoop/trunk/CHANGES.txt Wed Jan 31 14:34:48 2007
@@ -118,6 +118,10 @@
     permitting it to take advantage of native compression facilities.
     (Sanjay Dahiya via cutting)
 
+37. HADOOP-962.  In contrib/ec2: make scripts executable in tar file;
+    add a README; make the environment file use a template.
+    (Tom White via cutting)
+
 
 Release 0.10.1 - 2007-01-10
 

Modified: lucene/hadoop/trunk/build.xml
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/build.xml?view=diff&rev=502024&r1=502023&r2=502024
==============================================================================
--- lucene/hadoop/trunk/build.xml (original)
+++ lucene/hadoop/trunk/build.xml Wed Jan 31 14:34:48 2007
@@ -20,6 +20,7 @@
   <property name="lib.dir" value="${basedir}/lib"/>
   <property name="conf.dir" value="${basedir}/conf"/>
   <property name="docs.dir" value="${basedir}/docs"/>
+  <property name="contrib.dir" value="${basedir}/src/contrib"/>
   <property name="docs.src" value="${basedir}/src/web"/>
   <property name="libhdfs.src" value="${basedir}/src/c++/libhdfs"/>
 
@@ -111,6 +112,7 @@
     <mkdir dir="${test.build.testjar}"/>
     <touch datetime="01/25/1971 2:00 pm">
       <fileset dir="${conf.dir}" includes="**/*.template"/>
+      <fileset dir="${contrib.dir}" includes="**/*.template"/>
     </touch>
 
     <!-- copy all of the jsp and static files -->
@@ -125,6 +127,11 @@
       <mapper type="glob" from="*.template" to="*"/>
     </copy>
 
+    <copy todir="${contrib.dir}" verbose="true">
+      <fileset dir="${contrib.dir}" includes="**/*.template"/>
+      <mapper type="glob" from="*.template" to="*"/>
+    </copy>
+
     <exec executable="sh">
        <arg line="src/saveVersion.sh ${version}"/>
     </exec>
@@ -266,7 +273,7 @@
 
   <target name="compile-contrib" depends="compile-core">
      <subant target="deploy">
-        <fileset file="${basedir}/src/contrib/build.xml"/>
+        <fileset file="${contrib.dir}/build.xml"/>
      </subant>  	
   </target>
   
@@ -449,7 +456,7 @@
 
   <target name="test-contrib" depends="compile-core, compile-core-test">
     <subant target="test">
-       <fileset file="${basedir}/src/contrib/build.xml"/>
+       <fileset file="${contrib.dir}/build.xml"/>
     </subant> 
   </target>   
   
@@ -561,10 +568,6 @@
       <fileset dir="${conf.dir}" excludes="**/*.template"/>
     </copy>
 
-    <chmod perm="ugo+x" type="file">
-        <fileset dir="${dist.dir}/bin"/>
-    </chmod>
-
     <copy todir="${dist.dir}/docs">
       <fileset dir="${build.docs}"/>
     </copy>
@@ -576,11 +579,17 @@
     </copy>
 
     <copy todir="${dist.dir}/src" includeEmptyDirs="true">
-      <fileset dir="src"/>
+      <fileset dir="src" excludes="**/*.template"/>
     </copy>
-
+  	
     <copy todir="${dist.dir}/" file="build.xml"/>
 
+    <chmod perm="ugo+x" type="file">
+        <fileset dir="${dist.dir}/bin"/>
+        <fileset dir="${dist.dir}/src/contrib/ec2/bin"/>
+        <fileset dir="${dist.dir}/src/contrib/ec2/bin/image"/>
+    </chmod>
+
   </target>
 
   <!-- ================================================================== -->
@@ -590,11 +599,15 @@
     <tar compression="gzip" longfile="gnu"
       destfile="${build.dir}/${final.name}.tar.gz">
       <tarfileset dir="${build.dir}" mode="664">
-	<exclude name="${final.name}/bin/*" />
+        <exclude name="${final.name}/bin/*" />
+        <exclude name="${final.name}/src/contrib/ec2/bin/*" />
+        <exclude name="${final.name}/src/contrib/ec2/bin/image/*" />
         <include name="${final.name}/**" />
       </tarfileset>
       <tarfileset dir="${build.dir}" mode="755">
         <include name="${final.name}/bin/*" />
+        <include name="${final.name}/src/contrib/ec2/bin/*" />
+        <include name="${final.name}/src/contrib/ec2/bin/image/*" />
       </tarfileset>
     </tar>
   </target>

Added: lucene/hadoop/trunk/src/contrib/ec2/README.txt
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/ec2/README.txt?view=auto&rev=502024
==============================================================================
--- lucene/hadoop/trunk/src/contrib/ec2/README.txt (added)
+++ lucene/hadoop/trunk/src/contrib/ec2/README.txt Wed Jan 31 14:34:48 2007
@@ -0,0 +1,9 @@
+Hadoop EC2
+
+This collection of scripts allows you to run Hadoop clusters on Amazon.com's Elastic Compute Cloud (EC2) service described at:
+
+  http://aws.amazon.com/ec2
+
+For full instructions, please visit the Hadoop wiki at:
+
+  http://wiki.apache.org/lucene-hadoop/AmazonEC2#AutomatedScripts
\ No newline at end of file

Added: lucene/hadoop/trunk/src/contrib/ec2/bin/hadoop-ec2-env.sh.template
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/ec2/bin/hadoop-ec2-env.sh.template?view=auto&rev=502024
==============================================================================
--- lucene/hadoop/trunk/src/contrib/ec2/bin/hadoop-ec2-env.sh.template (added)
+++ lucene/hadoop/trunk/src/contrib/ec2/bin/hadoop-ec2-env.sh.template Wed Jan 31 14:34:48 2007
@@ -0,0 +1,40 @@
+# Set environment variables for running Hadoop on Amazon EC2 here. All are required.
+
+# Your Amazon Account Number.
+AWS_ACCOUNT_ID=
+
+# Your Amazon AWS access key.
+AWS_ACCESS_KEY_ID=
+
+# Your Amazon AWS secret access key.
+AWS_SECRET_ACCESS_KEY=
+
+# The Amazon S3 bucket where the Hadoop AMI you create will be stored.
+S3_BUCKET=
+
+# SSH options used when connecting to EC2 instances.
+# Change the -i option to be the absolute path to your keypair that you set up in the Amazon Getting Started guide.
+SSH_OPTS='-i /home/<yourname>/id_rsa-gsg-keypair -o StrictHostKeyChecking=no'
+
+# Location of EC2 keys.
+# The default setting is probably OK if you set up EC2 following the Amazon Getting Started guide.
+EC2_KEYDIR=~/.ec2
+
+# The download URL for the Sun JDK. Visit http://java.sun.com/javase/downloads/index_jdk5.jsp and get the URL for the "Linux self-extracting file".
+JAVA_BINARY_URL=''
+
+# The version number of the installed JDK.
+JAVA_VERSION=1.5.0_11
+
+# The EC2 group to run your cluster in.
+GROUP=hadoop-cluster-group
+
+# The version of Hadoop to install.
+HADOOP_VERSION=0.10.1
+
+# The hostname of the master node in the cluster. You need to be able to set the DNS for this host to point to the master's IP address.
+# See http://www.dyndns.com/services/dns/dyndns/, for example.
+MASTER_HOST=
+
+# The number of nodes in your cluster.
+NO_INSTANCES=2