You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-commits@hadoop.apache.org by to...@apache.org on 2010/11/10 19:50:05 UTC
svn commit: r1033641 - in /hadoop/hdfs/branches/branch-0.21: CHANGES.txt
build.xml ivy/libraries.properties
Author: tomwhite
Date: Wed Nov 10 18:50:04 2010
New Revision: 1033641
URL: http://svn.apache.org/viewvc?rev=1033641&view=rev
Log:
Merge -r 1033638:1033639 from trunk to branch-0.21. Fixes: HDFS-1292
Modified:
hadoop/hdfs/branches/branch-0.21/CHANGES.txt
hadoop/hdfs/branches/branch-0.21/build.xml
hadoop/hdfs/branches/branch-0.21/ivy/libraries.properties
Modified: hadoop/hdfs/branches/branch-0.21/CHANGES.txt
URL: http://svn.apache.org/viewvc/hadoop/hdfs/branches/branch-0.21/CHANGES.txt?rev=1033641&r1=1033640&r2=1033641&view=diff
==============================================================================
--- hadoop/hdfs/branches/branch-0.21/CHANGES.txt (original)
+++ hadoop/hdfs/branches/branch-0.21/CHANGES.txt Wed Nov 10 18:50:04 2010
@@ -24,6 +24,9 @@ Release 0.21.1 - Unreleased
HDFS-1474. ant binary-system is broken (cos)
+ HDFS-1292. Allow artifacts to be published to the staging Apache Nexus
+ Maven Repository. (Giridharan Kesavan via tomwhite)
+
Release 0.21.0 - 2010-08-13
INCOMPATIBLE CHANGES
Modified: hadoop/hdfs/branches/branch-0.21/build.xml
URL: http://svn.apache.org/viewvc/hadoop/hdfs/branches/branch-0.21/build.xml?rev=1033641&r1=1033640&r2=1033641&view=diff
==============================================================================
--- hadoop/hdfs/branches/branch-0.21/build.xml (original)
+++ hadoop/hdfs/branches/branch-0.21/build.xml Wed Nov 10 18:50:04 2010
@@ -141,6 +141,10 @@
<loadproperties srcfile="${ivy.dir}/libraries.properties"/>
<property name="ivy.jar" location="${ivy.dir}/ivy-${ivy.version}.jar"/>
<property name="mvn.repo" value="http://repo2.maven.org/maven2"/>
+ <property name="asfrepo" value="https://repository.apache.org"/>
+ <property name="asfsnapshotrepo" value="${asfrepo}/content/repositories/snapshots"/>
+ <property name="asfstagingrepo"
+ value="${asfrepo}/service/local/staging/deploy/maven2"/>
<property name="ivy_repo_url" value="${mvn.repo}/org/apache/ivy/ivy/${ivy.version}/ivy-${ivy.version}.jar"/>
<property name="ant_task.jar" location="${ivy.dir}/maven-ant-tasks-${ant-task.version}.jar"/>
<property name="ant_task_repo_url" value="${mvn.repo}/org/apache/maven/maven-ant-tasks/${ant-task.version}/maven-ant-tasks-${ant-task.version}.jar"/>
@@ -186,6 +190,10 @@
</and>
</condition>
+ <condition property="staging">
+ <equals arg1="${repo}" arg2="staging"/>
+ </condition>
+
<!-- the normal classpath -->
<path id="classpath">
<pathelement location="${build.classes}"/>
@@ -1160,13 +1168,20 @@
<!-- ================================================================== -->
<!-- Clean. Delete the build files, and their directories -->
<!-- ================================================================== -->
- <target name="clean" depends="clean-contrib, clean-fi" description="Clean. Delete the build files, and their directories">
+ <target name="clean" depends="clean-contrib, clean-fi, clean-sign" description="Clean. Delete the build files, and their directories">
<delete dir="${build.dir}"/>
<delete dir="${build-fi.dir}"/>
<delete dir="${docs.src}/build"/>
<delete file="${hadoop-hdfs.pom}"/>
<delete file="${hadoop-hdfs-test.pom}"/>
<delete file="${hadoop-hdfs-instrumented.pom}"/>
+ <delete file="${hadoop-hdfs-instrumented-test.pom}"/>
+ </target>
+
+ <target name="clean-sign" description="Clean. Delete .asc files">
+ <delete>
+ <fileset dir="." includes="**/**/*.asc"/>
+ </delete>
</target>
<target name="veryclean" depends="clean-cache,clean"
@@ -1423,16 +1438,8 @@
uri="urn:maven-artifact-ant" classpathref="mvn-ant-task.classpath"/>
</target>
- <target name="mvn-install-hdfs" depends="mvn-taskdef,jar,set-version">
- <artifact:pom file="${hadoop-hdfs.pom}" id="hadoop.hdfs"/>
- <artifact:install file="${hadoop-hdfs.jar}">
- <pom refid="hadoop.hdfs"/>
- <attach file="${hadoop-hdfs-sources.jar}" classifier="sources" />
- </artifact:install>
- </target>
-
- <target name="mvn-install" depends="mvn-taskdef,jar,jar-hdfs-test,set-version,
- -mvn-system-install">
+ <target name="mvn-install" depends="mvn-taskdef,jar,jar-test,set-version,-mvn-system-install"
+ description="To install hadoop hdfs and test jars to local filesystem's m2 cache">
<artifact:pom file="${hadoop-hdfs.pom}" id="hadoop.hdfs"/>
<artifact:pom file="${hadoop-hdfs-test.pom}" id="hadoop.hdfs.test"/>
<artifact:install file="${hadoop-hdfs.jar}">
@@ -1443,38 +1450,150 @@
<pom refid="hadoop.hdfs.test"/>
<attach file="${hadoop-hdfs-test-sources.jar}" classifier="sources" />
</artifact:install>
+ </target>
+
+ <target name="mvn-deploy" depends="mvn-taskdef, jar, jar-test,
+ jar-system, jar-test-system, set-version, signanddeploy, simpledeploy"
+ description="To deploy hadoop hdfs and test jar's to apache
+ snapshot's repository"/>
+
+ <target name="signanddeploy" if="staging" depends="sign">
+ <artifact:pom file="${hadoop-hdfs.pom}" id="hadoop.hdfs"/>
+ <artifact:pom file="${hadoop-hdfs-test.pom}" id="hadoop.hdfs.test"/>
+ <artifact:pom file="${hadoop-hdfs-instrumented.pom}"
+ id="hadoop.hdfs.${herriot.suffix}"/>
+ <artifact:pom file="${hadoop-hdfs-instrumented-test.pom}"
+ id="hadoop.hdfs.${herriot.suffix}.test"/>
+ <artifact:install-provider artifactId="wagon-http"
+ version="${wagon-http.version}"/>
+
+ <artifact:deploy file="${hadoop-hdfs.jar}">
+ <remoteRepository id="apache.staging.https" url="${asfstagingrepo}"/>
+ <pom refid="hadoop.hdfs"/>
+ <attach file="${hadoop-hdfs.jar}.asc" type="jar.asc"/>
+ <attach file="${hadoop-hdfs.pom}.asc" type="pom.asc"/>
+ <attach file="${hadoop-hdfs-sources.jar}.asc" type="jar.asc"
+ classifier="sources" />
+ <attach file="${hadoop-hdfs-sources.jar}" classifier="sources"/>
+ </artifact:deploy>
+
+ <artifact:deploy file="${hadoop-hdfs-test.jar}">
+ <remoteRepository id="apache.staging.https" url="${asfstagingrepo}"/>
+ <pom refid="hadoop.hdfs.test"/>
+ <attach file="${hadoop-hdfs-test.jar}.asc" type="jar.asc"/>
+ <attach file="${hadoop-hdfs-test.pom}.asc" type="pom.asc"/>
+ <attach file="${hadoop-hdfs-test-sources.jar}.asc" type="jar.asc"
+ classifier="sources"/>
+ <attach file="${hadoop-hdfs-test-sources.jar}" classifier="sources"/>
+ </artifact:deploy>
+
+ <artifact:deploy file="${hadoop-hdfs-instrumented.jar}">
+ <remoteRepository id="apache.staging.https" url="${asfstagingrepo}"/>
+ <pom refid="hadoop.hdfs.${herriot.suffix}"/>
+ <attach file="${hadoop-hdfs-instrumented.jar}.asc" type="jar.asc"/>
+ <attach file="${hadoop-hdfs-instrumented.pom}.asc" type="pom.asc"/>
+ <attach file="${hadoop-hdfs-instrumented-sources.jar}.asc"
+ type="jar.asc" classifier="sources"/>
+ <attach file="${hadoop-hdfs-instrumented-sources.jar}"
+ classifier="sources"/>
+ </artifact:deploy>
+
+ <artifact:deploy file="${hadoop-hdfs-instrumented-test.jar}">
+ <remoteRepository id="apache.staging.https" url="${asfstagingrepo}"/>
+ <pom refid="hadoop.hdfs.${herriot.suffix}.test"/>
+ <attach file="${hadoop-hdfs-instrumented-test.jar}.asc" type="jar.asc"/>
+ <attach file="${hadoop-hdfs-instrumented-test.pom}.asc" type="pom.asc"/>
+ <attach file="${hadoop-hdfs-instrumented-test-sources.jar}.asc"
+ type="jar.asc" classifier="sources"/>
+ <attach file="${hadoop-hdfs-instrumented-test-sources.jar}"
+ classifier="sources"/>
+ </artifact:deploy>
+ </target>
+
+ <target name="sign" depends="clean-sign" if="staging">
+ <input message="password:>" addproperty="gpg.passphrase">
+ <handler classname="org.apache.tools.ant.input.SecureInputHandler" />
+ </input>
+ <macrodef name="sign-artifact" description="Signs the artifact">
+ <attribute name="input.file"/>
+ <attribute name="output.file" default="@{input.file}.asc"/>
+ <attribute name="gpg.passphrase"/>
+ <sequential>
+ <echo>Signing @{input.file} Sig File: @{output.file}</echo>
+ <exec executable="gpg" >
+ <arg value="--armor"/>
+ <arg value="--output"/>
+ <arg value="@{output.file}"/>
+ <arg value="--passphrase"/>
+ <arg value="@{gpg.passphrase}"/>
+ <arg value="--detach-sig"/>
+ <arg value="@{input.file}"/>
+ </exec>
+ </sequential>
+ </macrodef>
+ <sign-artifact input.file="${hadoop-hdfs.jar}"
+ output.file="${hadoop-hdfs.jar}.asc" gpg.passphrase="${gpg.passphrase}"/>
+ <sign-artifact input.file="${hadoop-hdfs-test.jar}"
+ output.file="${hadoop-hdfs-test.jar}.asc" gpg.passphrase="${gpg.passphrase}"/>
+ <sign-artifact input.file="${hadoop-hdfs-sources.jar}"
+ output.file="${hadoop-hdfs-sources.jar}.asc" gpg.passphrase="${gpg.passphrase}"/>
+ <sign-artifact input.file="${hadoop-hdfs-test-sources.jar}"
+ output.file="${hadoop-hdfs-test-sources.jar}.asc" gpg.passphrase="${gpg.passphrase}"/>
+ <sign-artifact input.file="${hadoop-hdfs.pom}"
+ output.file="${hadoop-hdfs.pom}.asc" gpg.passphrase="${gpg.passphrase}"/>
+ <sign-artifact input.file="${hadoop-hdfs-test.pom}"
+ output.file="${hadoop-hdfs-test.pom}.asc" gpg.passphrase="${gpg.passphrase}"/>
+ <sign-artifact input.file="${hadoop-hdfs-instrumented.jar}"
+ output.file="${hadoop-hdfs-instrumented.jar}.asc" gpg.passphrase="${gpg.passphrase}"/>
+ <sign-artifact input.file="${hadoop-hdfs-instrumented.pom}"
+ output.file="${hadoop-hdfs-instrumented.pom}.asc" gpg.passphrase="${gpg.passphrase}"/>
+ <sign-artifact input.file="${hadoop-hdfs-instrumented-sources.jar}"
+ output.file="${hadoop-hdfs-instrumented-sources.jar}.asc" gpg.passphrase="${gpg.passphrase}"/>
+ <sign-artifact input.file="${hadoop-hdfs-instrumented-test.jar}"
+ output.file="${hadoop-hdfs-instrumented-test.jar}.asc" gpg.passphrase="${gpg.passphrase}"/>
+ <sign-artifact input.file="${hadoop-hdfs-instrumented-test.pom}"
+ output.file="${hadoop-hdfs-instrumented-test.pom}.asc" gpg.passphrase="${gpg.passphrase}"/>
+ <sign-artifact input.file="${hadoop-hdfs-instrumented-test-sources.jar}"
+ output.file="${hadoop-hdfs-instrumented-test-sources.jar}.asc" gpg.passphrase="${gpg.passphrase}"/>
</target>
- <target name="mvn-deploy" depends="mvn-taskdef, jar, jar-hdfs-test, set-version,
- -mvn-system-deploy">
- <property name="repourl" value="https://repository.apache.org/content/repositories/snapshots" />
+ <target name="simpledeploy" unless="staging">
<artifact:pom file="${hadoop-hdfs.pom}" id="hadoop.hdfs"/>
<artifact:pom file="${hadoop-hdfs-test.pom}" id="hadoop.hdfs.test"/>
+ <artifact:pom file="${hadoop-hdfs-instrumented.pom}"
+ id="hadoop.hdfs.${herriot.suffix}"/>
- <artifact:install-provider artifactId="wagon-http" version="1.0-beta-2"/>
+ <artifact:install-provider artifactId="wagon-http" version="${wagon-http.version}"/>
<artifact:deploy file="${hadoop-hdfs.jar}">
- <remoteRepository id="apache.snapshots.https" url="${repourl}"/>
+ <remoteRepository id="apache.snapshots.https" url="${asfsnapshotrepo}"/>
<pom refid="hadoop.hdfs"/>
- <attach file="${hadoop-hdfs-sources.jar}" classifier="sources" />
+ <attach file="${hadoop-hdfs-sources.jar}" classifier="sources" />
</artifact:deploy>
+
<artifact:deploy file="${hadoop-hdfs-test.jar}">
- <remoteRepository id="apache.snapshots.https" url="${repourl}"/>
+ <remoteRepository id="apache.snapshots.https" url="${asfsnapshotrepo}"/>
<pom refid="hadoop.hdfs.test"/>
- <attach file="${hadoop-hdfs-test-sources.jar}" classifier="sources" />
+ <attach file="${hadoop-hdfs-test-sources.jar}" classifier="sources" />
+ </artifact:deploy>
+
+ <artifact:deploy file="${hadoop-hdfs-instrumented.jar}">
+ <remoteRepository id="apache.snapshots.https" url="${asfsnapshotrepo}"/>
+ <pom refid="hadoop.hdfs.${herriot.suffix}"/>
+ <attach file="${hadoop-hdfs-instrumented-sources.jar}" classifier="sources" />
</artifact:deploy>
</target>
-
+
<target name="set-version">
<delete file="${basedir}/ivy/hadoop-hdfs.xml"/>
<delete file="${basedir}/ivy/hadoop-hdfs-test.xml"/>
- <delete file="${hadoop-hdfs-instrumented.pom}"/>
- <delete file="${hadoop-hdfs-instrumented-test.pom}"/>
+ <delete file="${basedir}/ivy/hadoop-hdfs-${herriot.suffix}.xml"/>
+ <delete file="${basedir}/ivy/hadoop-hdfs-${herriot.suffix}-test.xml"/>
<copy file="${basedir}/ivy/hadoop-hdfs-template.xml" tofile="${basedir}/ivy/hadoop-hdfs.xml"/>
<copy file="${basedir}/ivy/hadoop-hdfs-test-template.xml" tofile="${basedir}/ivy/hadoop-hdfs-test.xml"/>
<copy file="${basedir}/ivy/hadoop-hdfs-${herriot.suffix}-template.xml"
- tofile="${hadoop-hdfs-instrumented.pom}"/>
+ tofile="${basedir}/ivy/hadoop-hdfs-${herriot.suffix}.xml"/>
<copy file="${basedir}/ivy/hadoop-hdfs-${herriot.suffix}-test-template.xml"
- tofile="${hadoop-hdfs-instrumented-test.pom}"/>
+ tofile="${basedir}/ivy/hadoop-hdfs-${herriot.suffix}-test.xml"/>
<replaceregexp byline="true">
<regexp pattern="@version"/>
<substitution expression="${version}"/>
@@ -1486,7 +1605,6 @@
</fileset>
</replaceregexp>
</target>
-
<!--
To avoid Ivy leaking things across big projects, always load Ivy in the same classloader.
Modified: hadoop/hdfs/branches/branch-0.21/ivy/libraries.properties
URL: http://svn.apache.org/viewvc/hadoop/hdfs/branches/branch-0.21/ivy/libraries.properties?rev=1033641&r1=1033640&r2=1033641&view=diff
==============================================================================
--- hadoop/hdfs/branches/branch-0.21/ivy/libraries.properties (original)
+++ hadoop/hdfs/branches/branch-0.21/ivy/libraries.properties Wed Nov 10 18:50:04 2010
@@ -63,6 +63,7 @@ servlet-api.version=2.5
slf4j-api.version=1.5.11
slf4j-log4j12.version=1.5.11
+wagon-http.version=1.0-beta-2
xmlenc.version=0.52
xerces.version=1.4.4