You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-issues@hadoop.apache.org by "Aaron Kimball (JIRA)" <ji...@apache.org> on 2009/10/30 22:35:59 UTC

[jira] Updated: (HADOOP-6342) Create a script to squash a common, hdfs, and mapreduce tarball into a single hadoop tarball

     [ https://issues.apache.org/jira/browse/HADOOP-6342?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Aaron Kimball updated HADOOP-6342:
----------------------------------

    Attachment: HADOOP-6342.patch

This patch provides a Makefile that can build all three projects and merge them. It's based on a script that I have been using myself that I finally got around to cleaning up a bit.

This adds a Makefile in {{src/buildscripts/}} as well as some shell scripts that it calls out to.

Running {{make combined-tar}} will create a combined tarball. It can also execute other commands across all three projects, e.g., {{make compile-core}} and combine their results into a single build directory. See the included README and the Makefile's comments for the full story.

In order to support mapred, hdfs, and core running in the same directory, this also modifies {{bin/hadoop-config.sh}} to assume that {{$HADOOP_CORE_HOME}} can satisify {{$HADOOP_MAPRED_HOME}} and {{$HADOOP_HDFS_HOME}} in the absense of {{hdfs/}} and {{mapred/}} subdirs.

Finally, this current version contains a hack in {{combine-bindirs.sh}} to reconcile the competing versions of various libraries used by mapred, hdfs, and common. These should be reconciled via their ivy.xml files -- separate ticket? They are left in for now for demonstration that this Makefile system works.

I've tested this by running the various commands and verifying they work. I also used 'make combined-tar' to build a tarball, unzipped it elsewhere, and used it to start a pseudodistributed instance and run pi. I also ran this from my hadoop-common/build/combined-hadoop directory created with {{make combined-binary}}.

If you commit this or something like it, you'll need to {{svn add src/buildscripts/}}

> Create a script to squash a common, hdfs, and mapreduce tarball into a single hadoop tarball
> --------------------------------------------------------------------------------------------
>
>                 Key: HADOOP-6342
>                 URL: https://issues.apache.org/jira/browse/HADOOP-6342
>             Project: Hadoop Common
>          Issue Type: New Feature
>          Components: build
>            Reporter: Owen O'Malley
>            Assignee: Aaron Kimball
>            Priority: Minor
>         Attachments: HADOOP-6342.patch
>
>
> It would be convenient for the transition if we had a script to take a set of common, hdfs, and mapreduce tarballs and merge them into a single tarball. This is intended just to help users who don't want to transition to split projects for deployment immediately.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.