You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by Apache Wiki <wi...@apache.org> on 2012/01/12 16:01:44 UTC

[Hadoop Wiki] Update of "HowToContribute" by SteveLoughran

Dear Wiki user,

You have subscribed to a wiki page or wiki category on "Hadoop Wiki" for change notification.

The "HowToContribute" page has been changed by SteveLoughran:
http://wiki.apache.org/hadoop/HowToContribute?action=diff&rev1=75&rev2=76

Comment:
fix some link entries

  First of all, you need the Hadoop source code. The official location for Hadoop is the Apache SVN repository; Git is also supported, and useful if you want to make lots of local changes -and keep those changes under some form or private or public revision control.
  
  ==== SVN Access ====
- Get the source code on your local drive using [[http://hadoop.apache.org/core/version_control.html|SVN]].  Most development is done on the "trunk":
+ Get the source code on your local drive using [[http://hadoop.apache.org/core/version_control.html |SVN]].  Most development is done on the "trunk":
  
  {{{
  svn checkout http://svn.apache.org/repos/asf/hadoop/common/trunk/ hadoop-trunk
@@ -71, +71 @@

  
  === Building ProtocolBuffers (for 0.23+) ===
  
- Hadoop 0.23+ must have Google's ProtocolBuffers for compilation to work. These are native binaries which need to be downloaded, compiled and then installed locally.  See [[http://svn.apache.org/repos/asf/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/README|YARN Readme]
+ Hadoop 0.23+ must have Google's ProtocolBuffers for compilation to work. These are native binaries which need to be downloaded, compiled and then installed locally.  See [[http://svn.apache.org/repos/asf/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/README |YARN Readme]]
  
  This is a good opportunity to get the GNU C/C++ toolchain installed, which is useful for working on the native code used in the HDFS project.
  
@@ -85, +85 @@

  
   * you need a copy of GCC 4.1+ including the {{{g++}}} C++ compiler, {{{make}}} and the rest of the GNU C++ development chain.
   * Linux: you need a copy of autoconf installed, which your local package manager will do -along with automake.
-  * Download the version of protocol buffers that the YARN readme recommends from [[http://code.google.com/p/protobuf/|the protocol buffers project]].
+  * Download the version of protocol buffers that the YARN readme recommends from [[http://code.google.com/p/protobuf/ |the protocol buffers project]].
   * unzip it/untar it
   * {{{cd}}} into the directory that has been created.
   * run {{{/.configure}}}.
@@ -108, +108 @@

  protoc: error while loading shared libraries: libprotobuf.so.7: cannot open shared object file: No such file or directory
  }}}
  
- This is a [[http://code.google.com/p/protobuf/issues/detail?id=213|known issue]] for Linux, and is caused by a stale cache of libraries. Run {{{ldconfig}}} and try again.
+ This is a [[http://code.google.com/p/protobuf/issues/detail?id=213 |known issue]] for Linux, and is caused by a stale cache of libraries. Run {{{ldconfig}}} and try again.
  
  === Making Changes ===
  Before you start, send a message to the [[http://hadoop.apache.org/core/mailing_lists.html|Hadoop developer mailing list]], or file a bug report in [[Jira]].  Describe your proposed changes and check that they fit in with what others are doing and have planned for the project.  Be patient, it may take folks a while to understand your requirements.