You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by mi...@apache.org on 2014/09/11 01:07:37 UTC

git commit: HBASE-11932 Docbook html-single build improvements

Repository: hbase
Updated Branches:
  refs/heads/master eec15bd17 -> 281648724


HBASE-11932 Docbook html-single build improvements


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/28164872
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/28164872
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/28164872

Branch: refs/heads/master
Commit: 281648724758f571300e0bc167f16b2bcc023fb3
Parents: eec15bd
Author: Misty Stanley-Jones <ms...@cloudera.com>
Authored: Thu Sep 11 09:06:17 2014 +1000
Committer: Misty Stanley-Jones <ms...@cloudera.com>
Committed: Thu Sep 11 09:06:17 2014 +1000

----------------------------------------------------------------------
 pom.xml                                   |   1 +
 src/main/docbkx/appendix_hfile_format.xml |  28 +++---
 src/main/docbkx/book.xml                  | 113 ++++++++++---------------
 src/main/docbkx/community.xml             |   5 +-
 src/main/docbkx/configuration.xml         |  60 +++++--------
 src/main/docbkx/cp.xml                    |  16 ++--
 src/main/docbkx/ops_mgt.xml               |  28 +++---
 src/main/docbkx/performance.xml           |  63 ++++++--------
 src/main/docbkx/preface.xml               |   2 +-
 src/main/docbkx/rpc.xml                   |   3 +-
 src/main/docbkx/security.xml              |   8 +-
 src/main/docbkx/troubleshooting.xml       |  23 +++--
 src/main/docbkx/upgrading.xml             |   8 +-
 src/main/docbkx/zookeeper.xml             |   4 +-
 14 files changed, 151 insertions(+), 211 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/hbase/blob/28164872/pom.xml
----------------------------------------------------------------------
diff --git a/pom.xml b/pom.xml
index 91fd6fa..3fbd4af 100644
--- a/pom.xml
+++ b/pom.xml
@@ -807,6 +807,7 @@
             <phase>pre-site</phase>
             <configuration>
               <targetDirectory>${basedir}/target/docbkx/</targetDirectory>
+              <includes>book.xml</includes>
               <preProcess>
                 <copy todir="target/docbkx/images">
                   <fileset dir="src/main/site/resources/images/"/>

http://git-wip-us.apache.org/repos/asf/hbase/blob/28164872/src/main/docbkx/appendix_hfile_format.xml
----------------------------------------------------------------------
diff --git a/src/main/docbkx/appendix_hfile_format.xml b/src/main/docbkx/appendix_hfile_format.xml
index aac58d3..ee43031 100644
--- a/src/main/docbkx/appendix_hfile_format.xml
+++ b/src/main/docbkx/appendix_hfile_format.xml
@@ -31,18 +31,22 @@
     <para>As we will be discussing changes to the HFile format, it is useful to give a short overview of the original (HFile version 1) format.</para>
     <section xml:id="hfilev1.overview">
       <title>Overview of Version 1</title>
-      <para>An HFile in version 1 format is structured as follows:
-           <inlinemediaobject>
-               <imageobject>
-                   <imagedata align="center" valign="middle" fileref="hfile.png" />
-               </imageobject>
-               <textobject>
-                 <phrase>HFile Version 1</phrase>
-               </textobject>
-               
-           </inlinemediaobject>
-           <footnote><para>Image courtesy of Lars George, <link xlink:href="http://www.larsgeorge.com/2009/10/hbase-architecture-101-storage.html">hbase-architecture-101-storage.html</link>.</para></footnote>
-       </para>
+      <para>An HFile in version 1 format is structured as follows:</para>
+      <figure>
+         <title>HFile V1 Format</title>
+         <mediaobject>
+            <imageobject>
+               <imagedata align="center" valign="middle" fileref="hfile.png"/>
+            </imageobject>
+            <textobject>
+               <phrase>HFile Version 1</phrase>
+            </textobject>
+            <caption><para>Image courtesy of Lars George, <link
+                     xlink:href="http://www.larsgeorge.com/2009/10/hbase-architecture-101-storage.html"
+                     >hbase-architecture-101-storage.html</link>.</para></caption>
+         </mediaobject>
+      </figure>
+
     </section>
        <section><title> Block index format in version 1 </title>
    <para>The block index in version 1 is very straightforward. For each entry, it contains: </para>

http://git-wip-us.apache.org/repos/asf/hbase/blob/28164872/src/main/docbkx/book.xml
----------------------------------------------------------------------
diff --git a/src/main/docbkx/book.xml b/src/main/docbkx/book.xml
index 683d8ef..00c5041 100644
--- a/src/main/docbkx/book.xml
+++ b/src/main/docbkx/book.xml
@@ -639,24 +639,19 @@ try {
         store file, the most recent values are found first.</para>
 
       <para>There is a lot of confusion over the semantics of <literal>cell</literal> versions, in
-        HBase. In particular, a couple questions that often come up are:</para>
+        HBase. In particular:</para>
       <itemizedlist>
         <listitem>
-          <para>If multiple writes to a cell have the same version, are all versions maintained or
-            just the last?<footnote>
-              <para>Currently, only the last written is fetchable.</para>
-            </footnote></para>
+          <para>If multiple writes to a cell have the same version, only the last written is
+            fetchable.</para>
         </listitem>
 
         <listitem>
-          <para>Is it OK to write cells in a non-increasing version order?<footnote>
-              <para>Yes</para>
-            </footnote></para>
+          <para>It is OK to write cells in a non-increasing version order.</para>
         </listitem>
       </itemizedlist>
 
-      <para>Below we describe how the version dimension in HBase currently works<footnote>
-          <para>See <link
+      <para>Below we describe how the version dimension in HBase currently works. See <link
               xlink:href="https://issues.apache.org/jira/browse/HBASE-2406">HBASE-2406</link> for
             discussion of HBase versions. <link
               xlink:href="http://outerthought.org/blog/417-ot.html">Bending time in HBase</link>
@@ -665,7 +660,6 @@ try {
               <emphasis>Overwriting values at existing timestamps</emphasis> mentioned in the
             article no longer holds in HBase. This section is basically a synopsis of this article
             by Bruno Dumon.</para>
-        </footnote>.</para>
 
       <section
         xml:id="versions.ops">
@@ -783,11 +777,10 @@ htable.put(put);
           xml:id="version.delete">
           <title>Delete</title>
 
-          <para>There are three different types of internal delete markers <footnote>
-              <para>See Lars Hofhansl's blog for discussion of his attempt adding another, <link
-                  xlink:href="http://hadoop-hbase.blogspot.com/2012/01/scanning-in-hbase.html">Scanning
-                  in HBase: Prefix Delete Marker</link></para>
-            </footnote>: </para>
+          <para>There are three different types of internal delete markers. See Lars Hofhansl's blog
+            for discussion of his attempt adding another, <link
+              xlink:href="http://hadoop-hbase.blogspot.com/2012/01/scanning-in-hbase.html">Scanning
+              in HBase: Prefix Delete Marker</link>. </para>
           <itemizedlist>
             <listitem>
               <para>Delete: for a specific version of a column.</para>
@@ -808,11 +801,10 @@ htable.put(put);
             modifies data in place, so for example a delete will not immediately delete (or mark as
             deleted) the entries in the storage file that correspond to the delete condition.
             Rather, a so-called <emphasis>tombstone</emphasis> is written, which will mask the
-            deleted values<footnote>
-              <para>When HBase does a major compaction, the tombstones are processed to actually
-                remove the dead values, together with the tombstones themselves.</para>
-            </footnote>. If the version you specified when deleting a row is larger than the version
-            of any value in the row, then you can consider the complete row to be deleted.</para>
+            deleted values. When HBase does a major compaction, the tombstones are processed to
+            actually remove the dead values, together with the tombstones themselves. If the version
+            you specified when deleting a row is larger than the version of any value in the row,
+            then you can consider the complete row to be deleted.</para>
           <para>For an informative discussion on how deletes and versioning interact, see the thread <link
               xlink:href="http://comments.gmane.org/gmane.comp.java.hadoop.hbase.user/28421">Put w/
               timestamp -> Deleteall -> Put w/ timestamp fails</link> up on the user mailing
@@ -846,10 +838,8 @@ htable.put(put);
           <title>Deletes mask Puts</title>
 
           <para>Deletes mask puts, even puts that happened after the delete
-          was entered<footnote>
-              <para><link
-              xlink:href="https://issues.apache.org/jira/browse/HBASE-2256">HBASE-2256</link></para>
-            </footnote>. Remember that a delete writes a tombstone, which only
+          was entered. See <link xlink:href="https://issues.apache.org/jira/browse/HBASE-2256"
+              >HBASE-2256</link>. Remember that a delete writes a tombstone, which only
           disappears after then next major compaction has run. Suppose you do
           a delete of everything &lt;= T. After this you do a new put with a
           timestamp &lt;= T. This put, even if it happened after the delete,
@@ -868,14 +858,12 @@ htable.put(put);
           <title>Major compactions change query results</title>
           
           <para><quote>...create three cell versions at t1, t2 and t3, with a maximum-versions
-            setting of 2. So when getting all versions, only the values at t2 and t3 will be
-            returned. But if you delete the version at t2 or t3, the one at t1 will appear again.
-            Obviously, once a major compaction has run, such behavior will not be the case anymore...<footnote>
-              <para>See <emphasis>Garbage Collection</emphasis> in <link
-                xlink:href="http://outerthought.org/blog/417-ot.html">Bending time in
-                HBase</link>
-              </para>
-            </footnote></quote></para>
+              setting of 2. So when getting all versions, only the values at t2 and t3 will be
+              returned. But if you delete the version at t2 or t3, the one at t1 will appear again.
+              Obviously, once a major compaction has run, such behavior will not be the case
+              anymore...</quote> (See <emphasis>Garbage Collection</emphasis> in <link
+              xlink:href="http://outerthought.org/blog/417-ot.html">Bending time in
+            HBase</link>.)</para>
         </section>
       </section>
     </section>
@@ -2020,14 +2008,13 @@ rs.close();
 	</section>  <!--  client.filter -->
 
     <section xml:id="master"><title>Master</title>
-       <para><code>HMaster</code> is the implementation of the Master Server.  The Master server
-       is responsible for monitoring all RegionServer instances in the cluster, and is
-       the interface for all metadata changes.  In a distributed cluster, the Master typically runs on the <xref linkend="arch.hdfs.nn" /><footnote>
-            <para>J Mohamed Zahoor goes into some more detail on the Master Architecture in this blog posting, <link
-            xlink:href="http://blog.zahoor.in/2012/08/hbase-hmaster-architecture/">HBase HMaster Architecture
-            </link>.</para>
-          </footnote>
-       </para>
+      <para><code>HMaster</code> is the implementation of the Master Server. The Master server is
+        responsible for monitoring all RegionServer instances in the cluster, and is the interface
+        for all metadata changes. In a distributed cluster, the Master typically runs on the <xref
+          linkend="arch.hdfs.nn"/>. J Mohamed Zahoor goes into some more detail on the Master
+        Architecture in this blog posting, <link
+          xlink:href="http://blog.zahoor.in/2012/08/hbase-hmaster-architecture/">HBase HMaster
+          Architecture </link>.</para>
        <section xml:id="master.startup"><title>Startup Behavior</title>
          <para>If run in a multi-Master environment, all Masters compete to run the cluster.  If the active
          Master loses its lease in ZooKeeper (or the Master shuts down), then then the remaining Masters jostle to
@@ -2469,17 +2456,16 @@ rs.close();
                         physical RAM, and is likely to be less than the total available RAM due to other
                         memory requirements and system constraints.
                 </para>
-                <para>You can see how much memory -- onheap and offheap/direct -- a RegionServer is configured to use
-                    and how much it is using at any one time by looking at the
-                    <emphasis>Server Metrics: Memory</emphasis> tab in the UI.
-                    It can also be gotten via JMX.  In particular the direct
-                    memory currently used by the server can be found on the
-                    <varname>java.nio.type=BufferPool,name=direct</varname>
-                    bean.
-                    <footnote><para>Terracotta has a <link xlink:href="http://terracotta.org/documentation/4.0/bigmemorygo/configuration/storage-options">good write up</link> on using offheap memory in java.
-                        It is for their product BigMemory but alot of the issues noted apply
-                        in general to any attempt at going offheap. Check it out.</para></footnote>
-                </para>
+              <para>You can see how much memory -- onheap and offheap/direct -- a RegionServer is
+                configured to use and how much it is using at any one time by looking at the
+                  <emphasis>Server Metrics: Memory</emphasis> tab in the UI. It can also be gotten
+                via JMX. In particular the direct memory currently used by the server can be found
+                on the <varname>java.nio.type=BufferPool,name=direct</varname> bean. Terracotta has
+                a <link
+                  xlink:href="http://terracotta.org/documentation/4.0/bigmemorygo/configuration/storage-options"
+                  >good write up</link> on using offheap memory in java. It is for their product
+                BigMemory but alot of the issues noted apply in general to any attempt at going
+                offheap. Check it out.</para>
             </note>
               <note xml:id="hbase.bucketcache.percentage.in.combinedcache"><title>hbase.bucketcache.percentage.in.combinedcache</title>
                   <para>This is a pre-HBase 1.0 configuration removed because it
@@ -2613,12 +2599,10 @@ rs.close();
             </itemizedlist>
             <para>If the <varname>hbase.hlog.split.skip.errors</varname> optionset to
                 <literal>false</literal>, the default, the exception will be propagated and the
-              split will be logged as failed.<footnote>
-                <para>See <link
+              split will be logged as failed. See <link
                     xlink:href="https://issues.apache.org/jira/browse/HBASE-2958">HBASE-2958 When
                     hbase.hlog.split.skip.errors is set to false, we fail the split but thats
                     it</link>. We need to do more than just fail split if this flag is set.</para>
-              </footnote></para>
             
             <section>
               <title>How EOFExceptions are treated when splitting a crashed RegionServers'
@@ -2628,11 +2612,9 @@ rs.close();
                   <varname>hbase.hlog.split.skip.errors</varname> is set to
                 <literal>false</literal>. An EOFException while reading the last log in the set of
                 files to split is likely, because the RegionServer is likely to be in the process of
-                writing a record at the time of a crash. <footnote>
-                  <para>For background, see <link
+                writing a record at the time of a crash. For background, see <link
                       xlink:href="https://issues.apache.org/jira/browse/HBASE-2643">HBASE-2643
                       Figure how to deal with eof splitting logs</link></para>
-                </footnote></para>
             </section>
           </section>
           
@@ -3042,9 +3024,9 @@ ctime = Sat Jun 23 11:13:40 PDT 2012
              </listitem>
              <listitem><para>Third replica is written on the same rack as the second, but on a different node chosen randomly</para>
              </listitem>
-             <listitem><para>Subsequent replicas are written on random nodes on the cluster
-<footnote><para>See <emphasis>Replica Placement: The First Baby Steps</emphasis> on this page: <link xlink:href="http://hadoop.apache.org/docs/stable/hadoop-project-dist/hadoop-hdfs/HdfsDesign.html">HDFS Architecture</link></para></footnote>
-             </para></listitem>
+             <listitem><para>Subsequent replicas are written on random nodes on the cluster. See <emphasis>Replica Placement: The First Baby Steps</emphasis> on this page: <link
+                xlink:href="http://hadoop.apache.org/docs/stable/hadoop-project-dist/hadoop-hdfs/HdfsDesign.html"
+                >HDFS Architecture</link></para></listitem>
            </orderedlist>
         <para>
           Thus, HBase eventually achieves locality for a region after a flush or a compaction.
@@ -5166,8 +5148,7 @@ This option should not normally be used, and it is not in <code>-fixAll</code>.
             <imageobject>
               <imagedata fileref="data_block_no_encoding.png" width="800"/>
             </imageobject>
-            <textobject><para></para>
-            </textobject>
+            <caption><para>A ColumnFamily with no encoding></para></caption>
           </mediaobject>
         </figure>
         <para>Here is the same data with prefix data encoding.</para>
@@ -5177,8 +5158,7 @@ This option should not normally be used, and it is not in <code>-fixAll</code>.
             <imageobject>
               <imagedata fileref="data_block_prefix_encoding.png" width="800"/>
             </imageobject>
-            <textobject><para></para>
-            </textobject>
+            <caption><para>A ColumnFamily with prefix encoding</para></caption>
           </mediaobject>
         </figure>
       </listitem>
@@ -5202,8 +5182,7 @@ This option should not normally be used, and it is not in <code>-fixAll</code>.
             <imageobject>
               <imagedata fileref="data_block_diff_encoding.png" width="800"/>
             </imageobject>
-            <textobject><para></para>
-            </textobject>
+            <caption><para>A ColumnFamily with diff encoding</para></caption>
           </mediaobject>
         </figure>
       </listitem>

http://git-wip-us.apache.org/repos/asf/hbase/blob/28164872/src/main/docbkx/community.xml
----------------------------------------------------------------------
diff --git a/src/main/docbkx/community.xml b/src/main/docbkx/community.xml
index 9bdaa39..813f356 100644
--- a/src/main/docbkx/community.xml
+++ b/src/main/docbkx/community.xml
@@ -40,12 +40,9 @@
         committer will add it for you. Thereafter you can file issues against your feature branch in
         Apache HBase JIRA. Your code you keep elsewhere -- it should be public so it can be observed
         -- and you can update dev mailing list on progress. When the feature is ready for commit, 3
-        +1s from committers will get your feature merged<footnote>
-          <para>See <link
+        +1s from committers will get your feature merged. See <link
               xlink:href="http://search-hadoop.com/m/asM982C5FkS1">HBase, mail # dev - Thoughts
               about large feature dev branches</link></para>
-        </footnote>
-      </para>
     </section>
     <section
       xml:id="patchplusonepolicy">

http://git-wip-us.apache.org/repos/asf/hbase/blob/28164872/src/main/docbkx/configuration.xml
----------------------------------------------------------------------
diff --git a/src/main/docbkx/configuration.xml b/src/main/docbkx/configuration.xml
index 898aa85..0af2b3c 100644
--- a/src/main/docbkx/configuration.xml
+++ b/src/main/docbkx/configuration.xml
@@ -212,12 +212,10 @@
         <term>DNS</term>
         <listitem>
           <para>HBase uses the local hostname to self-report its IP address. Both forward and
-            reverse DNS resolving must work in versions of HBase previous to 0.92.0.<footnote>
-              <para>The <link
+            reverse DNS resolving must work in versions of HBase previous to 0.92.0. The <link
                   xlink:href="https://github.com/sujee/hadoop-dns-checker">hadoop-dns-checker</link>
                 tool can be used to verify DNS is working correctly on the cluster. The project
                 README file provides detailed instructions on usage. </para>
-            </footnote></para>
 
           <para>If your server has multiple network interfaces, HBase defaults to using the
             interface that the primary hostname resolves to. To override this behavior, set the
@@ -306,11 +304,10 @@
             running the HBase process is an operating system configuration, rather than an HBase
             configuration. It is also important to be sure that the settings are changed for the
             user that actually runs HBase. To see which user started HBase, and that user's ulimit
-            configuration, look at the first line of the HBase log for that instance.<footnote>
-              <para>A useful read setting config on you hadoop cluster is Aaron Kimballs' <link
-                  xlink:href="http://www.cloudera.com/blog/2009/03/configuration-parameters-what-can-you-just-ignore/">Configuration
-                  Parameters: What can you just ignore?</link></para>
-            </footnote></para>
+            configuration, look at the first line of the HBase log for that instance. A useful read
+            setting config on you hadoop cluster is Aaron Kimballs' <link
+              xlink:href="http://www.cloudera.com/blog/2009/03/configuration-parameters-what-can-you-just-ignore/"
+              >Configuration Parameters: What can you just ignore?</link></para>
           <formalpara xml:id="ulimit_ubuntu">
             <title><command>ulimit</command> Settings on Ubuntu</title>
             <para>To configure <command>ulimit</command> settings on Ubuntu, edit
@@ -414,12 +411,8 @@ hadoop  -       nproc   32000
               <entry>HBase-0.92.x</entry>
               <entry>HBase-0.94.x</entry>
               <entry>HBase-0.96.x</entry>
-              <entry>HBase-0.98.x<footnote>
-                  <para>Support for Hadoop 1.x is deprecated.</para>
-                </footnote></entry>
-              <entry>HBase-1.0.x<footnote>
-                  <para>Hadoop 1.x is NOT supported</para>
-                </footnote></entry>
+              <entry><para>HBase-0.98.x (Support for Hadoop 1.x is deprecated.)</para></entry>
+              <entry><para>HBase-1.0.x (Hadoop 1.x is NOT supported)</para></entry>
             </row>
           </thead>
           <tbody>
@@ -440,11 +433,9 @@ hadoop  -       nproc   32000
               <entry>X</entry>
             </row>
             <row>
-              <entry>Hadoop-1.0.0-1.0.2<footnote>
-                  <para>HBase requires hadoop 1.0.3 at a minimum; there is an issue where we cannot
-                    find KerberosUtil compiling against earlier versions of Hadoop.</para>
-                </footnote>
-              </entry>
+              <entry><para>Hadoop-1.0.0-1.0.2 (HBase requires hadoop 1.0.3 at a minimum; there is an
+                  issue where we cannot find KerberosUtil compiling against earlier versions of
+                  Hadoop.)</para></entry>
               <entry>X</entry>
               <entry>X</entry>
               <entry>X</entry>
@@ -494,10 +485,9 @@ hadoop  -       nproc   32000
             <row>
               <entry>Hadoop-2.2.0 </entry>
               <entry>X</entry>
-              <entry>NT <footnote>
-                  <para>To get 0.94.x to run on hadoop 2.2.0, you need to change the hadoop 2 and
-                    protobuf versions in the <filename>pom.xml</filename>: Here is a diff with
-                    pom.xml changes: </para>
+              <entry><para>NT - To get 0.94.x to run on hadoop 2.2.0, you need to change the hadoop
+                  2 and protobuf versions in the <filename>pom.xml</filename>: Here is a diff with
+                  pom.xml changes: </para>
                   <programlisting><![CDATA[$ svn diff pom.xml
 Index: pom.xml
 ===================================================================
@@ -540,8 +530,7 @@ Index: pom.xml
                   </itemizedlist>
                   <para> Building against the hadoop 2 profile by running something like the
                     following command: </para>
-                  <screen language="bourne">$  mvn clean install assembly:single -Dhadoop.profile=2.0 -DskipTests</screen>
-                </footnote></entry>
+                  <screen language="bourne">$  mvn clean install assembly:single -Dhadoop.profile=2.0 -DskipTests</screen></entry>
               <entry>S</entry>
               <entry>S</entry>
               <entry>NT</entry>
@@ -601,11 +590,9 @@ Index: pom.xml
         <para> As of Apache HBase 0.96.x, Apache Hadoop 1.0.x at least is required. Hadoop 2 is
           strongly encouraged (faster but also has fixes that help MTTR). We will no longer run
           properly on older Hadoops such as 0.20.205 or branch-0.20-append. Do not move to Apache
-          HBase 0.96.x if you cannot upgrade your Hadoop.<footnote>
-            <para>See <link
+          HBase 0.96.x if you cannot upgrade your Hadoop.. See <link
                 xlink:href="http://search-hadoop.com/m/7vFVx4EsUb2">HBase, mail # dev - DISCUSS:
                 Have hbase require at least hadoop 1.0.0 in hbase 0.96.0?</link></para>
-          </footnote></para>
       </section>
 
       <section
@@ -615,13 +602,12 @@ Index: pom.xml
             <code>sync</code> implementation. DO NOT use Hadoop 0.20.2, Hadoop 0.20.203.0, and
           Hadoop 0.20.204.0 which DO NOT have this attribute. Currently only Hadoop versions
           0.20.205.x or any release in excess of this version -- this includes hadoop-1.0.0 -- have
-          a working, durable sync <footnote>
-            <para>The Cloudera blog post <link
-                xlink:href="http://www.cloudera.com/blog/2012/01/an-update-on-apache-hadoop-1-0/">An
-                update on Apache Hadoop 1.0</link> by Charles Zedlweski has a nice exposition on how
-              all the Hadoop versions relate. Its worth checking out if you are having trouble
-              making sense of the Hadoop version morass. </para>
-          </footnote>. Sync has to be explicitly enabled by setting
+          a working, durable sync. The Cloudera blog post <link
+            xlink:href="http://www.cloudera.com/blog/2012/01/an-update-on-apache-hadoop-1-0/">An
+            update on Apache Hadoop 1.0</link> by Charles Zedlweski has a nice exposition on how all
+          the Hadoop versions relate. Its worth checking out if you are having trouble making sense
+          of the Hadoop version morass. </para>
+        <para>Sync has to be explicitly enabled by setting
             <varname>dfs.support.append</varname> equal to true on both the client side -- in
             <filename>hbase-site.xml</filename> -- and on the serverside in
             <filename>hdfs-site.xml</filename> (The sync facility HBase needs is a subset of the
@@ -713,9 +699,7 @@ Index: pom.xml
       <para>Distributed mode can be subdivided into distributed but all daemons run on a single node
         -- a.k.a <emphasis>pseudo-distributed</emphasis>-- and
           <emphasis>fully-distributed</emphasis> where the daemons are spread across all nodes in
-        the cluster <footnote>
-          <para>The pseudo-distributed vs fully-distributed nomenclature comes from Hadoop.</para>
-        </footnote>.</para>
+        the cluster. The pseudo-distributed vs fully-distributed nomenclature comes from Hadoop.</para>
 
       <para>Pseudo-distributed mode can run against the local filesystem or it can run against an
         instance of the <emphasis>Hadoop Distributed File System</emphasis> (HDFS).

http://git-wip-us.apache.org/repos/asf/hbase/blob/28164872/src/main/docbkx/cp.xml
----------------------------------------------------------------------
diff --git a/src/main/docbkx/cp.xml b/src/main/docbkx/cp.xml
index 4ddc8c1..8624309 100644
--- a/src/main/docbkx/cp.xml
+++ b/src/main/docbkx/cp.xml
@@ -29,16 +29,14 @@
  */
 -->
   <title>Apache HBase Coprocessors</title>
-  <para> HBase coprocessors are modeled after the coprocessors which are part of Google's BigTable<footnote>
-      <para><link
-          xlink:href="http://www.scribd.com/doc/21631448/Dean-Keynote-Ladis2009" />, pages
-        66-67.</para>
-    </footnote>. Coprocessors function in a similar way to Linux kernel modules. They provide a way
-    to run server-level code against locally-stored data. The functionality they provide is very
+  <para> HBase coprocessors are modeled after the coprocessors which are part of Google's BigTable
+      (<link xlink:href="http://www.scribd.com/doc/21631448/Dean-Keynote-Ladis2009"/>, pages
+    66-67.). Coprocessors function in a similar way to Linux kernel modules. They provide a way to
+    run server-level code against locally-stored data. The functionality they provide is very
     powerful, but also carries great risk and can have adverse effects on the system, at the level
     of the operating system. The information in this chapter is primarily sourced and heavily reused
     from Mingjie Lai's blog post at <link
-      xlink:href="https://blogs.apache.org/hbase/entry/coprocessor_introduction" />. </para>
+      xlink:href="https://blogs.apache.org/hbase/entry/coprocessor_introduction"/>. </para>
 
   <para> Coprocessors are not designed to be used by end users of HBase, but by HBase developers who
     need to add specialized functionality to HBase. One example of the use of coprocessors is
@@ -418,10 +416,10 @@ coprocessors=[AggregateImplementation]
         <imageobject>
           <imagedata fileref="coprocessor_stats.png" width="100%"/>
         </imageobject>
-        <textobject>
+        <caption>
           <para>The Coprocessor Metrics UI shows statistics about time spent executing a given
             coprocessor, including min, max, average, and 90th, 95th, and 99th percentile.</para>
-        </textobject>
+        </caption>
       </mediaobject>
     </figure>
   </section>

http://git-wip-us.apache.org/repos/asf/hbase/blob/28164872/src/main/docbkx/ops_mgt.xml
----------------------------------------------------------------------
diff --git a/src/main/docbkx/ops_mgt.xml b/src/main/docbkx/ops_mgt.xml
index 7328f67..21045bd 100644
--- a/src/main/docbkx/ops_mgt.xml
+++ b/src/main/docbkx/ops_mgt.xml
@@ -757,11 +757,9 @@ false
         <para>Without this facility, decommissioning mulitple nodes may be non-optimal because
           regions that are being drained from one region server may be moved to other regionservers
           that are also draining. Marking RegionServers to be in the draining state prevents this
-          from happening<footnote>
-            <para>See this <link
-                xlink:href="http://inchoate-clatter.blogspot.com/2012/03/hbase-ops-automation.html">blog
-                post</link> for more details.</para>
-          </footnote>. </para>
+          from happening. See this <link
+            xlink:href="http://inchoate-clatter.blogspot.com/2012/03/hbase-ops-automation.html">blog
+            post</link> for more details.</para>
       </section>
 
       <section
@@ -1206,9 +1204,9 @@ $ for i in `cat conf/regionservers|sort`; do ./bin/graceful_stop.sh --restart --
           <imageobject>
             <imagedata fileref="bc_basic.png" width="100%"/>
           </imageobject>
-          <textobject>
+          <caption>
             <para>Shows the cache implementation</para>
-          </textobject>
+          </caption>
         </mediaobject>
       </figure>
       <figure>
@@ -1217,9 +1215,9 @@ $ for i in `cat conf/regionservers|sort`; do ./bin/graceful_stop.sh --restart --
           <imageobject>
             <imagedata fileref="bc_config.png" width="100%"/>
           </imageobject>
-          <textobject>
+          <caption>
             <para>Shows all cache configuration options.</para>
-          </textobject>
+          </caption>
         </mediaobject>
       </figure>
       <figure>
@@ -1228,9 +1226,9 @@ $ for i in `cat conf/regionservers|sort`; do ./bin/graceful_stop.sh --restart --
           <imageobject>
             <imagedata fileref="bc_stats.png" width="100%"/>
           </imageobject>
-          <textobject>
+          <caption>
             <para>Shows statistics about the performance of the cache.</para>
-          </textobject>
+          </caption>
         </mediaobject>
       </figure>
       <figure>
@@ -1242,9 +1240,9 @@ $ for i in `cat conf/regionservers|sort`; do ./bin/graceful_stop.sh --restart --
           <imageobject>
             <imagedata fileref="bc_l2_buckets.png" width="100%"/>
           </imageobject>
-          <textobject>
+          <caption>
             <para>Shows information about the L1 and L2 caches.</para>
-          </textobject>
+          </caption>
         </mediaobject>
       </figure>
       <para>This is not an exhaustive list of all the screens and reports available. Have a look in
@@ -1305,10 +1303,10 @@ $ for i in `cat conf/regionservers|sort`; do ./bin/graceful_stop.sh --restart --
         <imageobject>
           <imagedata fileref="replication_overview.png" />
         </imageobject>
-        <textobject>
+        <caption>
           <para>Illustration of the replication architecture in HBase, as described in the prior
             text.</para>
-        </textobject>
+        </caption>
       </mediaobject>
     </figure>
     

http://git-wip-us.apache.org/repos/asf/hbase/blob/28164872/src/main/docbkx/performance.xml
----------------------------------------------------------------------
diff --git a/src/main/docbkx/performance.xml b/src/main/docbkx/performance.xml
index 1940663..e7c0fc7 100644
--- a/src/main/docbkx/performance.xml
+++ b/src/main/docbkx/performance.xml
@@ -146,8 +146,7 @@
         xml:id="gcpause">
         <title>Long GC pauses</title>
 
-        <para
-          xml:id="mslab">In his presentation, <link
+        <para xml:id="mslab">In his presentation, <link
             xlink:href="http://www.slideshare.net/cloudera/hbase-hug-presentation">Avoiding Full GCs
             with MemStore-Local Allocation Buffers</link>, Todd Lipcon describes two cases of
           stop-the-world garbage collections common in HBase, especially during loading; CMS failure
@@ -158,16 +157,16 @@
           Todd added an experimental facility, <indexterm><primary>MSLAB</primary></indexterm>, that
           must be explicitly enabled in Apache HBase 0.90.x (Its defaulted to be on in Apache 0.92.x
           HBase). See <code>hbase.hregion.memstore.mslab.enabled</code> to true in your
-            <classname>Configuration</classname>. See the cited slides for background and detail<footnote>
-            <para>The latest jvms do better regards fragmentation so make sure you are running a
-              recent release. Read down in the message, <link
-                xlink:href="http://osdir.com/ml/hotspot-gc-use/2011-11/msg00002.html">Identifying
-                concurrent mode failures caused by fragmentation</link>.</para>
-          </footnote>. Be aware that when enabled, each MemStore instance will occupy at least an
-          MSLAB instance of memory. If you have thousands of regions or lots of regions each with
-          many column families, this allocation of MSLAB may be responsible for a good portion of
-          your heap allocation and in an extreme case cause you to OOME. Disable MSLAB in this case,
-          or lower the amount of memory it uses or float less regions per server. </para>
+            <classname>Configuration</classname>. See the cited slides for background and detail.
+          The latest jvms do better regards fragmentation so make sure you are running a recent
+          release. Read down in the message, <link
+            xlink:href="http://osdir.com/ml/hotspot-gc-use/2011-11/msg00002.html">Identifying
+            concurrent mode failures caused by fragmentation</link>. Be aware that when enabled,
+          each MemStore instance will occupy at least an MSLAB instance of memory. If you have
+          thousands of regions or lots of regions each with many column families, this allocation of
+          MSLAB may be responsible for a good portion of your heap allocation and in an extreme case
+          cause you to OOME. Disable MSLAB in this case, or lower the amount of memory it uses or
+          float less regions per server. </para>
         <para>If you have a write-heavy workload, check out <link
             xlink:href="https://issues.apache.org/jira/browse/HBASE-8163">HBASE-8163
             MemStoreChunkPool: An improvement for JAVA GC when using MSLAB</link>. It describes
@@ -916,24 +915,20 @@ htable.close();
         latencies.</para>
       <para><link
           xlink:href="http://en.wikipedia.org/wiki/Bloom_filter">Bloom filters</link> were developed
-        over in <link
-          xlink:href="https://issues.apache.org/jira/browse/HBASE-1200">HBase-1200 Add
-          bloomfilters</link>.<footnote>
-          <para>For description of the development process -- why static blooms rather than dynamic
+        over in <link xlink:href="https://issues.apache.org/jira/browse/HBASE-1200">HBase-1200 Add
+          bloomfilters</link>. For description of the development process -- why static blooms rather than dynamic
             -- and for an overview of the unique properties that pertain to blooms in HBase, as well
             as possible future directions, see the <emphasis>Development Process</emphasis> section
             of the document <link
               xlink:href="https://issues.apache.org/jira/secure/attachment/12444007/Bloom_Filters_in_HBase.pdf">BloomFilters
               in HBase</link> attached to <link
-              xlink:href="https://issues.apache.org/jira/browse/HBASE-1200">HBase-1200</link>.</para>
-        </footnote><footnote>
-          <para>The bloom filters described here are actually version two of blooms in HBase. In
+              xlink:href="https://issues.apache.org/jira/browse/HBASE-1200">HBase-1200</link>.  The bloom filters described here are actually version two of blooms in HBase. In
             versions up to 0.19.x, HBase had a dynamic bloom option based on work done by the <link
               xlink:href="http://www.one-lab.org">European Commission One-Lab Project 034819</link>.
             The core of the HBase bloom work was later pulled up into Hadoop to implement
             org.apache.hadoop.io.BloomMapFile. Version 1 of HBase blooms never worked that well.
             Version 2 is a rewrite from scratch though again it starts with the one-lab work.</para>
-        </footnote></para>
+        
       <para>See also <xref
           linkend="schema.bloom" />. </para>
 
@@ -1047,11 +1042,9 @@ htable.close();
         possible for the DFSClient to take a "short circuit" and read directly from the disk instead
         of going through the DataNode when the data is local. What this means for HBase is that the
         RegionServers can read directly off their machine's disks instead of having to open a socket
-        to talk to the DataNode, the former being generally much faster<footnote>
-          <para>See JD's <link
+        to talk to the DataNode, the former being generally much faster. See JD's <link
               xlink:href="http://files.meetup.com/1350427/hug_ebay_jdcryans.pdf">Performance
-              Talk</link></para>
-        </footnote>. Also see <link
+              Talk</link>. Also see <link
           xlink:href="http://search-hadoop.com/m/zV6dKrLCVh1">HBase, mail # dev - read short
           circuit</link> thread for more discussion around short circuit reads. </para>
       <para>To enable "short circuit" reads, it will depend on your version of Hadoop. The original
@@ -1081,19 +1074,17 @@ htable.close();
   </description>
 </property>]]></programlisting>
       <para>Be careful about permissions for the directory that hosts the shared domain socket;
-        dfsclient will complain if open to other than the hbase user. <footnote>
-          <para>If you are running on an old Hadoop, one that is without <link
-              xlink:href="https://issues.apache.org/jira/browse/HDFS-347">HDFS-347</link> but that
-            has <link
-              xlink:href="https://issues.apache.org/jira/browse/HDFS-2246">HDFS-2246</link>, you
-            must set two configurations. First, the hdfs-site.xml needs to be amended. Set the
-            property <varname>dfs.block.local-path-access.user</varname> to be the
-              <emphasis>only</emphasis> user that can use the shortcut. This has to be the user that
-            started HBase. Then in hbase-site.xml, set
-              <varname>dfs.client.read.shortcircuit</varname> to be <varname>true</varname>
-          </para>
-        </footnote>
+        dfsclient will complain if open to other than the hbase user. </para>
+      <para>If you are running on an old Hadoop, one that is without <link
+          xlink:href="https://issues.apache.org/jira/browse/HDFS-347">HDFS-347</link> but that has
+          <link xlink:href="https://issues.apache.org/jira/browse/HDFS-2246">HDFS-2246</link>, you
+        must set two configurations. First, the hdfs-site.xml needs to be amended. Set the property
+          <varname>dfs.block.local-path-access.user</varname> to be the <emphasis>only</emphasis>
+        user that can use the shortcut. This has to be the user that started HBase. Then in
+        hbase-site.xml, set <varname>dfs.client.read.shortcircuit</varname> to be
+          <varname>true</varname>
       </para>
+        
       <para> Services -- at least the HBase RegionServers -- will need to be restarted in order to
         pick up the new configurations. </para>
       <note

http://git-wip-us.apache.org/repos/asf/hbase/blob/28164872/src/main/docbkx/preface.xml
----------------------------------------------------------------------
diff --git a/src/main/docbkx/preface.xml b/src/main/docbkx/preface.xml
index 582d04a..a8f6895 100644
--- a/src/main/docbkx/preface.xml
+++ b/src/main/docbkx/preface.xml
@@ -43,7 +43,7 @@
         <title>About This Guide</title>
         <para>This reference guide is a work in progress. The source for this guide can be found in
             the <filename>src/main/docbkx</filename> directory of the HBase source. This reference
-            guide is marked up using <link xlink:href="http://www.docbook.com/">DocBook</link> from
+            guide is marked up using <link xlink:href="http://www.docbook.org/">DocBook</link> from
             which the the finished guide is generated as part of the 'site' build target. Run
             <programlisting language="bourne">mvn site</programlisting> to generate this documentation. Amendments and
             improvements to the documentation are welcomed. Click <link

http://git-wip-us.apache.org/repos/asf/hbase/blob/28164872/src/main/docbkx/rpc.xml
----------------------------------------------------------------------
diff --git a/src/main/docbkx/rpc.xml b/src/main/docbkx/rpc.xml
index 5745a33..2e5dd5f 100644
--- a/src/main/docbkx/rpc.xml
+++ b/src/main/docbkx/rpc.xml
@@ -104,7 +104,8 @@
 
                 <section>
                     <title>&lt;preamble&gt;</title>
-                    <para><programlisting>&lt;MAGIC 4 byte integer&gt; &lt;1 byte RPC Format Version&gt; &lt;1 byte auth type&gt;<footnote><para> We need the auth method spec. here so the connection header is encoded if auth enabled.</para></footnote></programlisting></para>
+                    <programlisting>&lt;MAGIC 4 byte integer&gt; &lt;1 byte RPC Format Version&gt; &lt;1 byte auth type&gt;</programlisting>
+                    <para> We need the auth method spec. here so the connection header is encoded if auth enabled.</para>
                     <para>E.g.: HBas0x000x50 -- 4 bytes of MAGIC -- ‘HBas’ -- plus one-byte of
                         version, 0 in this case, and one byte, 0x50 (SIMPLE). of an auth
                         type.</para>

http://git-wip-us.apache.org/repos/asf/hbase/blob/28164872/src/main/docbkx/security.xml
----------------------------------------------------------------------
diff --git a/src/main/docbkx/security.xml b/src/main/docbkx/security.xml
index 4fe5aa0..c74af9b 100644
--- a/src/main/docbkx/security.xml
+++ b/src/main/docbkx/security.xml
@@ -32,11 +32,9 @@
   <section
     xml:id="hbase.secure.configuration">
     <title>Secure Client Access to Apache HBase</title>
-    <para>Newer releases of Apache HBase (&gt;= 0.92) support optional SASL authentication of clients<footnote>
-        <para>See also Matteo Bertozzi's article on <link
+    <para>Newer releases of Apache HBase (&gt;= 0.92) support optional SASL authentication of clients. See also Matteo Bertozzi's article on <link
             xlink:href="http://www.cloudera.com/blog/2012/09/understanding-user-authentication-and-authorization-in-apache-hbase/">Understanding
             User Authentication and Authorization in Apache HBase</link>.</para>
-      </footnote>.</para>
     <para>This describes how to set up Apache HBase and clients for connection to secure HBase
       resources.</para>
 
@@ -339,11 +337,9 @@ grant 'rest_server', 'RWCA'
   <section
     xml:id="hbase.secure.simpleconfiguration">
     <title>Simple User Access to Apache HBase</title>
-    <para>Newer releases of Apache HBase (&gt;= 0.92) support optional SASL authentication of clients<footnote>
-        <para>See also Matteo Bertozzi's article on <link
+    <para>Newer releases of Apache HBase (&gt;= 0.92) support optional SASL authentication of clients. See also Matteo Bertozzi's article on <link
             xlink:href="http://www.cloudera.com/blog/2012/09/understanding-user-authentication-and-authorization-in-apache-hbase/">Understanding
             User Authentication and Authorization in Apache HBase</link>.</para>
-      </footnote>.</para>
     <para>This describes how to set up Apache HBase and clients for simple user access to HBase
       resources.</para>
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/28164872/src/main/docbkx/troubleshooting.xml
----------------------------------------------------------------------
diff --git a/src/main/docbkx/troubleshooting.xml b/src/main/docbkx/troubleshooting.xml
index ec6c1cf..f48e26f 100644
--- a/src/main/docbkx/troubleshooting.xml
+++ b/src/main/docbkx/troubleshooting.xml
@@ -233,19 +233,16 @@ export SERVER_GC_OPTS="$SERVER_GC_OPTS -XX:NewSize=64m -XX:MaxNewSize=64m"
     <section
       xml:id="trouble.resources.lists">
       <title>Mailing Lists</title>
-      <para>Ask a question on the <link
-          xlink:href="http://hbase.apache.org/mail-lists.html">Apache HBase mailing lists</link>.
-        The 'dev' mailing list is aimed at the community of developers actually building Apache
-        HBase and for features currently under development, and 'user' is generally used for
-        questions on released versions of Apache HBase. Before going to the mailing list, make sure
-        your question has not already been answered by searching the mailing list archives first.
-        Use <xref
-          linkend="trouble.resources.searchhadoop" />. Take some time crafting your question<footnote>
-          <para>See <link
-              xlink:href="http://www.mikeash.com/getting_answers.html">Getting Answers</link></para>
-        </footnote>; a quality question that includes all context and exhibits evidence the author
-        has tried to find answers in the manual and out on lists is more likely to get a prompt
-        response. </para>
+      <para>Ask a question on the <link xlink:href="http://hbase.apache.org/mail-lists.html">Apache
+          HBase mailing lists</link>. The 'dev' mailing list is aimed at the community of developers
+        actually building Apache HBase and for features currently under development, and 'user' is
+        generally used for questions on released versions of Apache HBase. Before going to the
+        mailing list, make sure your question has not already been answered by searching the mailing
+        list archives first. Use <xref linkend="trouble.resources.searchhadoop"/>. Take some time
+        crafting your question. See <link xlink:href="http://www.mikeash.com/getting_answers.html"
+          >Getting Answers</link> for ideas on crafting good questions. A quality question that
+        includes all context and exhibits evidence the author has tried to find answers in the
+        manual and out on lists is more likely to get a prompt response. </para>
     </section>
     <section
       xml:id="trouble.resources.irc">

http://git-wip-us.apache.org/repos/asf/hbase/blob/28164872/src/main/docbkx/upgrading.xml
----------------------------------------------------------------------
diff --git a/src/main/docbkx/upgrading.xml b/src/main/docbkx/upgrading.xml
index 2ba0143..b0a248b 100644
--- a/src/main/docbkx/upgrading.xml
+++ b/src/main/docbkx/upgrading.xml
@@ -71,12 +71,10 @@
                 out the jars of one version and replace them with the jars of another, compatible
                 version and all will just work. Unless otherwise specified, HBase point versions are
                 binary compatible. You can safely do rolling upgrades between binary compatible
-                versions; i.e. across point versions: e.g. from 0.94.5 to 0.94.6<footnote>
-                    <para>See <link
+                versions; i.e. across point versions: e.g. from 0.94.5 to 0.94.6. See <link
                             xlink:href="http://search-hadoop.com/m/bOOvwHGW981/Does+compatibility+between+versions+also+mean+binary+compatibility%253F&amp;subj=Re+Does+compatibility+between+versions+also+mean+binary+compatibility+">Does
                             compatibility between versions also mean binary compatibility?</link>
                         discussion on the hbaes dev mailing list. </para>
-                </footnote>. </para>
         </section>
         <section
             xml:id="hbase.rolling.restart">
@@ -506,12 +504,10 @@ Successfully completed Log splitting
             to change this (The 'normal'/default value is 64MB (67108864)). Run the script
                 <filename>bin/set_meta_memstore_size.rb</filename>. This will make the necessary
             edit to your <varname>.META.</varname> schema. Failure to run this change will make for
-            a slow cluster <footnote>
-                <para> See <link
+            a slow cluster. See <link
                         xlink:href="https://issues.apache.org/jira/browse/HBASE-3499">HBASE-3499
                         Users upgrading to 0.90.0 need to have their .META. table updated with the
                         right MEMSTORE_SIZE</link>
                 </para>
-            </footnote> . </para>
     </section>
 </chapter>

http://git-wip-us.apache.org/repos/asf/hbase/blob/28164872/src/main/docbkx/zookeeper.xml
----------------------------------------------------------------------
diff --git a/src/main/docbkx/zookeeper.xml b/src/main/docbkx/zookeeper.xml
index 9f34a8f..206ccf5 100644
--- a/src/main/docbkx/zookeeper.xml
+++ b/src/main/docbkx/zookeeper.xml
@@ -52,12 +52,10 @@
       <varname>hbase.zookeeper.property.clientPort</varname> property. For all default values used
     by HBase, including ZooKeeper configuration, see <xref
       linkend="hbase_default_configurations" />. Look for the
-      <varname>hbase.zookeeper.property</varname> prefix <footnote>
-      <para>For the full list of ZooKeeper configurations, see ZooKeeper's
+      <varname>hbase.zookeeper.property</varname> prefix. For the full list of ZooKeeper configurations, see ZooKeeper's
           <filename>zoo.cfg</filename>. HBase does not ship with a <filename>zoo.cfg</filename> so
         you will need to browse the <filename>conf</filename> directory in an appropriate ZooKeeper
         download.</para>
-    </footnote></para>
 
   <para>You must at least list the ensemble servers in <filename>hbase-site.xml</filename> using the
       <varname>hbase.zookeeper.quorum</varname> property. This property defaults to a single