You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by gi...@apache.org on 2020/05/04 14:47:35 UTC

[hbase-site] branch asf-site updated: Published site at c44bb2e99aa4a52280df74d46ff556cdfd14dac7.

This is an automated email from the ASF dual-hosted git repository.

git-site-role pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/hbase-site.git


The following commit(s) were added to refs/heads/asf-site by this push:
     new 05ecb54  Published site at c44bb2e99aa4a52280df74d46ff556cdfd14dac7.
05ecb54 is described below

commit 05ecb54ce783bb3db119541799a5442260fb8232
Author: jenkins <bu...@apache.org>
AuthorDate: Mon May 4 14:47:17 2020 +0000

    Published site at c44bb2e99aa4a52280df74d46ff556cdfd14dac7.
---
 acid-semantics.html                                |    2 +-
 apache_hbase_reference_guide.pdf                   |    4 +-
 book.html                                          |    2 +-
 bulk-loads.html                                    |    2 +-
 checkstyle-aggregate.html                          |   22 +-
 checkstyle.rss                                     |   16 +-
 coc.html                                           |    2 +-
 dependencies.html                                  |    2 +-
 dependency-convergence.html                        |    2 +-
 dependency-info.html                               |    2 +-
 dependency-management.html                         |    2 +-
 devapidocs/constant-values.html                    |   36 +-
 devapidocs/index-all.html                          |   16 +-
 .../apache/hadoop/hbase/backup/package-tree.html   |    4 +-
 .../org/apache/hadoop/hbase/class-use/Cell.html    |   34 +-
 .../apache/hadoop/hbase/class-use/TableName.html   |    2 +-
 .../client/class-use/AsyncClusterConnection.html   |    2 +-
 .../apache/hadoop/hbase/client/package-tree.html   |   18 +-
 .../hadoop/hbase/coprocessor/package-tree.html     |    2 +-
 .../apache/hadoop/hbase/executor/package-tree.html |    2 +-
 .../apache/hadoop/hbase/filter/package-tree.html   |    4 +-
 .../hadoop/hbase/hbtop/field/package-tree.html     |    2 +-
 .../hadoop/hbase/hbtop/terminal/package-tree.html  |    2 +-
 .../org/apache/hadoop/hbase/http/package-tree.html |    2 +-
 .../apache/hadoop/hbase/io/hfile/package-tree.html |    4 +-
 .../org/apache/hadoop/hbase/ipc/package-tree.html  |    2 +-
 .../hbase/mapreduce/HashTable.HashMapper.html      |   16 +-
 .../hbase/mapreduce/HashTable.ResultHasher.html    |   43 +-
 .../mapreduce/HashTable.TableHash.Reader.html      |   26 +-
 .../hbase/mapreduce/HashTable.TableHash.html       |   75 +-
 .../apache/hadoop/hbase/mapreduce/HashTable.html   |   59 +-
 .../SyncTable.SyncMapper.CellScanner.html          |   18 +-
 .../mapreduce/SyncTable.SyncMapper.Counter.html    |   34 +-
 .../hbase/mapreduce/SyncTable.SyncMapper.html      |  138 +-
 .../apache/hadoop/hbase/mapreduce/SyncTable.html   |   84 +-
 .../hadoop/hbase/mapreduce/package-tree.html       |    4 +-
 .../hbase/master/assignment/package-tree.html      |    2 +-
 .../hadoop/hbase/master/balancer/package-tree.html |    2 +-
 .../apache/hadoop/hbase/master/package-tree.html   |    8 +-
 .../hbase/master/procedure/package-tree.html       |    2 +-
 .../hadoop/hbase/monitoring/package-tree.html      |    2 +-
 .../org/apache/hadoop/hbase/package-tree.html      |   14 +-
 .../hadoop/hbase/procedure2/package-tree.html      |    6 +-
 .../hbase/procedure2/store/wal/package-tree.html   |    2 +-
 .../apache/hadoop/hbase/quotas/package-tree.html   |    4 +-
 .../hadoop/hbase/regionserver/package-tree.html    |   20 +-
 .../regionserver/querymatcher/package-tree.html    |    2 +-
 .../hbase/regionserver/wal/package-tree.html       |    4 +-
 .../hadoop/hbase/replication/package-tree.html     |    2 +-
 .../replication/regionserver/package-tree.html     |    2 +-
 .../hadoop/hbase/security/access/package-tree.html |    6 +-
 .../apache/hadoop/hbase/security/package-tree.html |    4 +-
 .../apache/hadoop/hbase/thrift/package-tree.html   |    2 +-
 .../hadoop/hbase/tool/BulkLoadHFilesTool.html      |   70 +-
 .../class-use/BulkLoadHFiles.LoadQueueItem.html    |    4 +-
 .../org/apache/hadoop/hbase/util/package-tree.html |    8 +-
 .../src-html/org/apache/hadoop/hbase/Version.html  |    4 +-
 .../hbase/mapreduce/HashTable.HashMapper.html      | 1350 +++++++++--------
 .../hbase/mapreduce/HashTable.ResultHasher.html    | 1350 +++++++++--------
 .../mapreduce/HashTable.TableHash.Reader.html      | 1350 +++++++++--------
 .../hbase/mapreduce/HashTable.TableHash.html       | 1350 +++++++++--------
 .../apache/hadoop/hbase/mapreduce/HashTable.html   | 1350 +++++++++--------
 .../SyncTable.SyncMapper.CellScanner.html          | 1603 ++++++++++----------
 .../mapreduce/SyncTable.SyncMapper.Counter.html    | 1603 ++++++++++----------
 .../hbase/mapreduce/SyncTable.SyncMapper.html      | 1603 ++++++++++----------
 .../apache/hadoop/hbase/mapreduce/SyncTable.html   | 1603 ++++++++++----------
 .../tool/BulkLoadHFilesTool.BulkHFileVisitor.html  | 1450 +++++++++---------
 .../hadoop/hbase/tool/BulkLoadHFilesTool.html      | 1450 +++++++++---------
 downloads.html                                     |    2 +-
 export_control.html                                |    2 +-
 index.html                                         |    2 +-
 issue-tracking.html                                |    2 +-
 mail-lists.html                                    |    2 +-
 metrics.html                                       |    2 +-
 old_news.html                                      |    2 +-
 plugin-management.html                             |    2 +-
 plugins.html                                       |    2 +-
 poweredbyhbase.html                                |    2 +-
 project-info.html                                  |    2 +-
 project-reports.html                               |    2 +-
 project-summary.html                               |    2 +-
 pseudo-distributed.html                            |    2 +-
 replication.html                                   |    2 +-
 resources.html                                     |    2 +-
 source-repository.html                             |    2 +-
 sponsors.html                                      |    2 +-
 supportingprojects.html                            |    2 +-
 team-list.html                                     |    2 +-
 testdevapidocs/allclasses-frame.html               |    1 -
 testdevapidocs/allclasses-noframe.html             |    1 -
 testdevapidocs/index-all.html                      |   18 +-
 .../hadoop/hbase/chaos/actions/package-tree.html   |    2 +-
 .../hadoop/hbase/class-use/HBaseClassTestRule.html |   14 +-
 .../hadoop/hbase/mapreduce/TestSyncTable.html      |   86 +-
 .../org/apache/hadoop/hbase/package-tree.html      |   10 +-
 .../hadoop/hbase/procedure/package-tree.html       |    8 +-
 .../hadoop/hbase/procedure2/package-tree.html      |    4 +-
 .../hadoop/hbase/regionserver/package-tree.html    |    6 +-
 .../org/apache/hadoop/hbase/test/package-tree.html |    2 +-
 .../hadoop/hbase/tool/TestBulkLoadHFiles.html      |  188 +--
 .../hbase/tool/TestBulkLoadHFilesByFamily.html     |  332 ----
 .../tool/TestBulkLoadHFilesSplitRecovery.html      |    4 +-
 .../hbase/tool/TestSecureBulkLoadHFiles.html       |    2 +-
 .../hbase/tool/class-use/TestBulkLoadHFiles.html   |    4 -
 .../tool/class-use/TestBulkLoadHFilesByFamily.html |  125 --
 .../apache/hadoop/hbase/tool/package-frame.html    |    1 -
 .../apache/hadoop/hbase/tool/package-summary.html  |   12 +-
 .../org/apache/hadoop/hbase/tool/package-tree.html |    1 -
 .../org/apache/hadoop/hbase/wal/package-tree.html  |    4 +-
 testdevapidocs/overview-tree.html                  |    1 -
 .../hadoop/hbase/mapreduce/TestSyncTable.html      | 1096 ++++++-------
 .../hadoop/hbase/tool/TestBulkLoadHFiles.html      | 1449 +++++++++---------
 .../hbase/tool/TestBulkLoadHFilesByFamily.html     |  117 --
 113 files changed, 10178 insertions(+), 10269 deletions(-)

diff --git a/acid-semantics.html b/acid-semantics.html
index 2b2307a..b02c550 100644
--- a/acid-semantics.html
+++ b/acid-semantics.html
@@ -467,7 +467,7 @@
         <div class="row">
             <p>Copyright &copy;2007&#x2013;2020
 <a href="https://www.apache.org/">The Apache Software Foundation</a>.
-All rights reserved.        <li id="publishDate" class="pull-right">Last Published: 2020-05-02</li>
+All rights reserved.        <li id="publishDate" class="pull-right">Last Published: 2020-05-04</li>
 </p>
         </div>
         <p id="poweredBy" class="pull-right"><a href="http://maven.apache.org/" title="Built by Maven" class="poweredBy"><img class="builtBy" alt="Built by Maven" src="./images/logos/maven-feather.png" /></a>
diff --git a/apache_hbase_reference_guide.pdf b/apache_hbase_reference_guide.pdf
index a5ebfa9..c2f2f62 100644
--- a/apache_hbase_reference_guide.pdf
+++ b/apache_hbase_reference_guide.pdf
@@ -5,8 +5,8 @@
 /Author (Apache HBase Team)
 /Creator (Asciidoctor PDF 1.5.0.rc.2, based on Prawn 2.2.2)
 /Producer (Apache HBase Team)
-/ModDate (D:20200502143249+00'00')
-/CreationDate (D:20200502144504+00'00')
+/ModDate (D:20200504143121+00'00')
+/CreationDate (D:20200504144354+00'00')
 >>
 endobj
 2 0 obj
diff --git a/book.html b/book.html
index 20fd911..dd61634 100644
--- a/book.html
+++ b/book.html
@@ -45261,7 +45261,7 @@ org/apache/hadoop/hbase/security/access/AccessControlClient.revoke:(Lorg/apache/
 <div id="footer">
 <div id="footer-text">
 Version 3.0.0-SNAPSHOT<br>
-Last updated 2020-05-02 14:32:49 UTC
+Last updated 2020-05-04 14:31:21 UTC
 </div>
 </div>
 <script type="text/x-mathjax-config">
diff --git a/bulk-loads.html b/bulk-loads.html
index 386870a..10ec38d 100644
--- a/bulk-loads.html
+++ b/bulk-loads.html
@@ -172,7 +172,7 @@
         <div class="row">
             <p>Copyright &copy;2007&#x2013;2020
 <a href="https://www.apache.org/">The Apache Software Foundation</a>.
-All rights reserved.        <li id="publishDate" class="pull-right">Last Published: 2020-05-02</li>
+All rights reserved.        <li id="publishDate" class="pull-right">Last Published: 2020-05-04</li>
 </p>
         </div>
         <p id="poweredBy" class="pull-right"><a href="http://maven.apache.org/" title="Built by Maven" class="poweredBy"><img class="builtBy" alt="Built by Maven" src="./images/logos/maven-feather.png" /></a>
diff --git a/checkstyle-aggregate.html b/checkstyle-aggregate.html
index 5adf90e..7cb7563 100644
--- a/checkstyle-aggregate.html
+++ b/checkstyle-aggregate.html
@@ -166,7 +166,7 @@
 <th><img src="images/icon_warning_sml.gif" alt="" />&#160;Warnings</th>
 <th><img src="images/icon_error_sml.gif" alt="" />&#160;Errors</th></tr>
 <tr class="b">
-<td>4237</td>
+<td>4236</td>
 <td>0</td>
 <td>0</td>
 <td>9851</td></tr></table></div>
@@ -7369,7 +7369,7 @@
 <tr class="a">
 <td>annotation</td>
 <td><a class="externalLink" href="http://checkstyle.sourceforge.net/config_annotation.html#MissingDeprecated">MissingDeprecated</a></td>
-<td>10</td>
+<td>9</td>
 <td><img src="images/icon_error_sml.gif" alt="" />&#160;Error</td></tr>
 <tr class="b">
 <td>blocks</td>
@@ -7468,7 +7468,7 @@
 <td><a class="externalLink" href="http://checkstyle.sourceforge.net/config_javadoc.html#JavadocTagContinuationIndentation">JavadocTagContinuationIndentation</a>
 <ul>
 <li>offset: <tt>&quot;2&quot;</tt></li></ul></td>
-<td>595</td>
+<td>596</td>
 <td><img src="images/icon_error_sml.gif" alt="" />&#160;Error</td></tr>
 <tr class="b">
 <td></td>
@@ -8489,8 +8489,8 @@
 <td>600</td></tr>
 <tr class="b">
 <td><img src="images/icon_error_sml.gif" alt="" />&#160;Error</td>
-<td>annotation</td>
-<td>MissingDeprecated</td>
+<td>javadoc</td>
+<td>JavadocTagContinuationIndentation</td>
 <td>Javadoc comment at column 33 has parse error. Details: no viable alternative at input '&lt;Map.Entry&lt;' while parsing HTML_ELEMENT</td>
 <td>642</td></tr>
 <tr class="a">
@@ -36584,7 +36584,7 @@
 <td>javadoc</td>
 <td>NonEmptyAtclauseDescription</td>
 <td>At-clause should have a non-empty description.</td>
-<td>320</td></tr></table></div>
+<td>322</td></tr></table></div>
 <div class="section">
 <h3 id="org.apache.hadoop.hbase.mapreduce.IdentityTableMapper.java">org/apache/hadoop/hbase/mapreduce/IdentityTableMapper.java</h3>
 <table border="0" class="table table-striped">
@@ -37481,25 +37481,25 @@
 <td>indentation</td>
 <td>Indentation</td>
 <td>'method def' child has incorrect indentation level 8, expected level should be 6.</td>
-<td>236</td></tr>
+<td>244</td></tr>
 <tr class="a">
 <td><img src="images/icon_error_sml.gif" alt="" />&#160;Error</td>
 <td>indentation</td>
 <td>Indentation</td>
 <td>'method def' child has incorrect indentation level 8, expected level should be 6.</td>
-<td>237</td></tr>
+<td>245</td></tr>
 <tr class="b">
 <td><img src="images/icon_error_sml.gif" alt="" />&#160;Error</td>
 <td>indentation</td>
 <td>Indentation</td>
 <td>'method def' child has incorrect indentation level 8, expected level should be 6.</td>
-<td>239</td></tr>
+<td>247</td></tr>
 <tr class="a">
 <td><img src="images/icon_error_sml.gif" alt="" />&#160;Error</td>
 <td>indentation</td>
 <td>Indentation</td>
 <td>'method def modifier' has incorrect indentation level 5, expected level should be 4.</td>
-<td>610</td></tr></table></div>
+<td>637</td></tr></table></div>
 <div class="section">
 <h3 id="org.apache.hadoop.hbase.mapreduce.TableInputFormat.java">org/apache/hadoop/hbase/mapreduce/TableInputFormat.java</h3>
 <table border="0" class="table table-striped">
@@ -79553,7 +79553,7 @@
         <div class="row">
             <p>Copyright &copy;2007&#x2013;2020
 <a href="https://www.apache.org/">The Apache Software Foundation</a>.
-All rights reserved.        <li id="publishDate" class="pull-right">Last Published: 2020-05-02</li>
+All rights reserved.        <li id="publishDate" class="pull-right">Last Published: 2020-05-04</li>
 </p>
         </div>
         <p id="poweredBy" class="pull-right"><a href="http://maven.apache.org/" title="Built by Maven" class="poweredBy"><img class="builtBy" alt="Built by Maven" src="./images/logos/maven-feather.png" /></a>
diff --git a/checkstyle.rss b/checkstyle.rss
index 19cafa9..6649d52 100644
--- a/checkstyle.rss
+++ b/checkstyle.rss
@@ -25,7 +25,7 @@ under the License.
     <language>en-us</language>
     <copyright>&#169;2007 - 2020 The Apache Software Foundation</copyright>
     <item>
-      <title>File: 4237,
+      <title>File: 4236,
              Errors: 9851,
              Warnings: 0,
              Infos: 0
@@ -8908,20 +8908,6 @@ under the License.
               </tr>
                           <tr>
                 <td>
-                  <a href="https://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.tool.TestBulkLoadHFilesByFamily.java">org/apache/hadoop/hbase/tool/TestBulkLoadHFilesByFamily.java</a>
-                </td>
-                <td>
-                  0
-                </td>
-                <td>
-                  0
-                </td>
-                <td>
-                  0
-                </td>
-              </tr>
-                          <tr>
-                <td>
                   <a href="https://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.master.balancer.TestStochasticLoadBalancerSmallCluster.java">org/apache/hadoop/hbase/master/balancer/TestStochasticLoadBalancerSmallCluster.java</a>
                 </td>
                 <td>
diff --git a/coc.html b/coc.html
index 2c843eb..c1be430 100644
--- a/coc.html
+++ b/coc.html
@@ -241,7 +241,7 @@ email to <a class="externalLink" href="mailto:private@hbase.apache.org">the priv
         <div class="row">
             <p>Copyright &copy;2007&#x2013;2020
 <a href="https://www.apache.org/">The Apache Software Foundation</a>.
-All rights reserved.        <li id="publishDate" class="pull-right">Last Published: 2020-05-02</li>
+All rights reserved.        <li id="publishDate" class="pull-right">Last Published: 2020-05-04</li>
 </p>
         </div>
         <p id="poweredBy" class="pull-right"><a href="http://maven.apache.org/" title="Built by Maven" class="poweredBy"><img class="builtBy" alt="Built by Maven" src="./images/logos/maven-feather.png" /></a>
diff --git a/dependencies.html b/dependencies.html
index 02c002a..8bd1a05 100644
--- a/dependencies.html
+++ b/dependencies.html
@@ -313,7 +313,7 @@
         <div class="row">
             <p>Copyright &copy;2007&#x2013;2020
 <a href="https://www.apache.org/">The Apache Software Foundation</a>.
-All rights reserved.        <li id="publishDate" class="pull-right">Last Published: 2020-05-02</li>
+All rights reserved.        <li id="publishDate" class="pull-right">Last Published: 2020-05-04</li>
 </p>
         </div>
         <p id="poweredBy" class="pull-right"><a href="http://maven.apache.org/" title="Built by Maven" class="poweredBy"><img class="builtBy" alt="Built by Maven" src="./images/logos/maven-feather.png" /></a>
diff --git a/dependency-convergence.html b/dependency-convergence.html
index 5b3ef02..4f167ac 100644
--- a/dependency-convergence.html
+++ b/dependency-convergence.html
@@ -774,7 +774,7 @@
         <div class="row">
             <p>Copyright &copy;2007&#x2013;2020
 <a href="https://www.apache.org/">The Apache Software Foundation</a>.
-All rights reserved.        <li id="publishDate" class="pull-right">Last Published: 2020-05-02</li>
+All rights reserved.        <li id="publishDate" class="pull-right">Last Published: 2020-05-04</li>
 </p>
         </div>
         <p id="poweredBy" class="pull-right"><a href="http://maven.apache.org/" title="Built by Maven" class="poweredBy"><img class="builtBy" alt="Built by Maven" src="./images/logos/maven-feather.png" /></a>
diff --git a/dependency-info.html b/dependency-info.html
index f3f5621..a8d9ec7 100644
--- a/dependency-info.html
+++ b/dependency-info.html
@@ -194,7 +194,7 @@
         <div class="row">
             <p>Copyright &copy;2007&#x2013;2020
 <a href="https://www.apache.org/">The Apache Software Foundation</a>.
-All rights reserved.        <li id="publishDate" class="pull-right">Last Published: 2020-05-02</li>
+All rights reserved.        <li id="publishDate" class="pull-right">Last Published: 2020-05-04</li>
 </p>
         </div>
         <p id="poweredBy" class="pull-right"><a href="http://maven.apache.org/" title="Built by Maven" class="poweredBy"><img class="builtBy" alt="Built by Maven" src="./images/logos/maven-feather.png" /></a>
diff --git a/dependency-management.html b/dependency-management.html
index 4440ceb..dc1ac11 100644
--- a/dependency-management.html
+++ b/dependency-management.html
@@ -1054,7 +1054,7 @@
         <div class="row">
             <p>Copyright &copy;2007&#x2013;2020
 <a href="https://www.apache.org/">The Apache Software Foundation</a>.
-All rights reserved.        <li id="publishDate" class="pull-right">Last Published: 2020-05-02</li>
+All rights reserved.        <li id="publishDate" class="pull-right">Last Published: 2020-05-04</li>
 </p>
         </div>
         <p id="poweredBy" class="pull-right"><a href="http://maven.apache.org/" title="Built by Maven" class="poweredBy"><img class="builtBy" alt="Built by Maven" src="./images/logos/maven-feather.png" /></a>
diff --git a/devapidocs/constant-values.html b/devapidocs/constant-values.html
index 9da9487..f467a76 100644
--- a/devapidocs/constant-values.html
+++ b/devapidocs/constant-values.html
@@ -4165,14 +4165,14 @@
 <!--   -->
 </a><code>public&nbsp;static&nbsp;final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a></code></td>
 <td><code><a href="org/apache/hadoop/hbase/Version.html#date">date</a></code></td>
-<td class="colLast"><code>"Sat May  2 14:41:34 UTC 2020"</code></td>
+<td class="colLast"><code>"Mon May  4 14:39:58 UTC 2020"</code></td>
 </tr>
 <tr class="rowColor">
 <td class="colFirst"><a name="org.apache.hadoop.hbase.Version.revision">
 <!--   -->
 </a><code>public&nbsp;static&nbsp;final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a></code></td>
 <td><code><a href="org/apache/hadoop/hbase/Version.html#revision">revision</a></code></td>
-<td class="colLast"><code>"4f9eecbe61934bb6a28f0d6dde7c94ff04b649ed"</code></td>
+<td class="colLast"><code>"c44bb2e99aa4a52280df74d46ff556cdfd14dac7"</code></td>
 </tr>
 <tr class="altColor">
 <td class="colFirst"><a name="org.apache.hadoop.hbase.Version.srcChecksum">
@@ -12842,34 +12842,41 @@
 <td class="colLast"><code>"hashes"</code></td>
 </tr>
 <tr class="rowColor">
+<td class="colFirst"><a name="org.apache.hadoop.hbase.mapreduce.HashTable.IGNORE_TIMESTAMPS">
+<!--   -->
+</a><code>static&nbsp;final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a></code></td>
+<td><code><a href="org/apache/hadoop/hbase/mapreduce/HashTable.html#IGNORE_TIMESTAMPS">IGNORE_TIMESTAMPS</a></code></td>
+<td class="colLast"><code>"ignoreTimestamps"</code></td>
+</tr>
+<tr class="altColor">
 <td class="colFirst"><a name="org.apache.hadoop.hbase.mapreduce.HashTable.MANIFEST_FILE_NAME">
 <!--   -->
 </a><code>static&nbsp;final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a></code></td>
 <td><code><a href="org/apache/hadoop/hbase/mapreduce/HashTable.html#MANIFEST_FILE_NAME">MANIFEST_FILE_NAME</a></code></td>
 <td class="colLast"><code>"manifest"</code></td>
 </tr>
-<tr class="altColor">
+<tr class="rowColor">
 <td class="colFirst"><a name="org.apache.hadoop.hbase.mapreduce.HashTable.NUM_ARGS">
 <!--   -->
 </a><code>private&nbsp;static&nbsp;final&nbsp;int</code></td>
 <td><code><a href="org/apache/hadoop/hbase/mapreduce/HashTable.html#NUM_ARGS">NUM_ARGS</a></code></td>
 <td class="colLast"><code>2</code></td>
 </tr>
-<tr class="rowColor">
+<tr class="altColor">
 <td class="colFirst"><a name="org.apache.hadoop.hbase.mapreduce.HashTable.OUTPUT_DATA_FILE_PREFIX">
 <!--   -->
 </a><code>static&nbsp;final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a></code></td>
 <td><code><a href="org/apache/hadoop/hbase/mapreduce/HashTable.html#OUTPUT_DATA_FILE_PREFIX">OUTPUT_DATA_FILE_PREFIX</a></code></td>
 <td class="colLast"><code>"part-r-"</code></td>
 </tr>
-<tr class="altColor">
+<tr class="rowColor">
 <td class="colFirst"><a name="org.apache.hadoop.hbase.mapreduce.HashTable.PARTITIONS_FILE_NAME">
 <!--   -->
 </a><code>static&nbsp;final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a></code></td>
 <td><code><a href="org/apache/hadoop/hbase/mapreduce/HashTable.html#PARTITIONS_FILE_NAME">PARTITIONS_FILE_NAME</a></code></td>
 <td class="colLast"><code>"partitions"</code></td>
 </tr>
-<tr class="rowColor">
+<tr class="altColor">
 <td class="colFirst"><a name="org.apache.hadoop.hbase.mapreduce.HashTable.TMP_MANIFEST_FILE_NAME">
 <!--   -->
 </a><code>private&nbsp;static&nbsp;final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a></code></td>
@@ -13582,41 +13589,48 @@
 <td class="colLast"><code>"sync.table.dry.run"</code></td>
 </tr>
 <tr class="rowColor">
+<td class="colFirst"><a name="org.apache.hadoop.hbase.mapreduce.SyncTable.IGNORE_TIMESTAMPS">
+<!--   -->
+</a><code>static&nbsp;final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a></code></td>
+<td><code><a href="org/apache/hadoop/hbase/mapreduce/SyncTable.html#IGNORE_TIMESTAMPS">IGNORE_TIMESTAMPS</a></code></td>
+<td class="colLast"><code>"sync.table.ignore.timestamps"</code></td>
+</tr>
+<tr class="altColor">
 <td class="colFirst"><a name="org.apache.hadoop.hbase.mapreduce.SyncTable.NUM_ARGS">
 <!--   -->
 </a><code>private&nbsp;static&nbsp;final&nbsp;int</code></td>
 <td><code><a href="org/apache/hadoop/hbase/mapreduce/SyncTable.html#NUM_ARGS">NUM_ARGS</a></code></td>
 <td class="colLast"><code>3</code></td>
 </tr>
-<tr class="altColor">
+<tr class="rowColor">
 <td class="colFirst"><a name="org.apache.hadoop.hbase.mapreduce.SyncTable.SOURCE_HASH_DIR_CONF_KEY">
 <!--   -->
 </a><code>static&nbsp;final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a></code></td>
 <td><code><a href="org/apache/hadoop/hbase/mapreduce/SyncTable.html#SOURCE_HASH_DIR_CONF_KEY">SOURCE_HASH_DIR_CONF_KEY</a></code></td>
 <td class="colLast"><code>"sync.table.source.hash.dir"</code></td>
 </tr>
-<tr class="rowColor">
+<tr class="altColor">
 <td class="colFirst"><a name="org.apache.hadoop.hbase.mapreduce.SyncTable.SOURCE_TABLE_CONF_KEY">
 <!--   -->
 </a><code>static&nbsp;final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a></code></td>
 <td><code><a href="org/apache/hadoop/hbase/mapreduce/SyncTable.html#SOURCE_TABLE_CONF_KEY">SOURCE_TABLE_CONF_KEY</a></code></td>
 <td class="colLast"><code>"sync.table.source.table.name"</code></td>
 </tr>
-<tr class="altColor">
+<tr class="rowColor">
 <td class="colFirst"><a name="org.apache.hadoop.hbase.mapreduce.SyncTable.SOURCE_ZK_CLUSTER_CONF_KEY">
 <!--   -->
 </a><code>static&nbsp;final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a></code></td>
 <td><code><a href="org/apache/hadoop/hbase/mapreduce/SyncTable.html#SOURCE_ZK_CLUSTER_CONF_KEY">SOURCE_ZK_CLUSTER_CONF_KEY</a></code></td>
 <td class="colLast"><code>"sync.table.source.zk.cluster"</code></td>
 </tr>
-<tr class="rowColor">
+<tr class="altColor">
 <td class="colFirst"><a name="org.apache.hadoop.hbase.mapreduce.SyncTable.TARGET_TABLE_CONF_KEY">
 <!--   -->
 </a><code>static&nbsp;final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a></code></td>
 <td><code><a href="org/apache/hadoop/hbase/mapreduce/SyncTable.html#TARGET_TABLE_CONF_KEY">TARGET_TABLE_CONF_KEY</a></code></td>
 <td class="colLast"><code>"sync.table.target.table.name"</code></td>
 </tr>
-<tr class="altColor">
+<tr class="rowColor">
 <td class="colFirst"><a name="org.apache.hadoop.hbase.mapreduce.SyncTable.TARGET_ZK_CLUSTER_CONF_KEY">
 <!--   -->
 </a><code>static&nbsp;final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a></code></td>
diff --git a/devapidocs/index-all.html b/devapidocs/index-all.html
index cb8ab32..e477607 100644
--- a/devapidocs/index-all.html
+++ b/devapidocs/index-all.html
@@ -10903,6 +10903,8 @@
 <div class="block">Check is a server of same host and port already exists,
  if not, or the existed one got a smaller start code, record it.</div>
 </dd>
+<dt><span class="memberNameLink"><a href="org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#checkAndResetTimestamp-org.apache.hadoop.hbase.Cell-">checkAndResetTimestamp(Cell)</a></span> - Method in class org.apache.hadoop.hbase.mapreduce.<a href="org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html" title="class in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper</a></dt>
+<dd>&nbsp;</dd>
 <dt><span class="memberNameLink"><a href="org/apache/hadoop/hbase/regionserver/HRegion.html#checkAndRowMutate-byte:A-byte:A-byte:A-org.apache.hadoop.hbase.CompareOperator-org.apache.hadoop.hbase.filter.ByteArrayComparable-org.apache.hadoop.hbase.io.TimeRange-org.apache.hadoop.hbase.client.RowMutations-">checkAndRowMutate(byte[], byte[], byte[], CompareOperator, ByteArrayComparable, TimeRange, RowMutations)</a></span> - Method in class org.apache.hadoop.hbase.regionserver.<a href="org/apa [...]
 <dd>&nbsp;</dd>
 <dt><span class="memberNameLink"><a href="org/apache/hadoop/hbase/regionserver/HRegion.html#checkAndRowMutate-byte:A-org.apache.hadoop.hbase.filter.Filter-org.apache.hadoop.hbase.io.TimeRange-org.apache.hadoop.hbase.client.RowMutations-">checkAndRowMutate(byte[], Filter, TimeRange, RowMutations)</a></span> - Method in class org.apache.hadoop.hbase.regionserver.<a href="org/apache/hadoop/hbase/regionserver/HRegion.html" title="class in org.apache.hadoop.hbase.regionserver">HRegion</a></dt>
@@ -15698,7 +15700,7 @@
 <dd>&nbsp;</dd>
 <dt><span class="memberNameLink"><a href="org/apache/hadoop/hbase/filter/FilterListBase.html#compareCell-org.apache.hadoop.hbase.Cell-org.apache.hadoop.hbase.Cell-">compareCell(Cell, Cell)</a></span> - Method in class org.apache.hadoop.hbase.filter.<a href="org/apache/hadoop/hbase/filter/FilterListBase.html" title="class in org.apache.hadoop.hbase.filter">FilterListBase</a></dt>
 <dd>&nbsp;</dd>
-<dt><span class="memberNameLink"><a href="org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#compareCellKeysWithinRow-org.apache.hadoop.hbase.Cell-org.apache.hadoop.hbase.Cell-">compareCellKeysWithinRow(Cell, Cell)</a></span> - Static method in class org.apache.hadoop.hbase.mapreduce.<a href="org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html" title="class in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper</a></dt>
+<dt><span class="memberNameLink"><a href="org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#compareCellKeysWithinRow-org.apache.hadoop.hbase.Cell-org.apache.hadoop.hbase.Cell-">compareCellKeysWithinRow(Cell, Cell)</a></span> - Method in class org.apache.hadoop.hbase.mapreduce.<a href="org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html" title="class in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper</a></dt>
 <dd>
 <div class="block">Compare families, qualifiers, and timestamps of the given Cells.</div>
 </dd>
@@ -61691,6 +61693,10 @@
 <dd>&nbsp;</dd>
 <dt><span class="memberNameLink"><a href="org/apache/hadoop/hbase/mapreduce/WALPlayer.html#IGNORE_MISSING_FILES">IGNORE_MISSING_FILES</a></span> - Static variable in class org.apache.hadoop.hbase.mapreduce.<a href="org/apache/hadoop/hbase/mapreduce/WALPlayer.html" title="class in org.apache.hadoop.hbase.mapreduce">WALPlayer</a></dt>
 <dd>&nbsp;</dd>
+<dt><span class="memberNameLink"><a href="org/apache/hadoop/hbase/mapreduce/HashTable.html#IGNORE_TIMESTAMPS">IGNORE_TIMESTAMPS</a></span> - Static variable in class org.apache.hadoop.hbase.mapreduce.<a href="org/apache/hadoop/hbase/mapreduce/HashTable.html" title="class in org.apache.hadoop.hbase.mapreduce">HashTable</a></dt>
+<dd>&nbsp;</dd>
+<dt><span class="memberNameLink"><a href="org/apache/hadoop/hbase/mapreduce/SyncTable.html#IGNORE_TIMESTAMPS">IGNORE_TIMESTAMPS</a></span> - Static variable in class org.apache.hadoop.hbase.mapreduce.<a href="org/apache/hadoop/hbase/mapreduce/SyncTable.html" title="class in org.apache.hadoop.hbase.mapreduce">SyncTable</a></dt>
+<dd>&nbsp;</dd>
 <dt><span class="memberNameLink"><a href="org/apache/hadoop/hbase/tool/BulkLoadHFiles.html#IGNORE_UNMATCHED_CF_CONF_KEY">IGNORE_UNMATCHED_CF_CONF_KEY</a></span> - Static variable in interface org.apache.hadoop.hbase.tool.<a href="org/apache/hadoop/hbase/tool/BulkLoadHFiles.html" title="interface in org.apache.hadoop.hbase.tool">BulkLoadHFiles</a></dt>
 <dd>&nbsp;</dd>
 <dt><span class="memberNameLink"><a href="org/apache/hadoop/hbase/hbtop/RecordFilter.FilterBuilder.html#ignoreCase">ignoreCase</a></span> - Variable in class org.apache.hadoop.hbase.hbtop.<a href="org/apache/hadoop/hbase/hbtop/RecordFilter.FilterBuilder.html" title="class in org.apache.hadoop.hbase.hbtop">RecordFilter.FilterBuilder</a></dt>
@@ -61705,6 +61711,14 @@
 <dd>
 <div class="block"><span class="deprecatedLabel">Deprecated.</span></div>
 &nbsp;</dd>
+<dt><span class="memberNameLink"><a href="org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#ignoreTimestamp">ignoreTimestamp</a></span> - Variable in class org.apache.hadoop.hbase.mapreduce.<a href="org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html" title="class in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper</a></dt>
+<dd>&nbsp;</dd>
+<dt><span class="memberNameLink"><a href="org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#ignoreTimestamps">ignoreTimestamps</a></span> - Variable in class org.apache.hadoop.hbase.mapreduce.<a href="org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html" title="class in org.apache.hadoop.hbase.mapreduce">HashTable.ResultHasher</a></dt>
+<dd>&nbsp;</dd>
+<dt><span class="memberNameLink"><a href="org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#ignoreTimestamps">ignoreTimestamps</a></span> - Variable in class org.apache.hadoop.hbase.mapreduce.<a href="org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html" title="class in org.apache.hadoop.hbase.mapreduce">HashTable.TableHash</a></dt>
+<dd>&nbsp;</dd>
+<dt><span class="memberNameLink"><a href="org/apache/hadoop/hbase/mapreduce/SyncTable.html#ignoreTimestamps">ignoreTimestamps</a></span> - Variable in class org.apache.hadoop.hbase.mapreduce.<a href="org/apache/hadoop/hbase/mapreduce/SyncTable.html" title="class in org.apache.hadoop.hbase.mapreduce">SyncTable</a></dt>
+<dd>&nbsp;</dd>
 <dt><span class="memberNameLink"><a href="org/apache/hadoop/hbase/coordination/ZKSplitLogManagerCoordination.html#ignoreZKDeleteForTesting">ignoreZKDeleteForTesting</a></span> - Variable in class org.apache.hadoop.hbase.coordination.<a href="org/apache/hadoop/hbase/coordination/ZKSplitLogManagerCoordination.html" title="class in org.apache.hadoop.hbase.coordination">ZKSplitLogManagerCoordination</a></dt>
 <dd>&nbsp;</dd>
 <dt><a href="org/apache/hadoop/hbase/exceptions/IllegalArgumentIOException.html" title="class in org.apache.hadoop.hbase.exceptions"><span class="typeNameLink">IllegalArgumentIOException</span></a> - Exception in <a href="org/apache/hadoop/hbase/exceptions/package-summary.html">org.apache.hadoop.hbase.exceptions</a></dt>
diff --git a/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html b/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html
index 53b6659..882a5aa 100644
--- a/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html
@@ -168,9 +168,9 @@
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
 <li type="circle">org.apache.hadoop.hbase.backup.<a href="../../../../../org/apache/hadoop/hbase/backup/BackupRestoreConstants.BackupCommand.html" title="enum in org.apache.hadoop.hbase.backup"><span class="typeNameLink">BackupRestoreConstants.BackupCommand</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.backup.<a href="../../../../../org/apache/hadoop/hbase/backup/BackupType.html" title="enum in org.apache.hadoop.hbase.backup"><span class="typeNameLink">BackupType</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.backup.<a href="../../../../../org/apache/hadoop/hbase/backup/BackupInfo.BackupState.html" title="enum in org.apache.hadoop.hbase.backup"><span class="typeNameLink">BackupInfo.BackupState</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.backup.<a href="../../../../../org/apache/hadoop/hbase/backup/BackupInfo.BackupPhase.html" title="enum in org.apache.hadoop.hbase.backup"><span class="typeNameLink">BackupInfo.BackupPhase</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.backup.<a href="../../../../../org/apache/hadoop/hbase/backup/BackupInfo.BackupState.html" title="enum in org.apache.hadoop.hbase.backup"><span class="typeNameLink">BackupInfo.BackupState</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.backup.<a href="../../../../../org/apache/hadoop/hbase/backup/BackupType.html" title="enum in org.apache.hadoop.hbase.backup"><span class="typeNameLink">BackupType</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/Cell.html b/devapidocs/org/apache/hadoop/hbase/class-use/Cell.html
index 871b0c1..7c9f919 100644
--- a/devapidocs/org/apache/hadoop/hbase/class-use/Cell.html
+++ b/devapidocs/org/apache/hadoop/hbase/class-use/Cell.html
@@ -4175,11 +4175,15 @@ service.</div>
 </tr>
 <tbody>
 <tr class="altColor">
+<td class="colFirst"><code>private <a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a></code></td>
+<td class="colLast"><span class="typeNameLabel">SyncTable.SyncMapper.</span><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#checkAndResetTimestamp-org.apache.hadoop.hbase.Cell-">checkAndResetTimestamp</a></span>(<a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a>&nbsp;sourceCell)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
 <td class="colFirst"><code>private static <a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a></code></td>
 <td class="colLast"><span class="typeNameLabel">Import.</span><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/Import.html#convertKv-org.apache.hadoop.hbase.Cell-java.util.Map-">convertKv</a></span>(<a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a>&nbsp;kv,
          <a href="https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true" title="class or interface in java.util">Map</a>&lt;byte[],byte[]&gt;&nbsp;cfRenameMap)</code>&nbsp;</td>
 </tr>
-<tr class="rowColor">
+<tr class="altColor">
 <td class="colFirst"><code><a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a></code></td>
 <td class="colLast"><span class="typeNameLabel">CellCreator.</span><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/CellCreator.html#create-byte:A-int-int-byte:A-int-int-byte:A-int-int-long-byte:A-int-int-">create</a></span>(byte[]&nbsp;row,
       int&nbsp;roffset,
@@ -4195,7 +4199,7 @@ service.</div>
       int&nbsp;voffset,
       int&nbsp;vlength)</code>&nbsp;</td>
 </tr>
-<tr class="altColor">
+<tr class="rowColor">
 <td class="colFirst"><code><a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a></code></td>
 <td class="colLast"><span class="typeNameLabel">CellCreator.</span><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/CellCreator.html#create-byte:A-int-int-byte:A-int-int-byte:A-int-int-long-byte:A-int-int-java.util.List-">create</a></span>(byte[]&nbsp;row,
       int&nbsp;roffset,
@@ -4212,7 +4216,7 @@ service.</div>
       int&nbsp;vlength,
       <a href="https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true" title="class or interface in java.util">List</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/Tag.html" title="interface in org.apache.hadoop.hbase">Tag</a>&gt;&nbsp;tags)</code>&nbsp;</td>
 </tr>
-<tr class="rowColor">
+<tr class="altColor">
 <td class="colFirst"><code><a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a></code></td>
 <td class="colLast"><span class="typeNameLabel">CellCreator.</span><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/CellCreator.html#create-byte:A-int-int-byte:A-int-int-byte:A-int-int-long-byte:A-int-int-java.lang.String-">create</a></span>(byte[]&nbsp;row,
       int&nbsp;roffset,
@@ -4233,18 +4237,18 @@ service.</div>
 </div>
 </td>
 </tr>
-<tr class="altColor">
+<tr class="rowColor">
 <td class="colFirst"><code>static <a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a></code></td>
 <td class="colLast"><span class="typeNameLabel">Import.</span><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/Import.html#filterKv-org.apache.hadoop.hbase.filter.Filter-org.apache.hadoop.hbase.Cell-">filterKv</a></span>(<a href="../../../../../org/apache/hadoop/hbase/filter/Filter.html" title="class in org.apache.hadoop.hbase.filter">Filter</a>&nbsp;filter,
         <a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a>&nbsp;c)</code>
 <div class="block">Attempt to filter out the keyvalue</div>
 </td>
 </tr>
-<tr class="rowColor">
+<tr class="altColor">
 <td class="colFirst"><code><a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a></code></td>
 <td class="colLast"><span class="typeNameLabel">HFileInputFormat.HFileRecordReader.</span><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HFileInputFormat.HFileRecordReader.html#getCurrentValue--">getCurrentValue</a></span>()</code>&nbsp;</td>
 </tr>
-<tr class="altColor">
+<tr class="rowColor">
 <td class="colFirst"><code><a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a></code></td>
 <td class="colLast"><span class="typeNameLabel">SyncTable.SyncMapper.CellScanner.</span><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html#nextCellInRow--">nextCellInRow</a></span>()</code>
 <div class="block">Returns the next Cell in the current row or null iff none remain.</div>
@@ -4283,40 +4287,44 @@ service.</div>
           <a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a>&nbsp;kv)</code>&nbsp;</td>
 </tr>
 <tr class="rowColor">
-<td class="colFirst"><code>private static int</code></td>
+<td class="colFirst"><code>private <a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a></code></td>
+<td class="colLast"><span class="typeNameLabel">SyncTable.SyncMapper.</span><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#checkAndResetTimestamp-org.apache.hadoop.hbase.Cell-">checkAndResetTimestamp</a></span>(<a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a>&nbsp;sourceCell)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>private int</code></td>
 <td class="colLast"><span class="typeNameLabel">SyncTable.SyncMapper.</span><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#compareCellKeysWithinRow-org.apache.hadoop.hbase.Cell-org.apache.hadoop.hbase.Cell-">compareCellKeysWithinRow</a></span>(<a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a>&nbsp;c1,
                         <a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a>&nbsp;c2)</code>
 <div class="block">Compare families, qualifiers, and timestamps of the given Cells.</div>
 </td>
 </tr>
-<tr class="altColor">
+<tr class="rowColor">
 <td class="colFirst"><code>private static <a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a></code></td>
 <td class="colLast"><span class="typeNameLabel">Import.</span><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/Import.html#convertKv-org.apache.hadoop.hbase.Cell-java.util.Map-">convertKv</a></span>(<a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a>&nbsp;kv,
          <a href="https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true" title="class or interface in java.util">Map</a>&lt;byte[],byte[]&gt;&nbsp;cfRenameMap)</code>&nbsp;</td>
 </tr>
-<tr class="rowColor">
+<tr class="altColor">
 <td class="colFirst"><code><a href="../../../../../org/apache/hadoop/hbase/KeyValue.html" title="class in org.apache.hadoop.hbase">KeyValue</a></code></td>
 <td class="colLast"><span class="typeNameLabel">CellSerialization.CellDeserializer.</span><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/CellSerialization.CellDeserializer.html#deserialize-org.apache.hadoop.hbase.Cell-">deserialize</a></span>(<a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a>&nbsp;ignore)</code>&nbsp;</td>
 </tr>
-<tr class="altColor">
+<tr class="rowColor">
 <td class="colFirst"><code>protected boolean</code></td>
 <td class="colLast"><span class="typeNameLabel">WALPlayer.WALMapper.</span><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/WALPlayer.WALMapper.html#filter-org.apache.hadoop.mapreduce.Mapper.Context-org.apache.hadoop.hbase.Cell-">filter</a></span>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context,
       <a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a>&nbsp;cell)</code>&nbsp;</td>
 </tr>
-<tr class="rowColor">
+<tr class="altColor">
 <td class="colFirst"><code>static <a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a></code></td>
 <td class="colLast"><span class="typeNameLabel">Import.</span><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/Import.html#filterKv-org.apache.hadoop.hbase.filter.Filter-org.apache.hadoop.hbase.Cell-">filterKv</a></span>(<a href="../../../../../org/apache/hadoop/hbase/filter/Filter.html" title="class in org.apache.hadoop.hbase.filter">Filter</a>&nbsp;filter,
         <a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a>&nbsp;c)</code>
 <div class="block">Attempt to filter out the keyvalue</div>
 </td>
 </tr>
-<tr class="altColor">
+<tr class="rowColor">
 <td class="colFirst"><code>int</code></td>
 <td class="colLast"><span class="typeNameLabel">Import.CellWritableComparablePartitioner.</span><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/Import.CellWritableComparablePartitioner.html#getPartition-org.apache.hadoop.hbase.mapreduce.Import.CellWritableComparable-org.apache.hadoop.hbase.Cell-int-">getPartition</a></span>(<a href="../../../../../org/apache/hadoop/hbase/mapreduce/Import.CellWritableComparable.html" title="class in org.apache. [...]
             <a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a>&nbsp;value,
             int&nbsp;numPartitions)</code>&nbsp;</td>
 </tr>
-<tr class="rowColor">
+<tr class="altColor">
 <td class="colFirst"><code>void</code></td>
 <td class="colLast"><span class="typeNameLabel">CellSerialization.CellSerializer.</span><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/CellSerialization.CellSerializer.html#serialize-org.apache.hadoop.hbase.Cell-">serialize</a></span>(<a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a>&nbsp;kv)</code>&nbsp;</td>
 </tr>
diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
index bc91dcb..2331da9 100644
--- a/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
+++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
@@ -11436,7 +11436,7 @@ service.</div>
 <td class="colLast"><span class="typeNameLabel">BulkLoadHFilesTool.</span><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#throwAndLogTableNotFoundException-org.apache.hadoop.hbase.TableName-">throwAndLogTableNotFoundException</a></span>(<a href="../../../../../org/apache/hadoop/hbase/TableName.html" title="class in org.apache.hadoop.hbase">TableName</a>&nbsp;tn)</code>&nbsp;</td>
 </tr>
 <tr class="rowColor">
-<td class="colFirst"><code>private <a href="https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true" title="class or interface in java.util.concurrent">CompletableFuture</a>&lt;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true" title="class or interface in java.util">Collection</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache. [...]
+<td class="colFirst"><code>protected <a href="https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true" title="class or interface in java.util.concurrent">CompletableFuture</a>&lt;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true" title="class or interface in java.util">Collection</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apach [...]
 <td class="colLast"><span class="typeNameLabel">BulkLoadHFilesTool.</span><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#tryAtomicRegionLoad-org.apache.hadoop.hbase.client.AsyncClusterConnection-org.apache.hadoop.hbase.TableName-boolean-byte:A-java.util.Collection-">tryAtomicRegionLoad</a></span>(<a href="../../../../../org/apache/hadoop/hbase/client/AsyncClusterConnection.html" title="interface in org.apache.hadoop.hbase.c [...]
                    <a href="../../../../../org/apache/hadoop/hbase/TableName.html" title="class in org.apache.hadoop.hbase">TableName</a>&nbsp;tableName,
                    boolean&nbsp;copyFiles,
diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncClusterConnection.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncClusterConnection.html
index 64c9812..ef4b4bf 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncClusterConnection.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncClusterConnection.html
@@ -505,7 +505,7 @@
            <a href="../../../../../../org/apache/hadoop/hbase/TableName.html" title="class in org.apache.hadoop.hbase">TableName</a>&nbsp;tableName)</code>&nbsp;</td>
 </tr>
 <tr class="rowColor">
-<td class="colFirst"><code>private <a href="https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true" title="class or interface in java.util.concurrent">CompletableFuture</a>&lt;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true" title="class or interface in java.util">Collection</a>&lt;<a href="../../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apac [...]
+<td class="colFirst"><code>protected <a href="https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true" title="class or interface in java.util.concurrent">CompletableFuture</a>&lt;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true" title="class or interface in java.util">Collection</a>&lt;<a href="../../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.ap [...]
 <td class="colLast"><span class="typeNameLabel">BulkLoadHFilesTool.</span><code><span class="memberNameLink"><a href="../../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#tryAtomicRegionLoad-org.apache.hadoop.hbase.client.AsyncClusterConnection-org.apache.hadoop.hbase.TableName-boolean-byte:A-java.util.Collection-">tryAtomicRegionLoad</a></span>(<a href="../../../../../../org/apache/hadoop/hbase/client/AsyncClusterConnection.html" title="interface in org.apache.hadoop.h [...]
                    <a href="../../../../../../org/apache/hadoop/hbase/TableName.html" title="class in org.apache.hadoop.hbase">TableName</a>&nbsp;tableName,
                    boolean&nbsp;copyFiles,
diff --git a/devapidocs/org/apache/hadoop/hbase/client/package-tree.html b/devapidocs/org/apache/hadoop/hbase/client/package-tree.html
index 6dea48e..2ea0d82 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/package-tree.html
@@ -433,22 +433,22 @@
 <ul>
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
+<li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/Scan.ReadType.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">Scan.ReadType</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/AbstractResponse.ResponseType.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">AbstractResponse.ResponseType</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/LogQueryFilter.Type.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">LogQueryFilter.Type</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/Durability.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">Durability</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/RequestController.ReturnCode.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">RequestController.ReturnCode</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/AsyncScanSingleRegionRpcRetryingCaller.ScanResumerState.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">AsyncScanSingleRegionRpcRetryingCaller.ScanResumerState</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/IsolationLevel.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">IsolationLevel</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/SnapshotType.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">SnapshotType</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/AsyncScanSingleRegionRpcRetryingCaller.ScanControllerState.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">AsyncScanSingleRegionRpcRetryingCaller.ScanControllerState</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/TableState.State.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">TableState.State</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/MobCompactPartitionPolicy.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">MobCompactPartitionPolicy</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/RequestController.ReturnCode.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">RequestController.ReturnCode</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/SnapshotType.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">SnapshotType</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/CompactionState.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">CompactionState</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/MasterSwitchType.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">MasterSwitchType</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/AsyncScanSingleRegionRpcRetryingCaller.ScanResumerState.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">AsyncScanSingleRegionRpcRetryingCaller.ScanResumerState</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/IsolationLevel.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">IsolationLevel</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/Durability.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">Durability</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/Consistency.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">Consistency</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/RegionLocateType.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">RegionLocateType</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/MasterSwitchType.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">MasterSwitchType</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/CompactType.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">CompactType</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/Consistency.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">Consistency</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/Scan.ReadType.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">Scan.ReadType</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.client.<a href="../../../../../org/apache/hadoop/hbase/client/AbstractResponse.ResponseType.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">AbstractResponse.ResponseType</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/coprocessor/package-tree.html b/devapidocs/org/apache/hadoop/hbase/coprocessor/package-tree.html
index db38cd2..46638ef 100644
--- a/devapidocs/org/apache/hadoop/hbase/coprocessor/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/coprocessor/package-tree.html
@@ -197,8 +197,8 @@
 <ul>
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
-<li type="circle">org.apache.hadoop.hbase.coprocessor.<a href="../../../../../org/apache/hadoop/hbase/coprocessor/MetaTableMetrics.MetaTableOps.html" title="enum in org.apache.hadoop.hbase.coprocessor"><span class="typeNameLink">MetaTableMetrics.MetaTableOps</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.coprocessor.<a href="../../../../../org/apache/hadoop/hbase/coprocessor/RegionObserver.MutationType.html" title="enum in org.apache.hadoop.hbase.coprocessor"><span class="typeNameLink">RegionObserver.MutationType</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.coprocessor.<a href="../../../../../org/apache/hadoop/hbase/coprocessor/MetaTableMetrics.MetaTableOps.html" title="enum in org.apache.hadoop.hbase.coprocessor"><span class="typeNameLink">MetaTableMetrics.MetaTableOps</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/executor/package-tree.html b/devapidocs/org/apache/hadoop/hbase/executor/package-tree.html
index 0192935..880a54a 100644
--- a/devapidocs/org/apache/hadoop/hbase/executor/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/executor/package-tree.html
@@ -104,8 +104,8 @@
 <ul>
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
-<li type="circle">org.apache.hadoop.hbase.executor.<a href="../../../../../org/apache/hadoop/hbase/executor/ExecutorType.html" title="enum in org.apache.hadoop.hbase.executor"><span class="typeNameLink">ExecutorType</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.executor.<a href="../../../../../org/apache/hadoop/hbase/executor/EventType.html" title="enum in org.apache.hadoop.hbase.executor"><span class="typeNameLink">EventType</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.executor.<a href="../../../../../org/apache/hadoop/hbase/executor/ExecutorType.html" title="enum in org.apache.hadoop.hbase.executor"><span class="typeNameLink">ExecutorType</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/filter/package-tree.html b/devapidocs/org/apache/hadoop/hbase/filter/package-tree.html
index dd0f5bc..dae05c2 100644
--- a/devapidocs/org/apache/hadoop/hbase/filter/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/filter/package-tree.html
@@ -190,12 +190,12 @@
 <ul>
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
+<li type="circle">org.apache.hadoop.hbase.filter.<a href="../../../../../org/apache/hadoop/hbase/filter/FilterWrapper.FilterRowRetCode.html" title="enum in org.apache.hadoop.hbase.filter"><span class="typeNameLink">FilterWrapper.FilterRowRetCode</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.filter.<a href="../../../../../org/apache/hadoop/hbase/filter/Filter.ReturnCode.html" title="enum in org.apache.hadoop.hbase.filter"><span class="typeNameLink">Filter.ReturnCode</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.filter.<a href="../../../../../org/apache/hadoop/hbase/filter/BitComparator.BitwiseOp.html" title="enum in org.apache.hadoop.hbase.filter"><span class="typeNameLink">BitComparator.BitwiseOp</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.filter.<a href="../../../../../org/apache/hadoop/hbase/filter/FuzzyRowFilter.Order.html" title="enum in org.apache.hadoop.hbase.filter"><span class="typeNameLink">FuzzyRowFilter.Order</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.filter.<a href="../../../../../org/apache/hadoop/hbase/filter/FilterList.Operator.html" title="enum in org.apache.hadoop.hbase.filter"><span class="typeNameLink">FilterList.Operator</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.filter.<a href="../../../../../org/apache/hadoop/hbase/filter/FilterWrapper.FilterRowRetCode.html" title="enum in org.apache.hadoop.hbase.filter"><span class="typeNameLink">FilterWrapper.FilterRowRetCode</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.filter.<a href="../../../../../org/apache/hadoop/hbase/filter/FuzzyRowFilter.SatisfiesCode.html" title="enum in org.apache.hadoop.hbase.filter"><span class="typeNameLink">FuzzyRowFilter.SatisfiesCode</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.filter.<a href="../../../../../org/apache/hadoop/hbase/filter/BitComparator.BitwiseOp.html" title="enum in org.apache.hadoop.hbase.filter"><span class="typeNameLink">BitComparator.BitwiseOp</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.filter.<a href="../../../../../org/apache/hadoop/hbase/filter/RegexStringComparator.EngineType.html" title="enum in org.apache.hadoop.hbase.filter"><span class="typeNameLink">RegexStringComparator.EngineType</span></a></li>
 </ul>
 </li>
diff --git a/devapidocs/org/apache/hadoop/hbase/hbtop/field/package-tree.html b/devapidocs/org/apache/hadoop/hbase/hbtop/field/package-tree.html
index cd793c2..d9bf09d 100644
--- a/devapidocs/org/apache/hadoop/hbase/hbtop/field/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/hbtop/field/package-tree.html
@@ -92,8 +92,8 @@
 <ul>
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
-<li type="circle">org.apache.hadoop.hbase.hbtop.field.<a href="../../../../../../org/apache/hadoop/hbase/hbtop/field/Field.html" title="enum in org.apache.hadoop.hbase.hbtop.field"><span class="typeNameLink">Field</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.hbtop.field.<a href="../../../../../../org/apache/hadoop/hbase/hbtop/field/FieldValueType.html" title="enum in org.apache.hadoop.hbase.hbtop.field"><span class="typeNameLink">FieldValueType</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.hbtop.field.<a href="../../../../../../org/apache/hadoop/hbase/hbtop/field/Field.html" title="enum in org.apache.hadoop.hbase.hbtop.field"><span class="typeNameLink">Field</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/hbtop/terminal/package-tree.html b/devapidocs/org/apache/hadoop/hbase/hbtop/terminal/package-tree.html
index b8f3866..007e90f 100644
--- a/devapidocs/org/apache/hadoop/hbase/hbtop/terminal/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/hbtop/terminal/package-tree.html
@@ -107,8 +107,8 @@
 <ul>
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
-<li type="circle">org.apache.hadoop.hbase.hbtop.terminal.<a href="../../../../../../org/apache/hadoop/hbase/hbtop/terminal/Color.html" title="enum in org.apache.hadoop.hbase.hbtop.terminal"><span class="typeNameLink">Color</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.hbtop.terminal.<a href="../../../../../../org/apache/hadoop/hbase/hbtop/terminal/KeyPress.Type.html" title="enum in org.apache.hadoop.hbase.hbtop.terminal"><span class="typeNameLink">KeyPress.Type</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.hbtop.terminal.<a href="../../../../../../org/apache/hadoop/hbase/hbtop/terminal/Color.html" title="enum in org.apache.hadoop.hbase.hbtop.terminal"><span class="typeNameLink">Color</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/http/package-tree.html b/devapidocs/org/apache/hadoop/hbase/http/package-tree.html
index 0e4be64..1c8e128 100644
--- a/devapidocs/org/apache/hadoop/hbase/http/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/http/package-tree.html
@@ -145,8 +145,8 @@
 <ul>
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
-<li type="circle">org.apache.hadoop.hbase.http.<a href="../../../../../org/apache/hadoop/hbase/http/ProfileServlet.Output.html" title="enum in org.apache.hadoop.hbase.http"><span class="typeNameLink">ProfileServlet.Output</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.http.<a href="../../../../../org/apache/hadoop/hbase/http/HttpConfig.Policy.html" title="enum in org.apache.hadoop.hbase.http"><span class="typeNameLink">HttpConfig.Policy</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.http.<a href="../../../../../org/apache/hadoop/hbase/http/ProfileServlet.Output.html" title="enum in org.apache.hadoop.hbase.http"><span class="typeNameLink">ProfileServlet.Output</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.http.<a href="../../../../../org/apache/hadoop/hbase/http/ProfileServlet.Event.html" title="enum in org.apache.hadoop.hbase.http"><span class="typeNameLink">ProfileServlet.Event</span></a></li>
 </ul>
 </li>
diff --git a/devapidocs/org/apache/hadoop/hbase/io/hfile/package-tree.html b/devapidocs/org/apache/hadoop/hbase/io/hfile/package-tree.html
index f79a766..648440f 100644
--- a/devapidocs/org/apache/hadoop/hbase/io/hfile/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/io/hfile/package-tree.html
@@ -305,12 +305,12 @@
 <ul>
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
-<li type="circle">org.apache.hadoop.hbase.io.hfile.<a href="../../../../../../org/apache/hadoop/hbase/io/hfile/HFileBlock.Writer.State.html" title="enum in org.apache.hadoop.hbase.io.hfile"><span class="typeNameLink">HFileBlock.Writer.State</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.io.hfile.<a href="../../../../../../org/apache/hadoop/hbase/io/hfile/ReaderContext.ReaderType.html" title="enum in org.apache.hadoop.hbase.io.hfile"><span class="typeNameLink">ReaderContext.ReaderType</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.io.hfile.<a href="../../../../../../org/apache/hadoop/hbase/io/hfile/BlockPriority.html" title="enum in org.apache.hadoop.hbase.io.hfile"><span class="typeNameLink">BlockPriority</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.io.hfile.<a href="../../../../../../org/apache/hadoop/hbase/io/hfile/BlockCacheFactory.ExternalBlockCaches.html" title="enum in org.apache.hadoop.hbase.io.hfile"><span class="typeNameLink">BlockCacheFactory.ExternalBlockCaches</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.io.hfile.<a href="../../../../../../org/apache/hadoop/hbase/io/hfile/BlockType.BlockCategory.html" title="enum in org.apache.hadoop.hbase.io.hfile"><span class="typeNameLink">BlockType.BlockCategory</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.io.hfile.<a href="../../../../../../org/apache/hadoop/hbase/io/hfile/BlockPriority.html" title="enum in org.apache.hadoop.hbase.io.hfile"><span class="typeNameLink">BlockPriority</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.io.hfile.<a href="../../../../../../org/apache/hadoop/hbase/io/hfile/BlockType.html" title="enum in org.apache.hadoop.hbase.io.hfile"><span class="typeNameLink">BlockType</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.io.hfile.<a href="../../../../../../org/apache/hadoop/hbase/io/hfile/HFileBlock.Writer.State.html" title="enum in org.apache.hadoop.hbase.io.hfile"><span class="typeNameLink">HFileBlock.Writer.State</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/package-tree.html b/devapidocs/org/apache/hadoop/hbase/ipc/package-tree.html
index f2ea339..c0294f9 100644
--- a/devapidocs/org/apache/hadoop/hbase/ipc/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/ipc/package-tree.html
@@ -360,8 +360,8 @@
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
 <li type="circle">org.apache.hadoop.hbase.ipc.<a href="../../../../../org/apache/hadoop/hbase/ipc/BufferCallBeforeInitHandler.BufferCallAction.html" title="enum in org.apache.hadoop.hbase.ipc"><span class="typeNameLink">BufferCallBeforeInitHandler.BufferCallAction</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.ipc.<a href="../../../../../org/apache/hadoop/hbase/ipc/MetricsHBaseServerSourceFactoryImpl.SourceStorage.html" title="enum in org.apache.hadoop.hbase.ipc"><span class="typeNameLink">MetricsHBaseServerSourceFactoryImpl.SourceStorage</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.ipc.<a href="../../../../../org/apache/hadoop/hbase/ipc/CallEvent.Type.html" title="enum in org.apache.hadoop.hbase.ipc"><span class="typeNameLink">CallEvent.Type</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.ipc.<a href="../../../../../org/apache/hadoop/hbase/ipc/MetricsHBaseServerSourceFactoryImpl.SourceStorage.html" title="enum in org.apache.hadoop.hbase.ipc"><span class="typeNameLink">MetricsHBaseServerSourceFactoryImpl.SourceStorage</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/mapreduce/HashTable.HashMapper.html b/devapidocs/org/apache/hadoop/hbase/mapreduce/HashTable.HashMapper.html
index 8739e16..35d4c1d 100644
--- a/devapidocs/org/apache/hadoop/hbase/mapreduce/HashTable.HashMapper.html
+++ b/devapidocs/org/apache/hadoop/hbase/mapreduce/HashTable.HashMapper.html
@@ -123,7 +123,7 @@ var activeTableTab = "activeTableTab";
 </dl>
 <hr>
 <br>
-<pre>public static class <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.542">HashTable.HashMapper</a>
+<pre>public static class <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.549">HashTable.HashMapper</a>
 extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.html" title="class in org.apache.hadoop.hbase.mapreduce">TableMapper</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>,<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&gt;</pre>
 </li>
 </ul>
@@ -251,7 +251,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>hasher</h4>
-<pre>private&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html" title="class in org.apache.hadoop.hbase.mapreduce">HashTable.ResultHasher</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.HashMapper.html#line.545">hasher</a></pre>
+<pre>private&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html" title="class in org.apache.hadoop.hbase.mapreduce">HashTable.ResultHasher</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.HashMapper.html#line.552">hasher</a></pre>
 </li>
 </ul>
 <a name="targetBatchSize">
@@ -260,7 +260,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>targetBatchSize</h4>
-<pre>private&nbsp;long <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.HashMapper.html#line.546">targetBatchSize</a></pre>
+<pre>private&nbsp;long <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.HashMapper.html#line.553">targetBatchSize</a></pre>
 </li>
 </ul>
 <a name="currentRow">
@@ -269,7 +269,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockListLast">
 <li class="blockList">
 <h4>currentRow</h4>
-<pre>private&nbsp;<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.HashMapper.html#line.548">currentRow</a></pre>
+<pre>private&nbsp;<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.HashMapper.html#line.555">currentRow</a></pre>
 </li>
 </ul>
 </li>
@@ -286,7 +286,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockListLast">
 <li class="blockList">
 <h4>HashMapper</h4>
-<pre>public&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.HashMapper.html#line.542">HashMapper</a>()</pre>
+<pre>public&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.HashMapper.html#line.549">HashMapper</a>()</pre>
 </li>
 </ul>
 </li>
@@ -303,7 +303,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>setup</h4>
-<pre>protected&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.HashMapper.html#line.551">setup</a>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context)
+<pre>protected&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.HashMapper.html#line.558">setup</a>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context)
               throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a>,
                      <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true" title="class or interface in java.lang">InterruptedException</a></pre>
 <dl>
@@ -321,7 +321,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>map</h4>
-<pre>protected&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.HashMapper.html#line.561">map</a>(<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&nbsp;key,
+<pre>protected&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.HashMapper.html#line.569">map</a>(<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&nbsp;key,
                    <a href="../../../../../org/apache/hadoop/hbase/client/Result.html" title="class in org.apache.hadoop.hbase.client">Result</a>&nbsp;value,
                    org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context)
             throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a>,
@@ -341,7 +341,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockListLast">
 <li class="blockList">
 <h4>cleanup</h4>
-<pre>protected&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.HashMapper.html#line.578">cleanup</a>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context)
+<pre>protected&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.HashMapper.html#line.586">cleanup</a>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context)
                 throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a>,
                        <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true" title="class or interface in java.lang">InterruptedException</a></pre>
 <dl>
diff --git a/devapidocs/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html b/devapidocs/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html
index c62e350..4b3d2ec 100644
--- a/devapidocs/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html
+++ b/devapidocs/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html
@@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab";
 </dl>
 <hr>
 <br>
-<pre>static class <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.467">HashTable.ResultHasher</a>
+<pre>static class <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.470">HashTable.ResultHasher</a>
 extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true" title="class or interface in java.lang">Object</a></pre>
 </li>
 </ul>
@@ -153,6 +153,10 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <td class="colFirst"><code>private <a href="https://docs.oracle.com/javase/8/docs/api/java/security/MessageDigest.html?is-external=true" title="class or interface in java.security">MessageDigest</a></code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#digest">digest</a></span></code>&nbsp;</td>
 </tr>
+<tr class="rowColor">
+<td class="colFirst"><code>(package private) boolean</code></td>
+<td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#ignoreTimestamps">ignoreTimestamps</a></span></code>&nbsp;</td>
+</tr>
 </table>
 </li>
 </ul>
@@ -241,7 +245,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>digest</h4>
-<pre>private&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/security/MessageDigest.html?is-external=true" title="class or interface in java.security">MessageDigest</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.468">digest</a></pre>
+<pre>private&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/security/MessageDigest.html?is-external=true" title="class or interface in java.security">MessageDigest</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.471">digest</a></pre>
 </li>
 </ul>
 <a name="batchStarted">
@@ -250,7 +254,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>batchStarted</h4>
-<pre>private&nbsp;boolean <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.470">batchStarted</a></pre>
+<pre>private&nbsp;boolean <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.473">batchStarted</a></pre>
 </li>
 </ul>
 <a name="batchStartKey">
@@ -259,7 +263,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>batchStartKey</h4>
-<pre>private&nbsp;<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.471">batchStartKey</a></pre>
+<pre>private&nbsp;<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.474">batchStartKey</a></pre>
 </li>
 </ul>
 <a name="batchHash">
@@ -268,16 +272,25 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>batchHash</h4>
-<pre>private&nbsp;<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.472">batchHash</a></pre>
+<pre>private&nbsp;<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.475">batchHash</a></pre>
 </li>
 </ul>
 <a name="batchSize">
 <!--   -->
 </a>
-<ul class="blockListLast">
+<ul class="blockList">
 <li class="blockList">
 <h4>batchSize</h4>
-<pre>private&nbsp;long <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.473">batchSize</a></pre>
+<pre>private&nbsp;long <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.476">batchSize</a></pre>
+</li>
+</ul>
+<a name="ignoreTimestamps">
+<!--   -->
+</a>
+<ul class="blockListLast">
+<li class="blockList">
+<h4>ignoreTimestamps</h4>
+<pre>boolean <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.477">ignoreTimestamps</a></pre>
 </li>
 </ul>
 </li>
@@ -294,7 +307,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockListLast">
 <li class="blockList">
 <h4>ResultHasher</h4>
-<pre>public&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.476">ResultHasher</a>()</pre>
+<pre>public&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.480">ResultHasher</a>()</pre>
 </li>
 </ul>
 </li>
@@ -311,7 +324,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>startBatch</h4>
-<pre>public&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.484">startBatch</a>(<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&nbsp;row)</pre>
+<pre>public&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.488">startBatch</a>(<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&nbsp;row)</pre>
 </li>
 </ul>
 <a name="hashResult-org.apache.hadoop.hbase.client.Result-">
@@ -320,7 +333,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>hashResult</h4>
-<pre>public&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.494">hashResult</a>(<a href="../../../../../org/apache/hadoop/hbase/client/Result.html" title="class in org.apache.hadoop.hbase.client">Result</a>&nbsp;result)</pre>
+<pre>public&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.498">hashResult</a>(<a href="../../../../../org/apache/hadoop/hbase/client/Result.html" title="class in org.apache.hadoop.hbase.client">Result</a>&nbsp;result)</pre>
 </li>
 </ul>
 <a name="finishBatch--">
@@ -329,7 +342,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>finishBatch</h4>
-<pre>public&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.517">finishBatch</a>()</pre>
+<pre>public&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.524">finishBatch</a>()</pre>
 </li>
 </ul>
 <a name="isBatchStarted--">
@@ -338,7 +351,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>isBatchStarted</h4>
-<pre>public&nbsp;boolean&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.525">isBatchStarted</a>()</pre>
+<pre>public&nbsp;boolean&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.532">isBatchStarted</a>()</pre>
 </li>
 </ul>
 <a name="getBatchStartKey--">
@@ -347,7 +360,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>getBatchStartKey</h4>
-<pre>public&nbsp;<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.529">getBatchStartKey</a>()</pre>
+<pre>public&nbsp;<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.536">getBatchStartKey</a>()</pre>
 </li>
 </ul>
 <a name="getBatchHash--">
@@ -356,7 +369,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>getBatchHash</h4>
-<pre>public&nbsp;<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.533">getBatchHash</a>()</pre>
+<pre>public&nbsp;<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.540">getBatchHash</a>()</pre>
 </li>
 </ul>
 <a name="getBatchSize--">
@@ -365,7 +378,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockListLast">
 <li class="blockList">
 <h4>getBatchSize</h4>
-<pre>public&nbsp;long&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.537">getBatchSize</a>()</pre>
+<pre>public&nbsp;long&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html#line.544">getBatchSize</a>()</pre>
 </li>
 </ul>
 </li>
diff --git a/devapidocs/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html b/devapidocs/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html
index a93fb85..9bfdc9a 100644
--- a/devapidocs/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html
+++ b/devapidocs/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html
@@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab";
 </dl>
 <hr>
 <br>
-<pre>public class <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.327">HashTable.TableHash.Reader</a>
+<pre>public class <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.329">HashTable.TableHash.Reader</a>
 extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true" title="class or interface in java.lang">Object</a>
 implements <a href="https://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html?is-external=true" title="class or interface in java.io">Closeable</a></pre>
 </li>
@@ -249,7 +249,7 @@ implements <a href="https://docs.oracle.com/javase/8/docs/api/java/io/Closeable.
 <ul class="blockList">
 <li class="blockList">
 <h4>conf</h4>
-<pre>private final&nbsp;org.apache.hadoop.conf.Configuration <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html#line.328">conf</a></pre>
+<pre>private final&nbsp;org.apache.hadoop.conf.Configuration <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html#line.330">conf</a></pre>
 </li>
 </ul>
 <a name="hashFileIndex">
@@ -258,7 +258,7 @@ implements <a href="https://docs.oracle.com/javase/8/docs/api/java/io/Closeable.
 <ul class="blockList">
 <li class="blockList">
 <h4>hashFileIndex</h4>
-<pre>private&nbsp;int <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html#line.330">hashFileIndex</a></pre>
+<pre>private&nbsp;int <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html#line.332">hashFileIndex</a></pre>
 </li>
 </ul>
 <a name="mapFileReader">
@@ -267,7 +267,7 @@ implements <a href="https://docs.oracle.com/javase/8/docs/api/java/io/Closeable.
 <ul class="blockList">
 <li class="blockList">
 <h4>mapFileReader</h4>
-<pre>private&nbsp;org.apache.hadoop.io.MapFile.Reader <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html#line.331">mapFileReader</a></pre>
+<pre>private&nbsp;org.apache.hadoop.io.MapFile.Reader <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html#line.333">mapFileReader</a></pre>
 </li>
 </ul>
 <a name="cachedNext">
@@ -276,7 +276,7 @@ implements <a href="https://docs.oracle.com/javase/8/docs/api/java/io/Closeable.
 <ul class="blockList">
 <li class="blockList">
 <h4>cachedNext</h4>
-<pre>private&nbsp;boolean <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html#line.333">cachedNext</a></pre>
+<pre>private&nbsp;boolean <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html#line.335">cachedNext</a></pre>
 </li>
 </ul>
 <a name="key">
@@ -285,7 +285,7 @@ implements <a href="https://docs.oracle.com/javase/8/docs/api/java/io/Closeable.
 <ul class="blockList">
 <li class="blockList">
 <h4>key</h4>
-<pre>private&nbsp;<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html#line.334">key</a></pre>
+<pre>private&nbsp;<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html#line.336">key</a></pre>
 </li>
 </ul>
 <a name="hash">
@@ -294,7 +294,7 @@ implements <a href="https://docs.oracle.com/javase/8/docs/api/java/io/Closeable.
 <ul class="blockListLast">
 <li class="blockList">
 <h4>hash</h4>
-<pre>private&nbsp;<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html#line.335">hash</a></pre>
+<pre>private&nbsp;<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html#line.337">hash</a></pre>
 </li>
 </ul>
 </li>
@@ -311,7 +311,7 @@ implements <a href="https://docs.oracle.com/javase/8/docs/api/java/io/Closeable.
 <ul class="blockListLast">
 <li class="blockList">
 <h4>Reader</h4>
-<pre><a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html#line.337">Reader</a>(org.apache.hadoop.conf.Configuration&nbsp;conf,
+<pre><a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html#line.339">Reader</a>(org.apache.hadoop.conf.Configuration&nbsp;conf,
        <a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&nbsp;startKey)
 throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
 <dl>
@@ -334,7 +334,7 @@ throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>next</h4>
-<pre>public&nbsp;boolean&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html#line.365">next</a>()
+<pre>public&nbsp;boolean&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html#line.367">next</a>()
              throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
 <div class="block">Read the next key/hash pair.
  Returns true if such a pair exists and false when at the end of the data.</div>
@@ -350,7 +350,7 @@ throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>getCurrentKey</h4>
-<pre>public&nbsp;<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html#line.393">getCurrentKey</a>()</pre>
+<pre>public&nbsp;<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html#line.395">getCurrentKey</a>()</pre>
 <div class="block">Get the current key</div>
 <dl>
 <dt><span class="returnLabel">Returns:</span></dt>
@@ -364,7 +364,7 @@ throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>getCurrentHash</h4>
-<pre>public&nbsp;<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html#line.401">getCurrentHash</a>()</pre>
+<pre>public&nbsp;<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html#line.403">getCurrentHash</a>()</pre>
 <div class="block">Get the current hash</div>
 <dl>
 <dt><span class="returnLabel">Returns:</span></dt>
@@ -378,7 +378,7 @@ throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>openHashFile</h4>
-<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html#line.405">openHashFile</a>()
+<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html#line.407">openHashFile</a>()
                    throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
 <dl>
 <dt><span class="throwsLabel">Throws:</span></dt>
@@ -392,7 +392,7 @@ throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.ht
 <ul class="blockListLast">
 <li class="blockList">
 <h4>close</h4>
-<pre>public&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html#line.415">close</a>()
+<pre>public&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html#line.417">close</a>()
            throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
 <dl>
 <dt><span class="overrideSpecifyLabel">Specified by:</span></dt>
diff --git a/devapidocs/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html b/devapidocs/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html
index 8865339..292f6be 100644
--- a/devapidocs/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html
+++ b/devapidocs/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html
@@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab";
 </dl>
 <hr>
 <br>
-<pre>public static class <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.85">HashTable.TableHash</a>
+<pre>public static class <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.86">HashTable.TableHash</a>
 extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true" title="class or interface in java.lang">Object</a></pre>
 </li>
 </ul>
@@ -169,34 +169,38 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#hashDir">hashDir</a></span></code>&nbsp;</td>
 </tr>
 <tr class="altColor">
+<td class="colFirst"><code>(package private) boolean</code></td>
+<td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#ignoreTimestamps">ignoreTimestamps</a></span></code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
 <td class="colFirst"><code>(package private) int</code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#numHashFiles">numHashFiles</a></span></code>&nbsp;</td>
 </tr>
-<tr class="rowColor">
+<tr class="altColor">
 <td class="colFirst"><code>(package private) <a href="https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true" title="class or interface in java.util">List</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&gt;</code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#partitions">partitions</a></span></code>&nbsp;</td>
 </tr>
-<tr class="altColor">
+<tr class="rowColor">
 <td class="colFirst"><code>(package private) int</code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#scanBatch">scanBatch</a></span></code>&nbsp;</td>
 </tr>
-<tr class="rowColor">
+<tr class="altColor">
 <td class="colFirst"><code>(package private) byte[]</code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#startRow">startRow</a></span></code>&nbsp;</td>
 </tr>
-<tr class="altColor">
+<tr class="rowColor">
 <td class="colFirst"><code>(package private) long</code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#startTime">startTime</a></span></code>&nbsp;</td>
 </tr>
-<tr class="rowColor">
+<tr class="altColor">
 <td class="colFirst"><code>(package private) byte[]</code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#stopRow">stopRow</a></span></code>&nbsp;</td>
 </tr>
-<tr class="altColor">
+<tr class="rowColor">
 <td class="colFirst"><code>(package private) <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a></code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#tableName">tableName</a></span></code>&nbsp;</td>
 </tr>
-<tr class="rowColor">
+<tr class="altColor">
 <td class="colFirst"><code>(package private) int</code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#versions">versions</a></span></code>&nbsp;</td>
 </tr>
@@ -313,7 +317,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>hashDir</h4>
-<pre>org.apache.hadoop.fs.Path <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.87">hashDir</a></pre>
+<pre>org.apache.hadoop.fs.Path <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.88">hashDir</a></pre>
 </li>
 </ul>
 <a name="tableName">
@@ -322,7 +326,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>tableName</h4>
-<pre><a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.89">tableName</a></pre>
+<pre><a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.90">tableName</a></pre>
 </li>
 </ul>
 <a name="families">
@@ -331,7 +335,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>families</h4>
-<pre><a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.90">families</a></pre>
+<pre><a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.91">families</a></pre>
 </li>
 </ul>
 <a name="batchSize">
@@ -340,7 +344,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>batchSize</h4>
-<pre>long <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.91">batchSize</a></pre>
+<pre>long <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.92">batchSize</a></pre>
 </li>
 </ul>
 <a name="numHashFiles">
@@ -349,7 +353,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>numHashFiles</h4>
-<pre>int <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.92">numHashFiles</a></pre>
+<pre>int <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.93">numHashFiles</a></pre>
 </li>
 </ul>
 <a name="startRow">
@@ -358,7 +362,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>startRow</h4>
-<pre>byte[] <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.93">startRow</a></pre>
+<pre>byte[] <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.94">startRow</a></pre>
 </li>
 </ul>
 <a name="stopRow">
@@ -367,7 +371,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>stopRow</h4>
-<pre>byte[] <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.94">stopRow</a></pre>
+<pre>byte[] <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.95">stopRow</a></pre>
 </li>
 </ul>
 <a name="scanBatch">
@@ -376,7 +380,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>scanBatch</h4>
-<pre>int <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.95">scanBatch</a></pre>
+<pre>int <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.96">scanBatch</a></pre>
 </li>
 </ul>
 <a name="versions">
@@ -385,7 +389,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>versions</h4>
-<pre>int <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.96">versions</a></pre>
+<pre>int <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.97">versions</a></pre>
 </li>
 </ul>
 <a name="startTime">
@@ -394,7 +398,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>startTime</h4>
-<pre>long <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.97">startTime</a></pre>
+<pre>long <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.98">startTime</a></pre>
 </li>
 </ul>
 <a name="endTime">
@@ -403,7 +407,16 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>endTime</h4>
-<pre>long <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.98">endTime</a></pre>
+<pre>long <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.99">endTime</a></pre>
+</li>
+</ul>
+<a name="ignoreTimestamps">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>ignoreTimestamps</h4>
+<pre>boolean <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.100">ignoreTimestamps</a></pre>
 </li>
 </ul>
 <a name="partitions">
@@ -412,7 +425,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockListLast">
 <li class="blockList">
 <h4>partitions</h4>
-<pre><a href="https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true" title="class or interface in java.util">List</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&gt; <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.100">partitions</a></pre>
+<pre><a href="https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true" title="class or interface in java.util">List</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&gt; <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.102">partitions</a></pre>
 </li>
 </ul>
 </li>
@@ -429,7 +442,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockListLast">
 <li class="blockList">
 <h4>TableHash</h4>
-<pre>public&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.85">TableHash</a>()</pre>
+<pre>public&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.86">TableHash</a>()</pre>
 </li>
 </ul>
 </li>
@@ -446,7 +459,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>read</h4>
-<pre>public static&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html" title="class in org.apache.hadoop.hbase.mapreduce">HashTable.TableHash</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.102">read</a>(org.apache.hadoop.conf.Configuration&nbsp;conf,
+<pre>public static&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html" title="class in org.apache.hadoop.hbase.mapreduce">HashTable.TableHash</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.104">read</a>(org.apache.hadoop.conf.Configuration&nbsp;conf,
                                        org.apache.hadoop.fs.Path&nbsp;hashDir)
                                 throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
 <dl>
@@ -461,7 +474,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>writePropertiesFile</h4>
-<pre>void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.111">writePropertiesFile</a>(org.apache.hadoop.fs.FileSystem&nbsp;fs,
+<pre>void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.113">writePropertiesFile</a>(org.apache.hadoop.fs.FileSystem&nbsp;fs,
                          org.apache.hadoop.fs.Path&nbsp;path)
                   throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
 <dl>
@@ -476,7 +489,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>readPropertiesFile</h4>
-<pre>void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.143">readPropertiesFile</a>(org.apache.hadoop.fs.FileSystem&nbsp;fs,
+<pre>void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.145">readPropertiesFile</a>(org.apache.hadoop.fs.FileSystem&nbsp;fs,
                         org.apache.hadoop.fs.Path&nbsp;path)
                  throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
 <dl>
@@ -491,7 +504,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>initScan</h4>
-<pre><a href="../../../../../org/apache/hadoop/hbase/client/Scan.html" title="class in org.apache.hadoop.hbase.client">Scan</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.185">initScan</a>()
+<pre><a href="../../../../../org/apache/hadoop/hbase/client/Scan.html" title="class in org.apache.hadoop.hbase.client">Scan</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.187">initScan</a>()
        throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
 <dl>
 <dt><span class="throwsLabel">Throws:</span></dt>
@@ -505,7 +518,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>selectPartitions</h4>
-<pre>void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.216">selectPartitions</a>(<a href="../../../../../org/apache/hadoop/hbase/util/Pair.html" title="class in org.apache.hadoop.hbase.util">Pair</a>&lt;byte[][],byte[][]&gt;&nbsp;regionStartEndKeys)</pre>
+<pre>void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.218">selectPartitions</a>(<a href="../../../../../org/apache/hadoop/hbase/util/Pair.html" title="class in org.apache.hadoop.hbase.util">Pair</a>&lt;byte[][],byte[][]&gt;&nbsp;regionStartEndKeys)</pre>
 <div class="block">Choose partitions between row ranges to hash to a single output file
  Selects region boundaries that fall within the scan range, and groups them
  into the desired number of partitions.</div>
@@ -517,7 +530,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>writePartitionFile</h4>
-<pre>void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.256">writePartitionFile</a>(org.apache.hadoop.conf.Configuration&nbsp;conf,
+<pre>void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.258">writePartitionFile</a>(org.apache.hadoop.conf.Configuration&nbsp;conf,
                         org.apache.hadoop.fs.Path&nbsp;path)
                  throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
 <dl>
@@ -532,7 +545,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>readPartitionFile</h4>
-<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.268">readPartitionFile</a>(org.apache.hadoop.fs.FileSystem&nbsp;fs,
+<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.270">readPartitionFile</a>(org.apache.hadoop.fs.FileSystem&nbsp;fs,
                                org.apache.hadoop.conf.Configuration&nbsp;conf,
                                org.apache.hadoop.fs.Path&nbsp;path)
                         throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
@@ -548,7 +561,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>toString</h4>
-<pre>public&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.285">toString</a>()</pre>
+<pre>public&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.287">toString</a>()</pre>
 <dl>
 <dt><span class="overrideSpecifyLabel">Overrides:</span></dt>
 <dd><code><a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--" title="class or interface in java.lang">toString</a></code>&nbsp;in class&nbsp;<code><a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true" title="class or interface in java.lang">Object</a></code></dd>
@@ -561,7 +574,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>getDataFileName</h4>
-<pre>static&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.314">getDataFileName</a>(int&nbsp;hashFileIndex)</pre>
+<pre>static&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.316">getDataFileName</a>(int&nbsp;hashFileIndex)</pre>
 </li>
 </ul>
 <a name="newReader-org.apache.hadoop.conf.Configuration-org.apache.hadoop.hbase.io.ImmutableBytesWritable-">
@@ -570,7 +583,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockListLast">
 <li class="blockList">
 <h4>newReader</h4>
-<pre>public&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html" title="class in org.apache.hadoop.hbase.mapreduce">HashTable.TableHash.Reader</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.322">newReader</a>(org.apache.hadoop.conf.Configuration&nbsp;conf,
+<pre>public&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html" title="class in org.apache.hadoop.hbase.mapreduce">HashTable.TableHash.Reader</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html#line.324">newReader</a>(org.apache.hadoop.conf.Configuration&nbsp;conf,
                                             <a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&nbsp;startKey)
                                      throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
 <div class="block">Open a TableHash.Reader starting at the first hash at or after the given key.</div>
diff --git a/devapidocs/org/apache/hadoop/hbase/mapreduce/HashTable.html b/devapidocs/org/apache/hadoop/hbase/mapreduce/HashTable.html
index 2fd8483..dc81702 100644
--- a/devapidocs/org/apache/hadoop/hbase/mapreduce/HashTable.html
+++ b/devapidocs/org/apache/hadoop/hbase/mapreduce/HashTable.html
@@ -184,30 +184,34 @@ implements org.apache.hadoop.util.Tool</pre>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.html#HASH_DATA_DIR">HASH_DATA_DIR</a></span></code>&nbsp;</td>
 </tr>
 <tr class="altColor">
+<td class="colFirst"><code>(package private) static <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a></code></td>
+<td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.html#IGNORE_TIMESTAMPS">IGNORE_TIMESTAMPS</a></span></code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
 <td class="colFirst"><code>private static org.slf4j.Logger</code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.html#LOG">LOG</a></span></code>&nbsp;</td>
 </tr>
-<tr class="rowColor">
+<tr class="altColor">
 <td class="colFirst"><code>(package private) static <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a></code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.html#MANIFEST_FILE_NAME">MANIFEST_FILE_NAME</a></span></code>&nbsp;</td>
 </tr>
-<tr class="altColor">
+<tr class="rowColor">
 <td class="colFirst"><code>private static int</code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.html#NUM_ARGS">NUM_ARGS</a></span></code>&nbsp;</td>
 </tr>
-<tr class="rowColor">
+<tr class="altColor">
 <td class="colFirst"><code>(package private) static <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a></code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.html#OUTPUT_DATA_FILE_PREFIX">OUTPUT_DATA_FILE_PREFIX</a></span></code>&nbsp;</td>
 </tr>
-<tr class="altColor">
+<tr class="rowColor">
 <td class="colFirst"><code>(package private) static <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a></code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.html#PARTITIONS_FILE_NAME">PARTITIONS_FILE_NAME</a></span></code>&nbsp;</td>
 </tr>
-<tr class="rowColor">
+<tr class="altColor">
 <td class="colFirst"><code>(package private) <a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html" title="class in org.apache.hadoop.hbase.mapreduce">HashTable.TableHash</a></code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.html#tableHash">tableHash</a></span></code>&nbsp;</td>
 </tr>
-<tr class="altColor">
+<tr class="rowColor">
 <td class="colFirst"><code>private static <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a></code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.html#TMP_MANIFEST_FILE_NAME">TMP_MANIFEST_FILE_NAME</a></span></code>&nbsp;</td>
 </tr>
@@ -408,13 +412,26 @@ implements org.apache.hadoop.util.Tool</pre>
 </dl>
 </li>
 </ul>
+<a name="IGNORE_TIMESTAMPS">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>IGNORE_TIMESTAMPS</h4>
+<pre>static final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.76">IGNORE_TIMESTAMPS</a></pre>
+<dl>
+<dt><span class="seeLabel">See Also:</span></dt>
+<dd><a href="../../../../../constant-values.html#org.apache.hadoop.hbase.mapreduce.HashTable.IGNORE_TIMESTAMPS">Constant Field Values</a></dd>
+</dl>
+</li>
+</ul>
 <a name="TMP_MANIFEST_FILE_NAME">
 <!--   -->
 </a>
 <ul class="blockList">
 <li class="blockList">
 <h4>TMP_MANIFEST_FILE_NAME</h4>
-<pre>private static final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.76">TMP_MANIFEST_FILE_NAME</a></pre>
+<pre>private static final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.77">TMP_MANIFEST_FILE_NAME</a></pre>
 <dl>
 <dt><span class="seeLabel">See Also:</span></dt>
 <dd><a href="../../../../../constant-values.html#org.apache.hadoop.hbase.mapreduce.HashTable.TMP_MANIFEST_FILE_NAME">Constant Field Values</a></dd>
@@ -427,7 +444,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>tableHash</h4>
-<pre><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html" title="class in org.apache.hadoop.hbase.mapreduce">HashTable.TableHash</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.78">tableHash</a></pre>
+<pre><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html" title="class in org.apache.hadoop.hbase.mapreduce">HashTable.TableHash</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.79">tableHash</a></pre>
 </li>
 </ul>
 <a name="destPath">
@@ -436,7 +453,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>destPath</h4>
-<pre>org.apache.hadoop.fs.Path <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.79">destPath</a></pre>
+<pre>org.apache.hadoop.fs.Path <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.80">destPath</a></pre>
 </li>
 </ul>
 <a name="NUM_ARGS">
@@ -445,7 +462,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockListLast">
 <li class="blockList">
 <h4>NUM_ARGS</h4>
-<pre>private static final&nbsp;int <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.597">NUM_ARGS</a></pre>
+<pre>private static final&nbsp;int <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.605">NUM_ARGS</a></pre>
 <dl>
 <dt><span class="seeLabel">See Also:</span></dt>
 <dd><a href="../../../../../constant-values.html#org.apache.hadoop.hbase.mapreduce.HashTable.NUM_ARGS">Constant Field Values</a></dd>
@@ -466,7 +483,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockListLast">
 <li class="blockList">
 <h4>HashTable</h4>
-<pre>public&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.81">HashTable</a>(org.apache.hadoop.conf.Configuration&nbsp;conf)</pre>
+<pre>public&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.82">HashTable</a>(org.apache.hadoop.conf.Configuration&nbsp;conf)</pre>
 </li>
 </ul>
 </li>
@@ -483,7 +500,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>isTableStartRow</h4>
-<pre>static&nbsp;boolean&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.421">isTableStartRow</a>(byte[]&nbsp;row)</pre>
+<pre>static&nbsp;boolean&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.423">isTableStartRow</a>(byte[]&nbsp;row)</pre>
 </li>
 </ul>
 <a name="isTableEndRow-byte:A-">
@@ -492,7 +509,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>isTableEndRow</h4>
-<pre>static&nbsp;boolean&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.425">isTableEndRow</a>(byte[]&nbsp;row)</pre>
+<pre>static&nbsp;boolean&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.427">isTableEndRow</a>(byte[]&nbsp;row)</pre>
 </li>
 </ul>
 <a name="createSubmittableJob-java.lang.String:A-">
@@ -501,7 +518,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>createSubmittableJob</h4>
-<pre>public&nbsp;org.apache.hadoop.mapreduce.Job&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.429">createSubmittableJob</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>[]&nbsp;args)
+<pre>public&nbsp;org.apache.hadoop.mapreduce.Job&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.431">createSubmittableJob</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>[]&nbsp;args)
                                                      throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
 <dl>
 <dt><span class="throwsLabel">Throws:</span></dt>
@@ -515,7 +532,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>generatePartitions</h4>
-<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.455">generatePartitions</a>(org.apache.hadoop.fs.Path&nbsp;partitionsPath)
+<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.458">generatePartitions</a>(org.apache.hadoop.fs.Path&nbsp;partitionsPath)
                          throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
 <dl>
 <dt><span class="throwsLabel">Throws:</span></dt>
@@ -529,7 +546,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>writeTempManifestFile</h4>
-<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.584">writeTempManifestFile</a>()
+<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.592">writeTempManifestFile</a>()
                             throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
 <dl>
 <dt><span class="throwsLabel">Throws:</span></dt>
@@ -543,7 +560,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>completeManifest</h4>
-<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.590">completeManifest</a>()
+<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.598">completeManifest</a>()
                        throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
 <dl>
 <dt><span class="throwsLabel">Throws:</span></dt>
@@ -557,7 +574,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>printUsage</h4>
-<pre>private static&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.598">printUsage</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&nbsp;errorMsg)</pre>
+<pre>private static&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.606">printUsage</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&nbsp;errorMsg)</pre>
 </li>
 </ul>
 <a name="doCommandLine-java.lang.String:A-">
@@ -566,7 +583,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>doCommandLine</h4>
-<pre>private&nbsp;boolean&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.634">doCommandLine</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>[]&nbsp;args)</pre>
+<pre>private&nbsp;boolean&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.645">doCommandLine</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>[]&nbsp;args)</pre>
 </li>
 </ul>
 <a name="main-java.lang.String:A-">
@@ -575,7 +592,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>main</h4>
-<pre>public static&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.726">main</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>[]&nbsp;args)
+<pre>public static&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.744">main</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>[]&nbsp;args)
                  throws <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true" title="class or interface in java.lang">Exception</a></pre>
 <div class="block">Main entry point.</div>
 <dl>
@@ -590,7 +607,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockListLast">
 <li class="blockList">
 <h4>run</h4>
-<pre>public&nbsp;int&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.732">run</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>[]&nbsp;args)
+<pre>public&nbsp;int&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html#line.750">run</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>[]&nbsp;args)
         throws <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true" title="class or interface in java.lang">Exception</a></pre>
 <dl>
 <dt><span class="overrideSpecifyLabel">Specified by:</span></dt>
diff --git a/devapidocs/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html b/devapidocs/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html
index a77585b..6718c9d 100644
--- a/devapidocs/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html
+++ b/devapidocs/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html
@@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab";
 </dl>
 <hr>
 <br>
-<pre>private static class <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.393">SyncTable.SyncMapper.CellScanner</a>
+<pre>private static class <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.401">SyncTable.SyncMapper.CellScanner</a>
 extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true" title="class or interface in java.lang">Object</a></pre>
 </li>
 </ul>
@@ -225,7 +225,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>results</h4>
-<pre>private final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Iterator.html?is-external=true" title="class or interface in java.util">Iterator</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/client/Result.html" title="class in org.apache.hadoop.hbase.client">Result</a>&gt; <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html#line.394">results</a></pre>
+<pre>private final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Iterator.html?is-external=true" title="class or interface in java.util">Iterator</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/client/Result.html" title="class in org.apache.hadoop.hbase.client">Result</a>&gt; <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html#line.402">results</a></pre>
 </li>
 </ul>
 <a name="currentRow">
@@ -234,7 +234,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>currentRow</h4>
-<pre>private&nbsp;byte[] <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html#line.396">currentRow</a></pre>
+<pre>private&nbsp;byte[] <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html#line.404">currentRow</a></pre>
 </li>
 </ul>
 <a name="currentRowResult">
@@ -243,7 +243,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>currentRowResult</h4>
-<pre>private&nbsp;<a href="../../../../../org/apache/hadoop/hbase/client/Result.html" title="class in org.apache.hadoop.hbase.client">Result</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html#line.397">currentRowResult</a></pre>
+<pre>private&nbsp;<a href="../../../../../org/apache/hadoop/hbase/client/Result.html" title="class in org.apache.hadoop.hbase.client">Result</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html#line.405">currentRowResult</a></pre>
 </li>
 </ul>
 <a name="nextCellInRow">
@@ -252,7 +252,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>nextCellInRow</h4>
-<pre>private&nbsp;int <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html#line.398">nextCellInRow</a></pre>
+<pre>private&nbsp;int <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html#line.406">nextCellInRow</a></pre>
 </li>
 </ul>
 <a name="nextRowResult">
@@ -261,7 +261,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockListLast">
 <li class="blockList">
 <h4>nextRowResult</h4>
-<pre>private&nbsp;<a href="../../../../../org/apache/hadoop/hbase/client/Result.html" title="class in org.apache.hadoop.hbase.client">Result</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html#line.400">nextRowResult</a></pre>
+<pre>private&nbsp;<a href="../../../../../org/apache/hadoop/hbase/client/Result.html" title="class in org.apache.hadoop.hbase.client">Result</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html#line.408">nextRowResult</a></pre>
 </li>
 </ul>
 </li>
@@ -278,7 +278,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockListLast">
 <li class="blockList">
 <h4>CellScanner</h4>
-<pre>public&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html#line.402">CellScanner</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Iterator.html?is-external=true" title="class or interface in java.util">Iterator</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/client/Result.html" title="class in org.apache.hadoop.hbase.client">Result</a>&gt;&nbsp;results)</pre>
+<pre>public&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html#line.410">CellScanner</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Iterator.html?is-external=true" title="class or interface in java.util">Iterator</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/client/Result.html" title="class in org.apache.hadoop.hbase.client">Result</a>&gt;&nbsp;results)</pre>
 </li>
 </ul>
 </li>
@@ -295,7 +295,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockList">
 <li class="blockList">
 <h4>nextRow</h4>
-<pre>public&nbsp;byte[]&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html#line.410">nextRow</a>()</pre>
+<pre>public&nbsp;byte[]&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html#line.418">nextRow</a>()</pre>
 <div class="block">Advance to the next row and return its row key.
  Returns null iff there are no more rows.</div>
 </li>
@@ -306,7 +306,7 @@ extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 <ul class="blockListLast">
 <li class="blockList">
 <h4>nextCellInRow</h4>
-<pre>public&nbsp;<a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html#line.446">nextCellInRow</a>()</pre>
+<pre>public&nbsp;<a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html#line.454">nextCellInRow</a>()</pre>
 <div class="block">Returns the next Cell in the current row or null iff none remain.</div>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html b/devapidocs/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html
index 589788b..23f9034 100644
--- a/devapidocs/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html
+++ b/devapidocs/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html
@@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
 </dl>
 <hr>
 <br>
-<pre>public static enum <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.198">SyncTable.SyncMapper.Counter</a>
+<pre>public static enum <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.204">SyncTable.SyncMapper.Counter</a>
 extends <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang">Enum</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a>&gt;</pre>
 </li>
 </ul>
@@ -246,7 +246,7 @@ the order they are declared.</div>
 <ul class="blockList">
 <li class="blockList">
 <h4>BATCHES</h4>
-<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.198">BATCHES</a></pre>
+<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.204">BATCHES</a></pre>
 </li>
 </ul>
 <a name="HASHES_MATCHED">
@@ -255,7 +255,7 @@ the order they are declared.</div>
 <ul class="blockList">
 <li class="blockList">
 <h4>HASHES_MATCHED</h4>
-<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.198">HASHES_MATCHED</a></pre>
+<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.204">HASHES_MATCHED</a></pre>
 </li>
 </ul>
 <a name="HASHES_NOT_MATCHED">
@@ -264,7 +264,7 @@ the order they are declared.</div>
 <ul class="blockList">
 <li class="blockList">
 <h4>HASHES_NOT_MATCHED</h4>
-<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.198">HASHES_NOT_MATCHED</a></pre>
+<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.204">HASHES_NOT_MATCHED</a></pre>
 </li>
 </ul>
 <a name="SOURCEMISSINGROWS">
@@ -273,7 +273,7 @@ the order they are declared.</div>
 <ul class="blockList">
 <li class="blockList">
 <h4>SOURCEMISSINGROWS</h4>
-<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.198">SOURCEMISSINGROWS</a></pre>
+<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.204">SOURCEMISSINGROWS</a></pre>
 </li>
 </ul>
 <a name="SOURCEMISSINGCELLS">
@@ -282,7 +282,7 @@ the order they are declared.</div>
 <ul class="blockList">
 <li class="blockList">
 <h4>SOURCEMISSINGCELLS</h4>
-<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.199">SOURCEMISSINGCELLS</a></pre>
+<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.205">SOURCEMISSINGCELLS</a></pre>
 </li>
 </ul>
 <a name="TARGETMISSINGROWS">
@@ -291,7 +291,7 @@ the order they are declared.</div>
 <ul class="blockList">
 <li class="blockList">
 <h4>TARGETMISSINGROWS</h4>
-<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.199">TARGETMISSINGROWS</a></pre>
+<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.205">TARGETMISSINGROWS</a></pre>
 </li>
 </ul>
 <a name="TARGETMISSINGCELLS">
@@ -300,7 +300,7 @@ the order they are declared.</div>
 <ul class="blockList">
 <li class="blockList">
 <h4>TARGETMISSINGCELLS</h4>
-<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.199">TARGETMISSINGCELLS</a></pre>
+<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.205">TARGETMISSINGCELLS</a></pre>
 </li>
 </ul>
 <a name="ROWSWITHDIFFS">
@@ -309,7 +309,7 @@ the order they are declared.</div>
 <ul class="blockList">
 <li class="blockList">
 <h4>ROWSWITHDIFFS</h4>
-<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.199">ROWSWITHDIFFS</a></pre>
+<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.205">ROWSWITHDIFFS</a></pre>
 </li>
 </ul>
 <a name="DIFFERENTCELLVALUES">
@@ -318,7 +318,7 @@ the order they are declared.</div>
 <ul class="blockList">
 <li class="blockList">
 <h4>DIFFERENTCELLVALUES</h4>
-<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.199">DIFFERENTCELLVALUES</a></pre>
+<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.205">DIFFERENTCELLVALUES</a></pre>
 </li>
 </ul>
 <a name="MATCHINGROWS">
@@ -327,7 +327,7 @@ the order they are declared.</div>
 <ul class="blockList">
 <li class="blockList">
 <h4>MATCHINGROWS</h4>
-<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.200">MATCHINGROWS</a></pre>
+<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.206">MATCHINGROWS</a></pre>
 </li>
 </ul>
 <a name="MATCHINGCELLS">
@@ -336,7 +336,7 @@ the order they are declared.</div>
 <ul class="blockList">
 <li class="blockList">
 <h4>MATCHINGCELLS</h4>
-<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.200">MATCHINGCELLS</a></pre>
+<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.206">MATCHINGCELLS</a></pre>
 </li>
 </ul>
 <a name="EMPTY_BATCHES">
@@ -345,7 +345,7 @@ the order they are declared.</div>
 <ul class="blockList">
 <li class="blockList">
 <h4>EMPTY_BATCHES</h4>
-<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.200">EMPTY_BATCHES</a></pre>
+<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.206">EMPTY_BATCHES</a></pre>
 </li>
 </ul>
 <a name="RANGESMATCHED">
@@ -354,7 +354,7 @@ the order they are declared.</div>
 <ul class="blockList">
 <li class="blockList">
 <h4>RANGESMATCHED</h4>
-<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.200">RANGESMATCHED</a></pre>
+<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.206">RANGESMATCHED</a></pre>
 </li>
 </ul>
 <a name="RANGESNOTMATCHED">
@@ -363,7 +363,7 @@ the order they are declared.</div>
 <ul class="blockListLast">
 <li class="blockList">
 <h4>RANGESNOTMATCHED</h4>
-<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.200">RANGESNOTMATCHED</a></pre>
+<pre>public static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.206">RANGESNOTMATCHED</a></pre>
 </li>
 </ul>
 </li>
@@ -380,7 +380,7 @@ the order they are declared.</div>
 <ul class="blockList">
 <li class="blockList">
 <h4>values</h4>
-<pre>public static&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a>[]&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.198">values</a>()</pre>
+<pre>public static&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a>[]&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.204">values</a>()</pre>
 <div class="block">Returns an array containing the constants of this enum type, in
 the order they are declared.  This method may be used to iterate
 over the constants as follows:
@@ -400,7 +400,7 @@ for (SyncTable.SyncMapper.Counter c : SyncTable.SyncMapper.Counter.values())
 <ul class="blockListLast">
 <li class="blockList">
 <h4>valueOf</h4>
-<pre>public static&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.198">valueOf</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&nbsp;name)</pre>
+<pre>public static&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.Counter</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html#line.204">valueOf</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&nbsp;name)</pre>
 <div class="block">Returns the enum constant of this type with the specified name.
 The string must match <i>exactly</i> an identifier used to declare an
 enum constant in this type.  (Extraneous whitespace characters are 
diff --git a/devapidocs/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html b/devapidocs/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html
index 235dc41..2fcaa40 100644
--- a/devapidocs/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html
+++ b/devapidocs/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html
@@ -18,7 +18,7 @@
     catch(err) {
     }
 //-->
-var methods = {"i0":10,"i1":9,"i2":9,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":9,"i9":9,"i10":10,"i11":10,"i12":10,"i13":9};
+var methods = {"i0":10,"i1":10,"i2":10,"i3":9,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":9,"i10":9,"i11":10,"i12":10,"i13":10,"i14":9};
 var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -123,7 +123,7 @@ var activeTableTab = "activeTableTab";
 </dl>
 <hr>
 <br>
-<pre>public static class <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.179">SyncTable.SyncMapper</a>
+<pre>public static class <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.184">SyncTable.SyncMapper</a>
 extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.html" title="class in org.apache.hadoop.hbase.mapreduce">TableMapper</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>,<a href="../../../../../org/apache/hadoop/hbase/client/Mutation.html" title="class in org.apache.hadoop.hbase.client">Mutation</a>&gt;</pre>
 </li>
 </ul>
@@ -194,42 +194,46 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#EMPTY_CELL_SCANNER">EMPTY_CELL_SCANNER</a></span></code>&nbsp;</td>
 </tr>
 <tr class="rowColor">
+<td class="colFirst"><code>(package private) boolean</code></td>
+<td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#ignoreTimestamp">ignoreTimestamp</a></span></code>&nbsp;</td>
+</tr>
+<tr class="altColor">
 <td class="colFirst"><code>(package private) <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true" title="class or interface in java.lang">Throwable</a></code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#mapperException">mapperException</a></span></code>&nbsp;</td>
 </tr>
-<tr class="altColor">
+<tr class="rowColor">
 <td class="colFirst"><code>(package private) <a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a></code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#nextSourceKey">nextSourceKey</a></span></code>&nbsp;</td>
 </tr>
-<tr class="rowColor">
+<tr class="altColor">
 <td class="colFirst"><code>(package private) <a href="../../../../../org/apache/hadoop/hbase/client/Connection.html" title="interface in org.apache.hadoop.hbase.client">Connection</a></code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#sourceConnection">sourceConnection</a></span></code>&nbsp;</td>
 </tr>
-<tr class="altColor">
+<tr class="rowColor">
 <td class="colFirst"><code>(package private) org.apache.hadoop.fs.Path</code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#sourceHashDir">sourceHashDir</a></span></code>&nbsp;</td>
 </tr>
-<tr class="rowColor">
+<tr class="altColor">
 <td class="colFirst"><code>(package private) <a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html" title="class in org.apache.hadoop.hbase.mapreduce">HashTable.TableHash.Reader</a></code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#sourceHashReader">sourceHashReader</a></span></code>&nbsp;</td>
 </tr>
-<tr class="altColor">
+<tr class="rowColor">
 <td class="colFirst"><code>(package private) <a href="../../../../../org/apache/hadoop/hbase/client/Table.html" title="interface in org.apache.hadoop.hbase.client">Table</a></code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#sourceTable">sourceTable</a></span></code>&nbsp;</td>
 </tr>
-<tr class="rowColor">
+<tr class="altColor">
 <td class="colFirst"><code>(package private) <a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html" title="class in org.apache.hadoop.hbase.mapreduce">HashTable.TableHash</a></code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#sourceTableHash">sourceTableHash</a></span></code>&nbsp;</td>
 </tr>
-<tr class="altColor">
+<tr class="rowColor">
 <td class="colFirst"><code>(package private) <a href="../../../../../org/apache/hadoop/hbase/client/Connection.html" title="interface in org.apache.hadoop.hbase.client">Connection</a></code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#targetConnection">targetConnection</a></span></code>&nbsp;</td>
 </tr>
-<tr class="rowColor">
+<tr class="altColor">
 <td class="colFirst"><code>(package private) <a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html" title="class in org.apache.hadoop.hbase.mapreduce">HashTable.ResultHasher</a></code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#targetHasher">targetHasher</a></span></code>&nbsp;</td>
 </tr>
-<tr class="altColor">
+<tr class="rowColor">
 <td class="colFirst"><code>(package private) <a href="../../../../../org/apache/hadoop/hbase/client/Table.html" title="interface in org.apache.hadoop.hbase.client">Table</a></code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#targetTable">targetTable</a></span></code>&nbsp;</td>
 </tr>
@@ -266,68 +270,72 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <th class="colLast" scope="col">Method and Description</th>
 </tr>
 <tr id="i0" class="altColor">
+<td class="colFirst"><code>private <a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a></code></td>
+<td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#checkAndResetTimestamp-org.apache.hadoop.hbase.Cell-">checkAndResetTimestamp</a></span>(<a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a>&nbsp;sourceCell)</code>&nbsp;</td>
+</tr>
+<tr id="i1" class="rowColor">
 <td class="colFirst"><code>protected void</code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#cleanup-org.apache.hadoop.mapreduce.Mapper.Context-">cleanup</a></span>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context)</code>&nbsp;</td>
 </tr>
-<tr id="i1" class="rowColor">
-<td class="colFirst"><code>private static int</code></td>
+<tr id="i2" class="altColor">
+<td class="colFirst"><code>private int</code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#compareCellKeysWithinRow-org.apache.hadoop.hbase.Cell-org.apache.hadoop.hbase.Cell-">compareCellKeysWithinRow</a></span>(<a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a>&nbsp;c1,
                         <a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a>&nbsp;c2)</code>
 <div class="block">Compare families, qualifiers, and timestamps of the given Cells.</div>
 </td>
 </tr>
-<tr id="i2" class="altColor">
+<tr id="i3" class="rowColor">
 <td class="colFirst"><code>private static int</code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#compareRowKeys-byte:A-byte:A-">compareRowKeys</a></span>(byte[]&nbsp;r1,
               byte[]&nbsp;r2)</code>
 <div class="block">Compare row keys of the given Result objects.</div>
 </td>
 </tr>
-<tr id="i3" class="rowColor">
+<tr id="i4" class="altColor">
 <td class="colFirst"><code>private void</code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#findNextKeyHashPair--">findNextKeyHashPair</a></span>()</code>
 <div class="block">Attempt to read the next source key/hash pair.</div>
 </td>
 </tr>
-<tr id="i4" class="altColor">
+<tr id="i5" class="rowColor">
 <td class="colFirst"><code>private void</code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#finishBatchAndCompareHashes-org.apache.hadoop.mapreduce.Mapper.Context-">finishBatchAndCompareHashes</a></span>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context)</code>
 <div class="block">Finish the currently open hash batch.</div>
 </td>
 </tr>
-<tr id="i5" class="rowColor">
+<tr id="i6" class="altColor">
 <td class="colFirst"><code>private void</code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#finishRemainingHashRanges-org.apache.hadoop.mapreduce.Mapper.Context-">finishRemainingHashRanges</a></span>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context)</code>&nbsp;</td>
 </tr>
-<tr id="i6" class="altColor">
+<tr id="i7" class="rowColor">
 <td class="colFirst"><code>protected void</code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#map-org.apache.hadoop.hbase.io.ImmutableBytesWritable-org.apache.hadoop.hbase.client.Result-org.apache.hadoop.mapreduce.Mapper.Context-">map</a></span>(<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&nbsp;key,
    <a href="../../../../../org/apache/hadoop/hbase/client/Result.html" title="class in org.apache.hadoop.hbase.client">Result</a>&nbsp;value,
    org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context)</code>&nbsp;</td>
 </tr>
-<tr id="i7" class="rowColor">
+<tr id="i8" class="altColor">
 <td class="colFirst"><code>private void</code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#moveToNextBatch-org.apache.hadoop.mapreduce.Mapper.Context-">moveToNextBatch</a></span>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context)</code>
 <div class="block">If there is an open hash batch, complete it and sync if there are diffs.</div>
 </td>
 </tr>
-<tr id="i8" class="altColor">
+<tr id="i9" class="rowColor">
 <td class="colFirst"><code>private static <a href="../../../../../org/apache/hadoop/hbase/client/Connection.html" title="interface in org.apache.hadoop.hbase.client">Connection</a></code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#openConnection-org.apache.hadoop.conf.Configuration-java.lang.String-java.lang.String-">openConnection</a></span>(org.apache.hadoop.conf.Configuration&nbsp;conf,
               <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&nbsp;zkClusterConfKey,
               <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&nbsp;configPrefix)</code>&nbsp;</td>
 </tr>
-<tr id="i9" class="rowColor">
+<tr id="i10" class="altColor">
 <td class="colFirst"><code>private static <a href="../../../../../org/apache/hadoop/hbase/client/Table.html" title="interface in org.apache.hadoop.hbase.client">Table</a></code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#openTable-org.apache.hadoop.hbase.client.Connection-org.apache.hadoop.conf.Configuration-java.lang.String-">openTable</a></span>(<a href="../../../../../org/apache/hadoop/hbase/client/Connection.html" title="interface in org.apache.hadoop.hbase.client">Connection</a>&nbsp;connection,
          org.apache.hadoop.conf.Configuration&nbsp;conf,
          <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&nbsp;tableNameConfKey)</code>&nbsp;</td>
 </tr>
-<tr id="i10" class="altColor">
+<tr id="i11" class="rowColor">
 <td class="colFirst"><code>protected void</code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#setup-org.apache.hadoop.mapreduce.Mapper.Context-">setup</a></span>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context)</code>&nbsp;</td>
 </tr>
-<tr id="i11" class="rowColor">
+<tr id="i12" class="altColor">
 <td class="colFirst"><code>private void</code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#syncRange-org.apache.hadoop.mapreduce.Mapper.Context-org.apache.hadoop.hbase.io.ImmutableBytesWritable-org.apache.hadoop.hbase.io.ImmutableBytesWritable-">syncRange</a></span>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context,
          <a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&nbsp;startRow,
@@ -335,7 +343,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <div class="block">Rescan the given range directly from the source and target tables.</div>
 </td>
 </tr>
-<tr id="i12" class="altColor">
+<tr id="i13" class="rowColor">
 <td class="colFirst"><code>private boolean</code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#syncRowCells-org.apache.hadoop.mapreduce.Mapper.Context-byte:A-org.apache.hadoop.hbase.mapreduce.SyncTable.SyncMapper.CellScanner-org.apache.hadoop.hbase.mapreduce.SyncTable.SyncMapper.CellScanner-">syncRowCells</a></span>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context,
             byte[]&nbsp;rowKey,
@@ -344,7 +352,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <div class="block">Compare the cells for the given row from the source and target tables.</div>
 </td>
 </tr>
-<tr id="i13" class="rowColor">
+<tr id="i14" class="altColor">
 <td class="colFirst"><code>private static <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a></code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#toHex-org.apache.hadoop.hbase.io.ImmutableBytesWritable-">toHex</a></span>(<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&nbsp;bytes)</code>&nbsp;</td>
 </tr>
@@ -383,7 +391,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>sourceHashDir</h4>
-<pre>org.apache.hadoop.fs.Path <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.180">sourceHashDir</a></pre>
+<pre>org.apache.hadoop.fs.Path <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.185">sourceHashDir</a></pre>
 </li>
 </ul>
 <a name="sourceConnection">
@@ -392,7 +400,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>sourceConnection</h4>
-<pre><a href="../../../../../org/apache/hadoop/hbase/client/Connection.html" title="interface in org.apache.hadoop.hbase.client">Connection</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.182">sourceConnection</a></pre>
+<pre><a href="../../../../../org/apache/hadoop/hbase/client/Connection.html" title="interface in org.apache.hadoop.hbase.client">Connection</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.187">sourceConnection</a></pre>
 </li>
 </ul>
 <a name="targetConnection">
@@ -401,7 +409,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>targetConnection</h4>
-<pre><a href="../../../../../org/apache/hadoop/hbase/client/Connection.html" title="interface in org.apache.hadoop.hbase.client">Connection</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.183">targetConnection</a></pre>
+<pre><a href="../../../../../org/apache/hadoop/hbase/client/Connection.html" title="interface in org.apache.hadoop.hbase.client">Connection</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.188">targetConnection</a></pre>
 </li>
 </ul>
 <a name="sourceTable">
@@ -410,7 +418,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>sourceTable</h4>
-<pre><a href="../../../../../org/apache/hadoop/hbase/client/Table.html" title="interface in org.apache.hadoop.hbase.client">Table</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.184">sourceTable</a></pre>
+<pre><a href="../../../../../org/apache/hadoop/hbase/client/Table.html" title="interface in org.apache.hadoop.hbase.client">Table</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.189">sourceTable</a></pre>
 </li>
 </ul>
 <a name="targetTable">
@@ -419,7 +427,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>targetTable</h4>
-<pre><a href="../../../../../org/apache/hadoop/hbase/client/Table.html" title="interface in org.apache.hadoop.hbase.client">Table</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.185">targetTable</a></pre>
+<pre><a href="../../../../../org/apache/hadoop/hbase/client/Table.html" title="interface in org.apache.hadoop.hbase.client">Table</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.190">targetTable</a></pre>
 </li>
 </ul>
 <a name="dryRun">
@@ -428,7 +436,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>dryRun</h4>
-<pre>boolean <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.186">dryRun</a></pre>
+<pre>boolean <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.191">dryRun</a></pre>
 </li>
 </ul>
 <a name="doDeletes">
@@ -437,7 +445,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>doDeletes</h4>
-<pre>boolean <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.187">doDeletes</a></pre>
+<pre>boolean <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.192">doDeletes</a></pre>
 </li>
 </ul>
 <a name="doPuts">
@@ -446,7 +454,16 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>doPuts</h4>
-<pre>boolean <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.188">doPuts</a></pre>
+<pre>boolean <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.193">doPuts</a></pre>
+</li>
+</ul>
+<a name="ignoreTimestamp">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>ignoreTimestamp</h4>
+<pre>boolean <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.194">ignoreTimestamp</a></pre>
 </li>
 </ul>
 <a name="sourceTableHash">
@@ -455,7 +472,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>sourceTableHash</h4>
-<pre><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html" title="class in org.apache.hadoop.hbase.mapreduce">HashTable.TableHash</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.190">sourceTableHash</a></pre>
+<pre><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html" title="class in org.apache.hadoop.hbase.mapreduce">HashTable.TableHash</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.196">sourceTableHash</a></pre>
 </li>
 </ul>
 <a name="sourceHashReader">
@@ -464,7 +481,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>sourceHashReader</h4>
-<pre><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html" title="class in org.apache.hadoop.hbase.mapreduce">HashTable.TableHash.Reader</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.191">sourceHashReader</a></pre>
+<pre><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html" title="class in org.apache.hadoop.hbase.mapreduce">HashTable.TableHash.Reader</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.197">sourceHashReader</a></pre>
 </li>
 </ul>
 <a name="currentSourceHash">
@@ -473,7 +490,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>currentSourceHash</h4>
-<pre><a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.192">currentSourceHash</a></pre>
+<pre><a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.198">currentSourceHash</a></pre>
 </li>
 </ul>
 <a name="nextSourceKey">
@@ -482,7 +499,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>nextSourceKey</h4>
-<pre><a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.193">nextSourceKey</a></pre>
+<pre><a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.199">nextSourceKey</a></pre>
 </li>
 </ul>
 <a name="targetHasher">
@@ -491,7 +508,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>targetHasher</h4>
-<pre><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html" title="class in org.apache.hadoop.hbase.mapreduce">HashTable.ResultHasher</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.194">targetHasher</a></pre>
+<pre><a href="../../../../../org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html" title="class in org.apache.hadoop.hbase.mapreduce">HashTable.ResultHasher</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.200">targetHasher</a></pre>
 </li>
 </ul>
 <a name="mapperException">
@@ -500,7 +517,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>mapperException</h4>
-<pre><a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true" title="class or interface in java.lang">Throwable</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.196">mapperException</a></pre>
+<pre><a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true" title="class or interface in java.lang">Throwable</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.202">mapperException</a></pre>
 </li>
 </ul>
 <a name="EMPTY_CELL_SCANNER">
@@ -509,7 +526,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockListLast">
 <li class="blockList">
 <h4>EMPTY_CELL_SCANNER</h4>
-<pre>private static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html" title="class in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.CellScanner</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.332">EMPTY_CELL_SCANNER</a></pre>
+<pre>private static final&nbsp;<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html" title="class in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.CellScanner</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.340">EMPTY_CELL_SCANNER</a></pre>
 </li>
 </ul>
 </li>
@@ -526,7 +543,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockListLast">
 <li class="blockList">
 <h4>SyncMapper</h4>
-<pre>public&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.179">SyncMapper</a>()</pre>
+<pre>public&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.184">SyncMapper</a>()</pre>
 </li>
 </ul>
 </li>
@@ -543,7 +560,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>setup</h4>
-<pre>protected&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.204">setup</a>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context)
+<pre>protected&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.210">setup</a>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context)
               throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
 <dl>
 <dt><span class="overrideSpecifyLabel">Overrides:</span></dt>
@@ -559,7 +576,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>openConnection</h4>
-<pre>private static&nbsp;<a href="../../../../../org/apache/hadoop/hbase/client/Connection.html" title="interface in org.apache.hadoop.hbase.client">Connection</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.233">openConnection</a>(org.apache.hadoop.conf.Configuration&nbsp;conf,
+<pre>private static&nbsp;<a href="../../../../../org/apache/hadoop/hbase/client/Connection.html" title="interface in org.apache.hadoop.hbase.client">Connection</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.241">openConnection</a>(org.apache.hadoop.conf.Configuration&nbsp;conf,
                                          <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&nbsp;zkClusterConfKey,
                                          <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&nbsp;configPrefix)
                                   throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
@@ -575,7 +592,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>openTable</h4>
-<pre>private static&nbsp;<a href="../../../../../org/apache/hadoop/hbase/client/Table.html" title="interface in org.apache.hadoop.hbase.client">Table</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.242">openTable</a>(<a href="../../../../../org/apache/hadoop/hbase/client/Connection.html" title="interface in org.apache.hadoop.hbase.client">Connection</a>&nbsp;connection,
+<pre>private static&nbsp;<a href="../../../../../org/apache/hadoop/hbase/client/Table.html" title="interface in org.apache.hadoop.hbase.client">Table</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.250">openTable</a>(<a href="../../../../../org/apache/hadoop/hbase/client/Connection.html" title="interface in org.apache.hadoop.hbase.client">Connection</a>&nbsp;connection,
                                org.apache.hadoop.conf.Configuration&nbsp;conf,
                                <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&nbsp;tableNameConfKey)
                         throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
@@ -591,7 +608,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>findNextKeyHashPair</h4>
-<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.251">findNextKeyHashPair</a>()
+<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.259">findNextKeyHashPair</a>()
                           throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
 <div class="block">Attempt to read the next source key/hash pair.
  If there are no more, set nextSourceKey to null</div>
@@ -607,7 +624,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>map</h4>
-<pre>protected&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.262">map</a>(<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&nbsp;key,
+<pre>protected&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.270">map</a>(<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&nbsp;key,
                    <a href="../../../../../org/apache/hadoop/hbase/client/Result.html" title="class in org.apache.hadoop.hbase.client">Result</a>&nbsp;value,
                    org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context)
             throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a>,
@@ -627,7 +644,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>moveToNextBatch</h4>
-<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.286">moveToNextBatch</a>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context)
+<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.294">moveToNextBatch</a>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context)
                       throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a>,
                              <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true" title="class or interface in java.lang">InterruptedException</a></pre>
 <div class="block">If there is an open hash batch, complete it and sync if there are diffs.
@@ -645,7 +662,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>finishBatchAndCompareHashes</h4>
-<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.301">finishBatchAndCompareHashes</a>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context)
+<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.309">finishBatchAndCompareHashes</a>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context)
                                   throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a>,
                                          <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true" title="class or interface in java.lang">InterruptedException</a></pre>
 <div class="block">Finish the currently open hash batch.
@@ -664,7 +681,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>toHex</h4>
-<pre>private static&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.328">toHex</a>(<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&nbsp;bytes)</pre>
+<pre>private static&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.336">toHex</a>(<a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&nbsp;bytes)</pre>
 </li>
 </ul>
 <a name="syncRange-org.apache.hadoop.mapreduce.Mapper.Context-org.apache.hadoop.hbase.io.ImmutableBytesWritable-org.apache.hadoop.hbase.io.ImmutableBytesWritable-">
@@ -673,7 +690,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>syncRange</h4>
-<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.340">syncRange</a>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context,
+<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.348">syncRange</a>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context,
                        <a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&nbsp;startRow,
                        <a href="../../../../../org/apache/hadoop/hbase/io/ImmutableBytesWritable.html" title="class in org.apache.hadoop.hbase.io">ImmutableBytesWritable</a>&nbsp;stopRow)
                 throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a>,
@@ -688,13 +705,22 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 </dl>
 </li>
 </ul>
+<a name="checkAndResetTimestamp-org.apache.hadoop.hbase.Cell-">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>checkAndResetTimestamp</h4>
+<pre>private&nbsp;<a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.486">checkAndResetTimestamp</a>(<a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a>&nbsp;sourceCell)</pre>
+</li>
+</ul>
 <a name="syncRowCells-org.apache.hadoop.mapreduce.Mapper.Context-byte:A-org.apache.hadoop.hbase.mapreduce.SyncTable.SyncMapper.CellScanner-org.apache.hadoop.hbase.mapreduce.SyncTable.SyncMapper.CellScanner-">
 <!--   -->
 </a>
 <ul class="blockList">
 <li class="blockList">
 <h4>syncRowCells</h4>
-<pre>private&nbsp;boolean&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.484">syncRowCells</a>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context,
+<pre>private&nbsp;boolean&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.509">syncRowCells</a>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context,
                              byte[]&nbsp;rowKey,
                              <a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html" title="class in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.CellScanner</a>&nbsp;sourceCells,
                              <a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.CellScanner.html" title="class in org.apache.hadoop.hbase.mapreduce">SyncTable.SyncMapper.CellScanner</a>&nbsp;targetCells)
@@ -717,7 +743,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>compareRowKeys</h4>
-<pre>private static&nbsp;int&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.593">compareRowKeys</a>(byte[]&nbsp;r1,
+<pre>private static&nbsp;int&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.620">compareRowKeys</a>(byte[]&nbsp;r1,
                                   byte[]&nbsp;r2)</pre>
 <div class="block">Compare row keys of the given Result objects.
  Nulls are after non-nulls</div>
@@ -729,8 +755,8 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>compareCellKeysWithinRow</h4>
-<pre>private static&nbsp;int&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.610">compareCellKeysWithinRow</a>(<a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a>&nbsp;c1,
-                                            <a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a>&nbsp;c2)</pre>
+<pre>private&nbsp;int&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.637">compareCellKeysWithinRow</a>(<a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a>&nbsp;c1,
+                                     <a href="../../../../../org/apache/hadoop/hbase/Cell.html" title="interface in org.apache.hadoop.hbase">Cell</a>&nbsp;c2)</pre>
 <div class="block">Compare families, qualifiers, and timestamps of the given Cells.
  They are assumed to be of the same row.
  Nulls are after non-nulls.</div>
@@ -742,7 +768,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockList">
 <li class="blockList">
 <h4>cleanup</h4>
-<pre>protected&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.633">cleanup</a>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context)
+<pre>protected&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.664">cleanup</a>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context)
                 throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a>,
                        <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true" title="class or interface in java.lang">InterruptedException</a></pre>
 <dl>
@@ -760,7 +786,7 @@ extends <a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableMapper.ht
 <ul class="blockListLast">
 <li class="blockList">
 <h4>finishRemainingHashRanges</h4>
-<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.664">finishRemainingHashRanges</a>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context)
+<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.html#line.695">finishRemainingHashRanges</a>(org.apache.hadoop.mapreduce.Mapper.Context&nbsp;context)
                                 throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a>,
                                        <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true" title="class or interface in java.lang">InterruptedException</a></pre>
 <dl>
diff --git a/devapidocs/org/apache/hadoop/hbase/mapreduce/SyncTable.html b/devapidocs/org/apache/hadoop/hbase/mapreduce/SyncTable.html
index 854f91b..31d7a56 100644
--- a/devapidocs/org/apache/hadoop/hbase/mapreduce/SyncTable.html
+++ b/devapidocs/org/apache/hadoop/hbase/mapreduce/SyncTable.html
@@ -119,7 +119,7 @@ var activeTableTab = "activeTableTab";
 <hr>
 <br>
 <pre>@InterfaceAudience.Private
-public class <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.58">SyncTable</a>
+public class <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.60">SyncTable</a>
 extends org.apache.hadoop.conf.Configured
 implements org.apache.hadoop.util.Tool</pre>
 </li>
@@ -188,6 +188,14 @@ implements org.apache.hadoop.util.Tool</pre>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.html#dryRun">dryRun</a></span></code>&nbsp;</td>
 </tr>
 <tr class="rowColor">
+<td class="colFirst"><code>(package private) static <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a></code></td>
+<td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.html#IGNORE_TIMESTAMPS">IGNORE_TIMESTAMPS</a></span></code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>(package private) boolean</code></td>
+<td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.html#ignoreTimestamps">ignoreTimestamps</a></span></code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
 <td class="colFirst"><code>private static org.slf4j.Logger</code></td>
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.html#LOG">LOG</a></span></code>&nbsp;</td>
 </tr>
@@ -336,7 +344,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>LOG</h4>
-<pre>private static final&nbsp;org.slf4j.Logger <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.60">LOG</a></pre>
+<pre>private static final&nbsp;org.slf4j.Logger <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.62">LOG</a></pre>
 </li>
 </ul>
 <a name="SOURCE_HASH_DIR_CONF_KEY">
@@ -345,7 +353,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>SOURCE_HASH_DIR_CONF_KEY</h4>
-<pre>static final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.62">SOURCE_HASH_DIR_CONF_KEY</a></pre>
+<pre>static final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.64">SOURCE_HASH_DIR_CONF_KEY</a></pre>
 <dl>
 <dt><span class="seeLabel">See Also:</span></dt>
 <dd><a href="../../../../../constant-values.html#org.apache.hadoop.hbase.mapreduce.SyncTable.SOURCE_HASH_DIR_CONF_KEY">Constant Field Values</a></dd>
@@ -358,7 +366,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>SOURCE_TABLE_CONF_KEY</h4>
-<pre>static final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.63">SOURCE_TABLE_CONF_KEY</a></pre>
+<pre>static final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.65">SOURCE_TABLE_CONF_KEY</a></pre>
 <dl>
 <dt><span class="seeLabel">See Also:</span></dt>
 <dd><a href="../../../../../constant-values.html#org.apache.hadoop.hbase.mapreduce.SyncTable.SOURCE_TABLE_CONF_KEY">Constant Field Values</a></dd>
@@ -371,7 +379,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>TARGET_TABLE_CONF_KEY</h4>
-<pre>static final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.64">TARGET_TABLE_CONF_KEY</a></pre>
+<pre>static final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.66">TARGET_TABLE_CONF_KEY</a></pre>
 <dl>
 <dt><span class="seeLabel">See Also:</span></dt>
 <dd><a href="../../../../../constant-values.html#org.apache.hadoop.hbase.mapreduce.SyncTable.TARGET_TABLE_CONF_KEY">Constant Field Values</a></dd>
@@ -384,7 +392,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>SOURCE_ZK_CLUSTER_CONF_KEY</h4>
-<pre>static final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.65">SOURCE_ZK_CLUSTER_CONF_KEY</a></pre>
+<pre>static final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.67">SOURCE_ZK_CLUSTER_CONF_KEY</a></pre>
 <dl>
 <dt><span class="seeLabel">See Also:</span></dt>
 <dd><a href="../../../../../constant-values.html#org.apache.hadoop.hbase.mapreduce.SyncTable.SOURCE_ZK_CLUSTER_CONF_KEY">Constant Field Values</a></dd>
@@ -397,7 +405,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>TARGET_ZK_CLUSTER_CONF_KEY</h4>
-<pre>static final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.66">TARGET_ZK_CLUSTER_CONF_KEY</a></pre>
+<pre>static final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.68">TARGET_ZK_CLUSTER_CONF_KEY</a></pre>
 <dl>
 <dt><span class="seeLabel">See Also:</span></dt>
 <dd><a href="../../../../../constant-values.html#org.apache.hadoop.hbase.mapreduce.SyncTable.TARGET_ZK_CLUSTER_CONF_KEY">Constant Field Values</a></dd>
@@ -410,7 +418,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>DRY_RUN_CONF_KEY</h4>
-<pre>static final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.67">DRY_RUN_CONF_KEY</a></pre>
+<pre>static final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.69">DRY_RUN_CONF_KEY</a></pre>
 <dl>
 <dt><span class="seeLabel">See Also:</span></dt>
 <dd><a href="../../../../../constant-values.html#org.apache.hadoop.hbase.mapreduce.SyncTable.DRY_RUN_CONF_KEY">Constant Field Values</a></dd>
@@ -423,7 +431,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>DO_DELETES_CONF_KEY</h4>
-<pre>static final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.68">DO_DELETES_CONF_KEY</a></pre>
+<pre>static final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.70">DO_DELETES_CONF_KEY</a></pre>
 <dl>
 <dt><span class="seeLabel">See Also:</span></dt>
 <dd><a href="../../../../../constant-values.html#org.apache.hadoop.hbase.mapreduce.SyncTable.DO_DELETES_CONF_KEY">Constant Field Values</a></dd>
@@ -436,20 +444,33 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>DO_PUTS_CONF_KEY</h4>
-<pre>static final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.69">DO_PUTS_CONF_KEY</a></pre>
+<pre>static final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.71">DO_PUTS_CONF_KEY</a></pre>
 <dl>
 <dt><span class="seeLabel">See Also:</span></dt>
 <dd><a href="../../../../../constant-values.html#org.apache.hadoop.hbase.mapreduce.SyncTable.DO_PUTS_CONF_KEY">Constant Field Values</a></dd>
 </dl>
 </li>
 </ul>
+<a name="IGNORE_TIMESTAMPS">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>IGNORE_TIMESTAMPS</h4>
+<pre>static final&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.72">IGNORE_TIMESTAMPS</a></pre>
+<dl>
+<dt><span class="seeLabel">See Also:</span></dt>
+<dd><a href="../../../../../constant-values.html#org.apache.hadoop.hbase.mapreduce.SyncTable.IGNORE_TIMESTAMPS">Constant Field Values</a></dd>
+</dl>
+</li>
+</ul>
 <a name="sourceHashDir">
 <!--   -->
 </a>
 <ul class="blockList">
 <li class="blockList">
 <h4>sourceHashDir</h4>
-<pre>org.apache.hadoop.fs.Path <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.71">sourceHashDir</a></pre>
+<pre>org.apache.hadoop.fs.Path <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.74">sourceHashDir</a></pre>
 </li>
 </ul>
 <a name="sourceTableName">
@@ -458,7 +479,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>sourceTableName</h4>
-<pre><a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.72">sourceTableName</a></pre>
+<pre><a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.75">sourceTableName</a></pre>
 </li>
 </ul>
 <a name="targetTableName">
@@ -467,7 +488,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>targetTableName</h4>
-<pre><a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.73">targetTableName</a></pre>
+<pre><a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.76">targetTableName</a></pre>
 </li>
 </ul>
 <a name="sourceZkCluster">
@@ -476,7 +497,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>sourceZkCluster</h4>
-<pre><a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.75">sourceZkCluster</a></pre>
+<pre><a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.78">sourceZkCluster</a></pre>
 </li>
 </ul>
 <a name="targetZkCluster">
@@ -485,7 +506,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>targetZkCluster</h4>
-<pre><a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.76">targetZkCluster</a></pre>
+<pre><a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a> <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.79">targetZkCluster</a></pre>
 </li>
 </ul>
 <a name="dryRun">
@@ -494,7 +515,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>dryRun</h4>
-<pre>boolean <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.77">dryRun</a></pre>
+<pre>boolean <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.80">dryRun</a></pre>
 </li>
 </ul>
 <a name="doDeletes">
@@ -503,7 +524,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>doDeletes</h4>
-<pre>boolean <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.78">doDeletes</a></pre>
+<pre>boolean <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.81">doDeletes</a></pre>
 </li>
 </ul>
 <a name="doPuts">
@@ -512,7 +533,16 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>doPuts</h4>
-<pre>boolean <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.79">doPuts</a></pre>
+<pre>boolean <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.82">doPuts</a></pre>
+</li>
+</ul>
+<a name="ignoreTimestamps">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>ignoreTimestamps</h4>
+<pre>boolean <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.83">ignoreTimestamps</a></pre>
 </li>
 </ul>
 <a name="counters">
@@ -521,7 +551,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>counters</h4>
-<pre>org.apache.hadoop.mapreduce.Counters <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.81">counters</a></pre>
+<pre>org.apache.hadoop.mapreduce.Counters <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.85">counters</a></pre>
 </li>
 </ul>
 <a name="NUM_ARGS">
@@ -530,7 +560,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockListLast">
 <li class="blockList">
 <h4>NUM_ARGS</h4>
-<pre>private static final&nbsp;int <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.709">NUM_ARGS</a></pre>
+<pre>private static final&nbsp;int <a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.740">NUM_ARGS</a></pre>
 <dl>
 <dt><span class="seeLabel">See Also:</span></dt>
 <dd><a href="../../../../../constant-values.html#org.apache.hadoop.hbase.mapreduce.SyncTable.NUM_ARGS">Constant Field Values</a></dd>
@@ -551,7 +581,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockListLast">
 <li class="blockList">
 <h4>SyncTable</h4>
-<pre>public&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.83">SyncTable</a>(org.apache.hadoop.conf.Configuration&nbsp;conf)</pre>
+<pre>public&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.87">SyncTable</a>(org.apache.hadoop.conf.Configuration&nbsp;conf)</pre>
 </li>
 </ul>
 </li>
@@ -568,7 +598,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>initCredentialsForHBase</h4>
-<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.87">initCredentialsForHBase</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&nbsp;zookeeper,
+<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.91">initCredentialsForHBase</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&nbsp;zookeeper,
                                      org.apache.hadoop.mapreduce.Job&nbsp;job)
                               throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
 <dl>
@@ -583,7 +613,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>createSubmittableJob</h4>
-<pre>public&nbsp;org.apache.hadoop.mapreduce.Job&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.95">createSubmittableJob</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>[]&nbsp;args)
+<pre>public&nbsp;org.apache.hadoop.mapreduce.Job&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.99">createSubmittableJob</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>[]&nbsp;args)
                                                      throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
 <dl>
 <dt><span class="throwsLabel">Throws:</span></dt>
@@ -597,7 +627,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>printUsage</h4>
-<pre>private static&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.710">printUsage</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&nbsp;errorMsg)</pre>
+<pre>private static&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.741">printUsage</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&nbsp;errorMsg)</pre>
 </li>
 </ul>
 <a name="doCommandLine-java.lang.String:A-">
@@ -606,7 +636,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>doCommandLine</h4>
-<pre>private&nbsp;boolean&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.745">doCommandLine</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>[]&nbsp;args)</pre>
+<pre>private&nbsp;boolean&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.780">doCommandLine</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>[]&nbsp;args)</pre>
 </li>
 </ul>
 <a name="main-java.lang.String:A-">
@@ -615,7 +645,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockList">
 <li class="blockList">
 <h4>main</h4>
-<pre>public static&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.808">main</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>[]&nbsp;args)
+<pre>public static&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.849">main</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>[]&nbsp;args)
                  throws <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true" title="class or interface in java.lang">Exception</a></pre>
 <div class="block">Main entry point.</div>
 <dl>
@@ -630,7 +660,7 @@ implements org.apache.hadoop.util.Tool</pre>
 <ul class="blockListLast">
 <li class="blockList">
 <h4>run</h4>
-<pre>public&nbsp;int&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.814">run</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>[]&nbsp;args)
+<pre>public&nbsp;int&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/mapreduce/SyncTable.html#line.855">run</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>[]&nbsp;args)
         throws <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true" title="class or interface in java.lang">Exception</a></pre>
 <dl>
 <dt><span class="overrideSpecifyLabel">Specified by:</span></dt>
diff --git a/devapidocs/org/apache/hadoop/hbase/mapreduce/package-tree.html b/devapidocs/org/apache/hadoop/hbase/mapreduce/package-tree.html
index c3f203e..84f5245 100644
--- a/devapidocs/org/apache/hadoop/hbase/mapreduce/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/mapreduce/package-tree.html
@@ -296,10 +296,10 @@
 <ul>
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
-<li type="circle">org.apache.hadoop.hbase.mapreduce.<a href="../../../../../org/apache/hadoop/hbase/mapreduce/CellCounter.CellCounterMapper.Counters.html" title="enum in org.apache.hadoop.hbase.mapreduce"><span class="typeNameLink">CellCounter.CellCounterMapper.Counters</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.mapreduce.<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce"><span class="typeNameLink">SyncTable.SyncMapper.Counter</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.mapreduce.<a href="../../../../../org/apache/hadoop/hbase/mapreduce/RowCounter.RowCounterMapper.Counters.html" title="enum in org.apache.hadoop.hbase.mapreduce"><span class="typeNameLink">RowCounter.RowCounterMapper.Counters</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.mapreduce.<a href="../../../../../org/apache/hadoop/hbase/mapreduce/CellCounter.CellCounterMapper.Counters.html" title="enum in org.apache.hadoop.hbase.mapreduce"><span class="typeNameLink">CellCounter.CellCounterMapper.Counters</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.mapreduce.<a href="../../../../../org/apache/hadoop/hbase/mapreduce/TableSplit.Version.html" title="enum in org.apache.hadoop.hbase.mapreduce"><span class="typeNameLink">TableSplit.Version</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.mapreduce.<a href="../../../../../org/apache/hadoop/hbase/mapreduce/SyncTable.SyncMapper.Counter.html" title="enum in org.apache.hadoop.hbase.mapreduce"><span class="typeNameLink">SyncTable.SyncMapper.Counter</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/package-tree.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/package-tree.html
index f8add9f..5c89e3f 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/assignment/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/package-tree.html
@@ -151,8 +151,8 @@
 <ul>
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
-<li type="circle">org.apache.hadoop.hbase.master.assignment.<a href="../../../../../../org/apache/hadoop/hbase/master/assignment/ServerState.html" title="enum in org.apache.hadoop.hbase.master.assignment"><span class="typeNameLink">ServerState</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.master.assignment.<a href="../../../../../../org/apache/hadoop/hbase/master/assignment/TransitRegionStateProcedure.TransitionType.html" title="enum in org.apache.hadoop.hbase.master.assignment"><span class="typeNameLink">TransitRegionStateProcedure.TransitionType</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.master.assignment.<a href="../../../../../../org/apache/hadoop/hbase/master/assignment/ServerState.html" title="enum in org.apache.hadoop.hbase.master.assignment"><span class="typeNameLink">ServerState</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/master/balancer/package-tree.html b/devapidocs/org/apache/hadoop/hbase/master/balancer/package-tree.html
index 39196e0..4aadbb8 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/balancer/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/balancer/package-tree.html
@@ -199,8 +199,8 @@
 <ul>
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
-<li type="circle">org.apache.hadoop.hbase.master.balancer.<a href="../../../../../../org/apache/hadoop/hbase/master/balancer/BaseLoadBalancer.Cluster.LocalityType.html" title="enum in org.apache.hadoop.hbase.master.balancer"><span class="typeNameLink">BaseLoadBalancer.Cluster.LocalityType</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.master.balancer.<a href="../../../../../../org/apache/hadoop/hbase/master/balancer/BaseLoadBalancer.Cluster.Action.Type.html" title="enum in org.apache.hadoop.hbase.master.balancer"><span class="typeNameLink">BaseLoadBalancer.Cluster.Action.Type</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.master.balancer.<a href="../../../../../../org/apache/hadoop/hbase/master/balancer/BaseLoadBalancer.Cluster.LocalityType.html" title="enum in org.apache.hadoop.hbase.master.balancer"><span class="typeNameLink">BaseLoadBalancer.Cluster.LocalityType</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/master/package-tree.html b/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
index 6c468eb..76084ea 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
@@ -355,13 +355,13 @@
 <ul>
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
-<li type="circle">org.apache.hadoop.hbase.master.<a href="../../../../../org/apache/hadoop/hbase/master/MetricsMasterSourceFactoryImpl.FactoryStorage.html" title="enum in org.apache.hadoop.hbase.master"><span class="typeNameLink">MetricsMasterSourceFactoryImpl.FactoryStorage</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.master.<a href="../../../../../org/apache/hadoop/hbase/master/RegionState.State.html" title="enum in org.apache.hadoop.hbase.master"><span class="typeNameLink">RegionState.State</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.master.<a href="../../../../../org/apache/hadoop/hbase/master/ServerManager.ServerLiveState.html" title="enum in org.apache.hadoop.hbase.master"><span class="typeNameLink">ServerManager.ServerLiveState</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.master.<a href="../../../../../org/apache/hadoop/hbase/master/SplitLogManager.ResubmitDirective.html" title="enum in org.apache.hadoop.hbase.master"><span class="typeNameLink">SplitLogManager.ResubmitDirective</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.master.<a href="../../../../../org/apache/hadoop/hbase/master/MetaRegionLocationCache.ZNodeOpType.html" title="enum in org.apache.hadoop.hbase.master"><span class="typeNameLink">MetaRegionLocationCache.ZNodeOpType</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.master.<a href="../../../../../org/apache/hadoop/hbase/master/SplitLogManager.TerminationStatus.html" title="enum in org.apache.hadoop.hbase.master"><span class="typeNameLink">SplitLogManager.TerminationStatus</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.master.<a href="../../../../../org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html" title="enum in org.apache.hadoop.hbase.master"><span class="typeNameLink">MasterRpcServices.BalanceSwitchMode</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.master.<a href="../../../../../org/apache/hadoop/hbase/master/RegionState.State.html" title="enum in org.apache.hadoop.hbase.master"><span class="typeNameLink">RegionState.State</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.master.<a href="../../../../../org/apache/hadoop/hbase/master/MetricsMasterSourceFactoryImpl.FactoryStorage.html" title="enum in org.apache.hadoop.hbase.master"><span class="typeNameLink">MetricsMasterSourceFactoryImpl.FactoryStorage</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.master.<a href="../../../../../org/apache/hadoop/hbase/master/MetaRegionLocationCache.ZNodeOpType.html" title="enum in org.apache.hadoop.hbase.master"><span class="typeNameLink">MetaRegionLocationCache.ZNodeOpType</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.master.<a href="../../../../../org/apache/hadoop/hbase/master/ServerManager.ServerLiveState.html" title="enum in org.apache.hadoop.hbase.master"><span class="typeNameLink">ServerManager.ServerLiveState</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html b/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html
index fe74f67..47984a0 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html
@@ -223,8 +223,8 @@
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
 <li type="circle">org.apache.hadoop.hbase.master.procedure.<a href="../../../../../../org/apache/hadoop/hbase/master/procedure/MetaProcedureInterface.MetaOperationType.html" title="enum in org.apache.hadoop.hbase.master.procedure"><span class="typeNameLink">MetaProcedureInterface.MetaOperationType</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.master.procedure.<a href="../../../../../../org/apache/hadoop/hbase/master/procedure/PeerProcedureInterface.PeerOperationType.html" title="enum in org.apache.hadoop.hbase.master.procedure"><span class="typeNameLink">PeerProcedureInterface.PeerOperationType</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.master.procedure.<a href="../../../../../../org/apache/hadoop/hbase/master/procedure/TableProcedureInterface.TableOperationType.html" title="enum in org.apache.hadoop.hbase.master.procedure"><span class="typeNameLink">TableProcedureInterface.TableOperationType</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.master.procedure.<a href="../../../../../../org/apache/hadoop/hbase/master/procedure/PeerProcedureInterface.PeerOperationType.html" title="enum in org.apache.hadoop.hbase.master.procedure"><span class="typeNameLink">PeerProcedureInterface.PeerOperationType</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.master.procedure.<a href="../../../../../../org/apache/hadoop/hbase/master/procedure/ServerProcedureInterface.ServerOperationType.html" title="enum in org.apache.hadoop.hbase.master.procedure"><span class="typeNameLink">ServerProcedureInterface.ServerOperationType</span></a></li>
 </ul>
 </li>
diff --git a/devapidocs/org/apache/hadoop/hbase/monitoring/package-tree.html b/devapidocs/org/apache/hadoop/hbase/monitoring/package-tree.html
index de7293f..e0e8840 100644
--- a/devapidocs/org/apache/hadoop/hbase/monitoring/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/monitoring/package-tree.html
@@ -127,8 +127,8 @@
 <ul>
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
-<li type="circle">org.apache.hadoop.hbase.monitoring.<a href="../../../../../org/apache/hadoop/hbase/monitoring/TaskMonitor.TaskFilter.TaskType.html" title="enum in org.apache.hadoop.hbase.monitoring"><span class="typeNameLink">TaskMonitor.TaskFilter.TaskType</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.monitoring.<a href="../../../../../org/apache/hadoop/hbase/monitoring/MonitoredTask.State.html" title="enum in org.apache.hadoop.hbase.monitoring"><span class="typeNameLink">MonitoredTask.State</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.monitoring.<a href="../../../../../org/apache/hadoop/hbase/monitoring/TaskMonitor.TaskFilter.TaskType.html" title="enum in org.apache.hadoop.hbase.monitoring"><span class="typeNameLink">TaskMonitor.TaskFilter.TaskType</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/package-tree.html b/devapidocs/org/apache/hadoop/hbase/package-tree.html
index 13b9f14..ff0ed97 100644
--- a/devapidocs/org/apache/hadoop/hbase/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/package-tree.html
@@ -429,19 +429,19 @@
 <ul>
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
-<li type="circle">org.apache.hadoop.hbase.<a href="../../../../org/apache/hadoop/hbase/CompareOperator.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">CompareOperator</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.<a href="../../../../org/apache/hadoop/hbase/ClusterMetrics.Option.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">ClusterMetrics.Option</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.<a href="../../../../org/apache/hadoop/hbase/Cell.Type.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">Cell.Type</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.<a href="../../../../org/apache/hadoop/hbase/MetaTableAccessor.QueryType.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">MetaTableAccessor.QueryType</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.<a href="../../../../org/apache/hadoop/hbase/CellBuilderType.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">CellBuilderType</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.<a href="../../../../org/apache/hadoop/hbase/HConstants.OperationStatusCode.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">HConstants.OperationStatusCode</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.<a href="../../../../org/apache/hadoop/hbase/MemoryCompactionPolicy.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">MemoryCompactionPolicy</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.<a href="../../../../org/apache/hadoop/hbase/CompareOperator.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">CompareOperator</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.<a href="../../../../org/apache/hadoop/hbase/Coprocessor.State.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">Coprocessor.State</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.<a href="../../../../org/apache/hadoop/hbase/KeyValue.Type.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">KeyValue.Type</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.<a href="../../../../org/apache/hadoop/hbase/CompatibilitySingletonFactory.SingletonStorage.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">CompatibilitySingletonFactory.SingletonStorage</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.<a href="../../../../org/apache/hadoop/hbase/HConstants.OperationStatusCode.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">HConstants.OperationStatusCode</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.<a href="../../../../org/apache/hadoop/hbase/KeepDeletedCells.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">KeepDeletedCells</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.<a href="../../../../org/apache/hadoop/hbase/Cell.Type.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">Cell.Type</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.<a href="../../../../org/apache/hadoop/hbase/Size.Unit.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">Size.Unit</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.<a href="../../../../org/apache/hadoop/hbase/MemoryCompactionPolicy.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">MemoryCompactionPolicy</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.<a href="../../../../org/apache/hadoop/hbase/ClusterMetrics.Option.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">ClusterMetrics.Option</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.<a href="../../../../org/apache/hadoop/hbase/HealthChecker.HealthCheckerExitStatus.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">HealthChecker.HealthCheckerExitStatus</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.<a href="../../../../org/apache/hadoop/hbase/KeyValue.Type.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">KeyValue.Type</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.<a href="../../../../org/apache/hadoop/hbase/Size.Unit.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">Size.Unit</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/procedure2/package-tree.html b/devapidocs/org/apache/hadoop/hbase/procedure2/package-tree.html
index b822673..0b27c8c 100644
--- a/devapidocs/org/apache/hadoop/hbase/procedure2/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/procedure2/package-tree.html
@@ -216,11 +216,11 @@
 <ul>
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
-<li type="circle">org.apache.hadoop.hbase.procedure2.<a href="../../../../../org/apache/hadoop/hbase/procedure2/RootProcedureState.State.html" title="enum in org.apache.hadoop.hbase.procedure2"><span class="typeNameLink">RootProcedureState.State</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.procedure2.<a href="../../../../../org/apache/hadoop/hbase/procedure2/Procedure.LockState.html" title="enum in org.apache.hadoop.hbase.procedure2"><span class="typeNameLink">Procedure.LockState</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.procedure2.<a href="../../../../../org/apache/hadoop/hbase/procedure2/StateMachineProcedure.Flow.html" title="enum in org.apache.hadoop.hbase.procedure2"><span class="typeNameLink">StateMachineProcedure.Flow</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.procedure2.<a href="../../../../../org/apache/hadoop/hbase/procedure2/LockType.html" title="enum in org.apache.hadoop.hbase.procedure2"><span class="typeNameLink">LockType</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.procedure2.<a href="../../../../../org/apache/hadoop/hbase/procedure2/LockedResourceType.html" title="enum in org.apache.hadoop.hbase.procedure2"><span class="typeNameLink">LockedResourceType</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.procedure2.<a href="../../../../../org/apache/hadoop/hbase/procedure2/Procedure.LockState.html" title="enum in org.apache.hadoop.hbase.procedure2"><span class="typeNameLink">Procedure.LockState</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.procedure2.<a href="../../../../../org/apache/hadoop/hbase/procedure2/StateMachineProcedure.Flow.html" title="enum in org.apache.hadoop.hbase.procedure2"><span class="typeNameLink">StateMachineProcedure.Flow</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.procedure2.<a href="../../../../../org/apache/hadoop/hbase/procedure2/RootProcedureState.State.html" title="enum in org.apache.hadoop.hbase.procedure2"><span class="typeNameLink">RootProcedureState.State</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/procedure2/store/wal/package-tree.html b/devapidocs/org/apache/hadoop/hbase/procedure2/store/wal/package-tree.html
index 2525a0e..7b64815 100644
--- a/devapidocs/org/apache/hadoop/hbase/procedure2/store/wal/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/procedure2/store/wal/package-tree.html
@@ -133,8 +133,8 @@
 <ul>
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
-<li type="circle">org.apache.hadoop.hbase.procedure2.store.wal.<a href="../../../../../../../org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.PushType.html" title="enum in org.apache.hadoop.hbase.procedure2.store.wal"><span class="typeNameLink">WALProcedureStore.PushType</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.procedure2.store.wal.<a href="../../../../../../../org/apache/hadoop/hbase/procedure2/store/wal/ProcedureStoreTracker.DeleteState.html" title="enum in org.apache.hadoop.hbase.procedure2.store.wal"><span class="typeNameLink">ProcedureStoreTracker.DeleteState</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.procedure2.store.wal.<a href="../../../../../../../org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.PushType.html" title="enum in org.apache.hadoop.hbase.procedure2.store.wal"><span class="typeNameLink">WALProcedureStore.PushType</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html b/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html
index 702dc6f..6ae4b68 100644
--- a/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html
@@ -241,11 +241,11 @@
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
 <li type="circle">org.apache.hadoop.hbase.quotas.<a href="../../../../../org/apache/hadoop/hbase/quotas/RpcThrottlingException.Type.html" title="enum in org.apache.hadoop.hbase.quotas"><span class="typeNameLink">RpcThrottlingException.Type</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.quotas.<a href="../../../../../org/apache/hadoop/hbase/quotas/QuotaScope.html" title="enum in org.apache.hadoop.hbase.quotas"><span class="typeNameLink">QuotaScope</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.quotas.<a href="../../../../../org/apache/hadoop/hbase/quotas/QuotaType.html" title="enum in org.apache.hadoop.hbase.quotas"><span class="typeNameLink">QuotaType</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.quotas.<a href="../../../../../org/apache/hadoop/hbase/quotas/OperationQuota.OperationType.html" title="enum in org.apache.hadoop.hbase.quotas"><span class="typeNameLink">OperationQuota.OperationType</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.quotas.<a href="../../../../../org/apache/hadoop/hbase/quotas/QuotaScope.html" title="enum in org.apache.hadoop.hbase.quotas"><span class="typeNameLink">QuotaScope</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.quotas.<a href="../../../../../org/apache/hadoop/hbase/quotas/SpaceViolationPolicy.html" title="enum in org.apache.hadoop.hbase.quotas"><span class="typeNameLink">SpaceViolationPolicy</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.quotas.<a href="../../../../../org/apache/hadoop/hbase/quotas/ThrottleType.html" title="enum in org.apache.hadoop.hbase.quotas"><span class="typeNameLink">ThrottleType</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.quotas.<a href="../../../../../org/apache/hadoop/hbase/quotas/OperationQuota.OperationType.html" title="enum in org.apache.hadoop.hbase.quotas"><span class="typeNameLink">OperationQuota.OperationType</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
index 648b271..1a60e68 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
@@ -735,20 +735,20 @@
 <ul>
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
-<li type="circle">org.apache.hadoop.hbase.regionserver.<a href="../../../../../org/apache/hadoop/hbase/regionserver/ChunkCreator.ChunkType.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">ChunkCreator.ChunkType</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.regionserver.<a href="../../../../../org/apache/hadoop/hbase/regionserver/ScanType.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">ScanType</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.regionserver.<a href="../../../../../org/apache/hadoop/hbase/regionserver/ScannerContext.LimitScope.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">ScannerContext.LimitScope</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.regionserver.<a href="../../../../../org/apache/hadoop/hbase/regionserver/BloomType.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">BloomType</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.regionserver.<a href="../../../../../org/apache/hadoop/hbase/regionserver/TimeRangeTracker.Type.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">TimeRangeTracker.Type</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.regionserver.<a href="../../../../../org/apache/hadoop/hbase/regionserver/MemStoreCompactionStrategy.Action.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">MemStoreCompactionStrategy.Action</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.regionserver.<a href="../../../../../org/apache/hadoop/hbase/regionserver/MetricsRegionServerSourceFactoryImpl.FactoryStorage.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">MetricsRegionServerSourceFactoryImpl.FactoryStorage</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.regionserver.<a href="../../../../../org/apache/hadoop/hbase/regionserver/Region.Operation.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">Region.Operation</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.regionserver.<a href="../../../../../org/apache/hadoop/hbase/regionserver/ScanType.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">ScanType</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.regionserver.<a href="../../../../../org/apache/hadoop/hbase/regionserver/FlushType.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">FlushType</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.regionserver.<a href="../../../../../org/apache/hadoop/hbase/regionserver/DefaultHeapMemoryTuner.StepDirection.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">DefaultHeapMemoryTuner.StepDirection</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.regionserver.<a href="../../../../../org/apache/hadoop/hbase/regionserver/TimeRangeTracker.Type.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">TimeRangeTracker.Type</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.regionserver.<a href="../../../../../org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">HRegion.FlushResult.Result</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.regionserver.<a href="../../../../../org/apache/hadoop/hbase/regionserver/ScannerContext.NextState.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">ScannerContext.NextState</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.regionserver.<a href="../../../../../org/apache/hadoop/hbase/regionserver/ScannerContext.LimitScope.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">ScannerContext.LimitScope</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.regionserver.<a href="../../../../../org/apache/hadoop/hbase/regionserver/SplitLogWorker.TaskExecutor.Status.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">SplitLogWorker.TaskExecutor.Status</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.regionserver.<a href="../../../../../org/apache/hadoop/hbase/regionserver/FlushType.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">FlushType</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.regionserver.<a href="../../../../../org/apache/hadoop/hbase/regionserver/BloomType.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">BloomType</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.regionserver.<a href="../../../../../org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">HRegion.FlushResult.Result</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.regionserver.<a href="../../../../../org/apache/hadoop/hbase/regionserver/CompactingMemStore.IndexType.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">CompactingMemStore.IndexType</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.regionserver.<a href="../../../../../org/apache/hadoop/hbase/regionserver/Region.Operation.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">Region.Operation</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.regionserver.<a href="../../../../../org/apache/hadoop/hbase/regionserver/MetricsRegionServerSourceFactoryImpl.FactoryStorage.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">MetricsRegionServerSourceFactoryImpl.FactoryStorage</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.regionserver.<a href="../../../../../org/apache/hadoop/hbase/regionserver/ScannerContext.NextState.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">ScannerContext.NextState</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.regionserver.<a href="../../../../../org/apache/hadoop/hbase/regionserver/ChunkCreator.ChunkType.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">ChunkCreator.ChunkType</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html
index cf0a47c..6ef0a33 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html
@@ -131,8 +131,8 @@
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
 <li type="circle">org.apache.hadoop.hbase.regionserver.querymatcher.<a href="../../../../../../org/apache/hadoop/hbase/regionserver/querymatcher/ScanQueryMatcher.MatchCode.html" title="enum in org.apache.hadoop.hbase.regionserver.querymatcher"><span class="typeNameLink">ScanQueryMatcher.MatchCode</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.regionserver.querymatcher.<a href="../../../../../../org/apache/hadoop/hbase/regionserver/querymatcher/StripeCompactionScanQueryMatcher.DropDeletesInOutput.html" title="enum in org.apache.hadoop.hbase.regionserver.querymatcher"><span class="typeNameLink">StripeCompactionScanQueryMatcher.DropDeletesInOutput</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.regionserver.querymatcher.<a href="../../../../../../org/apache/hadoop/hbase/regionserver/querymatcher/DeleteTracker.DeleteResult.html" title="enum in org.apache.hadoop.hbase.regionserver.querymatcher"><span class="typeNameLink">DeleteTracker.DeleteResult</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.regionserver.querymatcher.<a href="../../../../../../org/apache/hadoop/hbase/regionserver/querymatcher/StripeCompactionScanQueryMatcher.DropDeletesInOutput.html" title="enum in org.apache.hadoop.hbase.regionserver.querymatcher"><span class="typeNameLink">StripeCompactionScanQueryMatcher.DropDeletesInOutput</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/wal/package-tree.html b/devapidocs/org/apache/hadoop/hbase/regionserver/wal/package-tree.html
index a68742c..01364b3 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/wal/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/wal/package-tree.html
@@ -247,10 +247,10 @@
 <ul>
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
-<li type="circle">org.apache.hadoop.hbase.regionserver.wal.<a href="../../../../../../org/apache/hadoop/hbase/regionserver/wal/ProtobufLogReader.WALHdrResult.html" title="enum in org.apache.hadoop.hbase.regionserver.wal"><span class="typeNameLink">ProtobufLogReader.WALHdrResult</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.regionserver.wal.<a href="../../../../../../org/apache/hadoop/hbase/regionserver/wal/WALActionsListener.RollRequestReason.html" title="enum in org.apache.hadoop.hbase.regionserver.wal"><span class="typeNameLink">WALActionsListener.RollRequestReason</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.regionserver.wal.<a href="../../../../../../org/apache/hadoop/hbase/regionserver/wal/CompressionContext.DictionaryIndex.html" title="enum in org.apache.hadoop.hbase.regionserver.wal"><span class="typeNameLink">CompressionContext.DictionaryIndex</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.regionserver.wal.<a href="../../../../../../org/apache/hadoop/hbase/regionserver/wal/ProtobufLogReader.WALHdrResult.html" title="enum in org.apache.hadoop.hbase.regionserver.wal"><span class="typeNameLink">ProtobufLogReader.WALHdrResult</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.regionserver.wal.<a href="../../../../../../org/apache/hadoop/hbase/regionserver/wal/RingBufferTruck.Type.html" title="enum in org.apache.hadoop.hbase.regionserver.wal"><span class="typeNameLink">RingBufferTruck.Type</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.regionserver.wal.<a href="../../../../../../org/apache/hadoop/hbase/regionserver/wal/CompressionContext.DictionaryIndex.html" title="enum in org.apache.hadoop.hbase.regionserver.wal"><span class="typeNameLink">CompressionContext.DictionaryIndex</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/replication/package-tree.html b/devapidocs/org/apache/hadoop/hbase/replication/package-tree.html
index 732dfb2..581ff15 100644
--- a/devapidocs/org/apache/hadoop/hbase/replication/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/replication/package-tree.html
@@ -166,8 +166,8 @@
 <ul>
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
-<li type="circle">org.apache.hadoop.hbase.replication.<a href="../../../../../org/apache/hadoop/hbase/replication/ReplicationPeer.PeerState.html" title="enum in org.apache.hadoop.hbase.replication"><span class="typeNameLink">ReplicationPeer.PeerState</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.replication.<a href="../../../../../org/apache/hadoop/hbase/replication/SyncReplicationState.html" title="enum in org.apache.hadoop.hbase.replication"><span class="typeNameLink">SyncReplicationState</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.replication.<a href="../../../../../org/apache/hadoop/hbase/replication/ReplicationPeer.PeerState.html" title="enum in org.apache.hadoop.hbase.replication"><span class="typeNameLink">ReplicationPeer.PeerState</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html b/devapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html
index f682cb8..d79209d 100644
--- a/devapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html
@@ -205,8 +205,8 @@
 <ul>
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
-<li type="circle">org.apache.hadoop.hbase.replication.regionserver.<a href="../../../../../../org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceShipper.WorkerState.html" title="enum in org.apache.hadoop.hbase.replication.regionserver"><span class="typeNameLink">ReplicationSourceShipper.WorkerState</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.replication.regionserver.<a href="../../../../../../org/apache/hadoop/hbase/replication/regionserver/MetricsReplicationSourceFactoryImpl.SourceHolder.html" title="enum in org.apache.hadoop.hbase.replication.regionserver"><span class="typeNameLink">MetricsReplicationSourceFactoryImpl.SourceHolder</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.replication.regionserver.<a href="../../../../../../org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceShipper.WorkerState.html" title="enum in org.apache.hadoop.hbase.replication.regionserver"><span class="typeNameLink">ReplicationSourceShipper.WorkerState</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/security/access/package-tree.html b/devapidocs/org/apache/hadoop/hbase/security/access/package-tree.html
index 36a0013..f8d5e56 100644
--- a/devapidocs/org/apache/hadoop/hbase/security/access/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/security/access/package-tree.html
@@ -162,12 +162,12 @@
 <ul>
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
-<li type="circle">org.apache.hadoop.hbase.security.access.<a href="../../../../../../org/apache/hadoop/hbase/security/access/Permission.Action.html" title="enum in org.apache.hadoop.hbase.security.access"><span class="typeNameLink">Permission.Action</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.security.access.<a href="../../../../../../org/apache/hadoop/hbase/security/access/SnapshotScannerHDFSAclHelper.HDFSAclOperation.OperationType.html" title="enum in org.apache.hadoop.hbase.security.access"><span class="typeNameLink">SnapshotScannerHDFSAclHelper.HDFSAclOperation.OperationType</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.security.access.<a href="../../../../../../org/apache/hadoop/hbase/security/access/SnapshotScannerHDFSAclHelper.HDFSAclOperation.AclType.html" title="enum in org.apache.hadoop.hbase.security.access"><span class="typeNameLink">SnapshotScannerHDFSAclHelper.HDFSAclOperation.AclType</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.security.access.<a href="../../../../../../org/apache/hadoop/hbase/security/access/Permission.Scope.html" title="enum in org.apache.hadoop.hbase.security.access"><span class="typeNameLink">Permission.Scope</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.security.access.<a href="../../../../../../org/apache/hadoop/hbase/security/access/AccessControlFilter.Strategy.html" title="enum in org.apache.hadoop.hbase.security.access"><span class="typeNameLink">AccessControlFilter.Strategy</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.security.access.<a href="../../../../../../org/apache/hadoop/hbase/security/access/SnapshotScannerHDFSAclHelper.HDFSAclOperation.OperationType.html" title="enum in org.apache.hadoop.hbase.security.access"><span class="typeNameLink">SnapshotScannerHDFSAclHelper.HDFSAclOperation.OperationType</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.security.access.<a href="../../../../../../org/apache/hadoop/hbase/security/access/AccessController.OpType.html" title="enum in org.apache.hadoop.hbase.security.access"><span class="typeNameLink">AccessController.OpType</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.security.access.<a href="../../../../../../org/apache/hadoop/hbase/security/access/SnapshotScannerHDFSAclHelper.HDFSAclOperation.AclType.html" title="enum in org.apache.hadoop.hbase.security.access"><span class="typeNameLink">SnapshotScannerHDFSAclHelper.HDFSAclOperation.AclType</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.security.access.<a href="../../../../../../org/apache/hadoop/hbase/security/access/Permission.Action.html" title="enum in org.apache.hadoop.hbase.security.access"><span class="typeNameLink">Permission.Action</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/security/package-tree.html b/devapidocs/org/apache/hadoop/hbase/security/package-tree.html
index 062071e..c793825 100644
--- a/devapidocs/org/apache/hadoop/hbase/security/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/security/package-tree.html
@@ -189,9 +189,9 @@
 <ul>
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
-<li type="circle">org.apache.hadoop.hbase.security.<a href="../../../../../org/apache/hadoop/hbase/security/SaslStatus.html" title="enum in org.apache.hadoop.hbase.security"><span class="typeNameLink">SaslStatus</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.security.<a href="../../../../../org/apache/hadoop/hbase/security/SaslUtil.QualityOfProtection.html" title="enum in org.apache.hadoop.hbase.security"><span class="typeNameLink">SaslUtil.QualityOfProtection</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.security.<a href="../../../../../org/apache/hadoop/hbase/security/AuthMethod.html" title="enum in org.apache.hadoop.hbase.security"><span class="typeNameLink">AuthMethod</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.security.<a href="../../../../../org/apache/hadoop/hbase/security/SaslUtil.QualityOfProtection.html" title="enum in org.apache.hadoop.hbase.security"><span class="typeNameLink">SaslUtil.QualityOfProtection</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.security.<a href="../../../../../org/apache/hadoop/hbase/security/SaslStatus.html" title="enum in org.apache.hadoop.hbase.security"><span class="typeNameLink">SaslStatus</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html b/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html
index 1c4097f..adc046f 100644
--- a/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html
@@ -211,9 +211,9 @@
 <ul>
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
+<li type="circle">org.apache.hadoop.hbase.thrift.<a href="../../../../../org/apache/hadoop/hbase/thrift/MetricsThriftServerSourceFactoryImpl.FactoryStorage.html" title="enum in org.apache.hadoop.hbase.thrift"><span class="typeNameLink">MetricsThriftServerSourceFactoryImpl.FactoryStorage</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.thrift.<a href="../../../../../org/apache/hadoop/hbase/thrift/ImplType.html" title="enum in org.apache.hadoop.hbase.thrift"><span class="typeNameLink">ImplType</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.thrift.<a href="../../../../../org/apache/hadoop/hbase/thrift/ThriftMetrics.ThriftServerType.html" title="enum in org.apache.hadoop.hbase.thrift"><span class="typeNameLink">ThriftMetrics.ThriftServerType</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.thrift.<a href="../../../../../org/apache/hadoop/hbase/thrift/MetricsThriftServerSourceFactoryImpl.FactoryStorage.html" title="enum in org.apache.hadoop.hbase.thrift"><span class="typeNameLink">MetricsThriftServerSourceFactoryImpl.FactoryStorage</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html b/devapidocs/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html
index 60d625e..28f5a2d 100644
--- a/devapidocs/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html
+++ b/devapidocs/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html
@@ -512,7 +512,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#throwAndLogTableNotFoundException-org.apache.hadoop.hbase.TableName-">throwAndLogTableNotFoundException</a></span>(<a href="../../../../../org/apache/hadoop/hbase/TableName.html" title="class in org.apache.hadoop.hbase">TableName</a>&nbsp;tn)</code>&nbsp;</td>
 </tr>
 <tr id="i35" class="rowColor">
-<td class="colFirst"><code>private <a href="https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true" title="class or interface in java.util.concurrent">CompletableFuture</a>&lt;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true" title="class or interface in java.util">Collection</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache. [...]
+<td class="colFirst"><code>protected <a href="https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true" title="class or interface in java.util.concurrent">CompletableFuture</a>&lt;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true" title="class or interface in java.util">Collection</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apach [...]
 <td class="colLast"><code><span class="memberNameLink"><a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#tryAtomicRegionLoad-org.apache.hadoop.hbase.client.AsyncClusterConnection-org.apache.hadoop.hbase.TableName-boolean-byte:A-java.util.Collection-">tryAtomicRegionLoad</a></span>(<a href="../../../../../org/apache/hadoop/hbase/client/AsyncClusterConnection.html" title="interface in org.apache.hadoop.hbase.client">AsyncClusterConnection</a>&nbsp;conn,
                    <a href="../../../../../org/apache/hadoop/hbase/TableName.html" title="class in org.apache.hadoop.hbase">TableName</a>&nbsp;tableName,
                    boolean&nbsp;copyFiles,
@@ -963,16 +963,22 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>tryAtomicRegionLoad</h4>
-<pre>private&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true" title="class or interface in java.util.concurrent">CompletableFuture</a>&lt;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true" title="class or interface in java.util">Collection</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool [...]
-                                                                                        <a href="../../../../../org/apache/hadoop/hbase/TableName.html" title="class in org.apache.hadoop.hbase">TableName</a>&nbsp;tableName,
-                                                                                        boolean&nbsp;copyFiles,
-                                                                                        byte[]&nbsp;first,
-                                                                                        <a href="https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true" title="class or interface in java.util">Collection</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>&gt;&nbsp;lqis)</pre>
+<pre>protected&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true" title="class or interface in java.util.concurrent">CompletableFuture</a>&lt;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true" title="class or interface in java.util">Collection</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.to [...]
+                                                                                          <a href="../../../../../org/apache/hadoop/hbase/TableName.html" title="class in org.apache.hadoop.hbase">TableName</a>&nbsp;tableName,
+                                                                                          boolean&nbsp;copyFiles,
+                                                                                          byte[]&nbsp;first,
+                                                                                          <a href="https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true" title="class or interface in java.util">Collection</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>&gt;&nbsp;lqis)</pre>
 <div class="block">Attempts to do an atomic load of many hfiles into a region. If it fails, it returns a list of
  hfiles that need to be retried. If it is successful it will return an empty list. NOTE: To
  maintain row atomicity guarantees, region server side should succeed atomically and fails
  atomically.</div>
 <dl>
+<dt><span class="paramLabel">Parameters:</span></dt>
+<dd><code>conn</code> - Connection to use</dd>
+<dd><code>tableName</code> - Table to which these hfiles should be loaded to</dd>
+<dd><code>copyFiles</code> - whether replicate to peer cluster while bulkloading</dd>
+<dd><code>first</code> - the start key of region</dd>
+<dd><code>lqis</code> - hfiles should be loaded</dd>
 <dt><span class="returnLabel">Returns:</span></dt>
 <dd>empty list if success, list of items to retry on recoverable failure</dd>
 </dl>
@@ -984,7 +990,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>bulkLoadPhase</h4>
-<pre>protected&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.427">bulkLoadPhase</a>(<a href="../../../../../org/apache/hadoop/hbase/client/AsyncClusterConnection.html" title="interface in org.apache.hadoop.hbase.client">AsyncClusterConnection</a>&nbsp;conn,
+<pre>protected&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.433">bulkLoadPhase</a>(<a href="../../../../../org/apache/hadoop/hbase/client/AsyncClusterConnection.html" title="interface in org.apache.hadoop.hbase.client">AsyncClusterConnection</a>&nbsp;conn,
                              <a href="../../../../../org/apache/hadoop/hbase/TableName.html" title="class in org.apache.hadoop.hbase">TableName</a>&nbsp;tableName,
                              <a href="https://docs.oracle.com/javase/8/docs/api/java/util/Deque.html?is-external=true" title="class or interface in java.util">Deque</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>&gt;&nbsp;queue,
                              org.apache.hbase.thirdparty.com.google.common.collect.Multimap&lt;<a href="https://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true" title="class or interface in java.nio">ByteBuffer</a>,<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>&gt;&nbsp;regionGroups,
@@ -1007,7 +1013,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>groupByFamilies</h4>
-<pre>private&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true" title="class or interface in java.util">Map</a>&lt;byte[],<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true" title="class or interface in java.util">Collection</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>&gt;&gt;& [...]
+<pre>private&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true" title="class or interface in java.util">Map</a>&lt;byte[],<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true" title="class or interface in java.util">Collection</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>&gt;&gt;& [...]
 </li>
 </ul>
 <a name="checkHFilesCountPerRegionPerFamily-org.apache.hbase.thirdparty.com.google.common.collect.Multimap-">
@@ -1016,7 +1022,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>checkHFilesCountPerRegionPerFamily</h4>
-<pre>private&nbsp;boolean&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.485">checkHFilesCountPerRegionPerFamily</a>(org.apache.hbase.thirdparty.com.google.common.collect.Multimap&lt;<a href="https://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true" title="class or interface in java.nio">ByteBuffer</a>,<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apac [...]
+<pre>private&nbsp;boolean&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.491">checkHFilesCountPerRegionPerFamily</a>(org.apache.hbase.thirdparty.com.google.common.collect.Multimap&lt;<a href="https://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true" title="class or interface in java.nio">ByteBuffer</a>,<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apac [...]
 </li>
 </ul>
 <a name="groupOrSplitPhase-org.apache.hadoop.hbase.client.AsyncClusterConnection-org.apache.hadoop.hbase.TableName-java.util.concurrent.ExecutorService-java.util.Deque-java.util.List-">
@@ -1025,7 +1031,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>groupOrSplitPhase</h4>
-<pre>private&nbsp;<a href="../../../../../org/apache/hadoop/hbase/util/Pair.html" title="class in org.apache.hadoop.hbase.util">Pair</a>&lt;org.apache.hbase.thirdparty.com.google.common.collect.Multimap&lt;<a href="https://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true" title="class or interface in java.nio">ByteBuffer</a>,<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">Bulk [...]
+<pre>private&nbsp;<a href="../../../../../org/apache/hadoop/hbase/util/Pair.html" title="class in org.apache.hadoop.hbase.util">Pair</a>&lt;org.apache.hbase.thirdparty.com.google.common.collect.Multimap&lt;<a href="https://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true" title="class or interface in java.nio">ByteBuffer</a>,<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">Bulk [...]
                                                                                                                                                     <a href="../../../../../org/apache/hadoop/hbase/TableName.html" title="class in org.apache.hadoop.hbase">TableName</a>&nbsp;tableName,
                                                                                                                                                     <a href="https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true" title="class or interface in java.util.concurrent">ExecutorService</a>&nbsp;pool,
                                                                                                                                                     <a href="https://docs.oracle.com/javase/8/docs/api/java/util/Deque.html?is-external=true" title="class or interface in java.util">Deque</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>&gt;&nbsp;queue,
@@ -1051,7 +1057,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>getUniqueName</h4>
-<pre>private&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.560">getUniqueName</a>()</pre>
+<pre>private&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.566">getUniqueName</a>()</pre>
 </li>
 </ul>
 <a name="splitStoreFile-org.apache.hadoop.hbase.tool.BulkLoadHFiles.LoadQueueItem-org.apache.hadoop.hbase.client.TableDescriptor-byte:A-">
@@ -1060,7 +1066,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>splitStoreFile</h4>
-<pre>private&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true" title="class or interface in java.util">List</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>&gt;&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.564">splitStoreFile</a>(<a href="../../../../../org/apache/hadoop/ [...]
+<pre>private&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true" title="class or interface in java.util">List</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>&gt;&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.570">splitStoreFile</a>(<a href="../../../../../org/apache/hadoop/ [...]
                                                           <a href="../../../../../org/apache/hadoop/hbase/client/TableDescriptor.html" title="interface in org.apache.hadoop.hbase.client">TableDescriptor</a>&nbsp;tableDesc,
                                                           byte[]&nbsp;splitKey)
                                                    throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
@@ -1076,7 +1082,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>groupOrSplit</h4>
-<pre>protected&nbsp;<a href="../../../../../org/apache/hadoop/hbase/util/Pair.html" title="class in org.apache.hadoop.hbase.util">Pair</a>&lt;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true" title="class or interface in java.util">List</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>&gt;,<a href="https://docs.oracle.com/javas [...]
+<pre>protected&nbsp;<a href="../../../../../org/apache/hadoop/hbase/util/Pair.html" title="class in org.apache.hadoop.hbase.util">Pair</a>&lt;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true" title="class or interface in java.util">List</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>&gt;,<a href="https://docs.oracle.com/javas [...]
                                                                        <a href="../../../../../org/apache/hadoop/hbase/TableName.html" title="class in org.apache.hadoop.hbase">TableName</a>&nbsp;tableName,
                                                                        org.apache.hbase.thirdparty.com.google.common.collect.Multimap&lt;<a href="https://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true" title="class or interface in java.nio">ByteBuffer</a>,<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>&gt;&nbsp;regionGroups,
                                                                        <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>&nbsp;item,
@@ -1099,7 +1105,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>splitStoreFile</h4>
-<pre>static&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.692">splitStoreFile</a>(org.apache.hadoop.conf.Configuration&nbsp;conf,
+<pre>static&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.698">splitStoreFile</a>(org.apache.hadoop.conf.Configuration&nbsp;conf,
                            org.apache.hadoop.fs.Path&nbsp;inFile,
                            <a href="../../../../../org/apache/hadoop/hbase/client/ColumnFamilyDescriptor.html" title="interface in org.apache.hadoop.hbase.client">ColumnFamilyDescriptor</a>&nbsp;familyDesc,
                            byte[]&nbsp;splitKey,
@@ -1120,7 +1126,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>copyHFileHalf</h4>
-<pre>private static&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.705">copyHFileHalf</a>(org.apache.hadoop.conf.Configuration&nbsp;conf,
+<pre>private static&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.711">copyHFileHalf</a>(org.apache.hadoop.conf.Configuration&nbsp;conf,
                                   org.apache.hadoop.fs.Path&nbsp;inFile,
                                   org.apache.hadoop.fs.Path&nbsp;outFile,
                                   <a href="../../../../../org/apache/hadoop/hbase/io/Reference.html" title="class in org.apache.hadoop.hbase.io">Reference</a>&nbsp;reference,
@@ -1139,7 +1145,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>inferBoundaries</h4>
-<pre>public static&nbsp;byte[][]&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.777">inferBoundaries</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/util/SortedMap.html?is-external=true" title="class or interface in java.util">SortedMap</a>&lt;byte[],<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true" title="class or interface in java.lang">Integer</a>&gt;&nbsp;bdryMap)</pre>
+<pre>public static&nbsp;byte[][]&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.783">inferBoundaries</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/util/SortedMap.html?is-external=true" title="class or interface in java.util">SortedMap</a>&lt;byte[],<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true" title="class or interface in java.lang">Integer</a>&gt;&nbsp;bdryMap)</pre>
 <div class="block">Infers region boundaries for a new table.
  <p/>
  Parameter: <br/>
@@ -1168,7 +1174,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>createTable</h4>
-<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.803">createTable</a>(<a href="../../../../../org/apache/hadoop/hbase/TableName.html" title="class in org.apache.hadoop.hbase">TableName</a>&nbsp;tableName,
+<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.809">createTable</a>(<a href="../../../../../org/apache/hadoop/hbase/TableName.html" title="class in org.apache.hadoop.hbase">TableName</a>&nbsp;tableName,
                          org.apache.hadoop.fs.Path&nbsp;hfofDir,
                          <a href="../../../../../org/apache/hadoop/hbase/client/AsyncAdmin.html" title="interface in org.apache.hadoop.hbase.client">AsyncAdmin</a>&nbsp;admin)
                   throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
@@ -1186,7 +1192,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>performBulkLoad</h4>
-<pre>private&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true" title="class or interface in java.util">Map</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>,<a href="https://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true" title="class or interface in java.nio">ByteBuffer</a>&gt;&nbsp;<a href=".. [...]
+<pre>private&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true" title="class or interface in java.util">Map</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>,<a href="https://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true" title="class or interface in java.nio">ByteBuffer</a>&gt;&nbsp;<a href=".. [...]
                                                                      <a href="../../../../../org/apache/hadoop/hbase/TableName.html" title="class in org.apache.hadoop.hbase">TableName</a>&nbsp;tableName,
                                                                      <a href="https://docs.oracle.com/javase/8/docs/api/java/util/Deque.html?is-external=true" title="class or interface in java.util">Deque</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>&gt;&nbsp;queue,
                                                                      <a href="https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true" title="class or interface in java.util.concurrent">ExecutorService</a>&nbsp;pool,
@@ -1204,7 +1210,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>cleanup</h4>
-<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.903">cleanup</a>(<a href="../../../../../org/apache/hadoop/hbase/client/AsyncClusterConnection.html" title="interface in org.apache.hadoop.hbase.client">AsyncClusterConnection</a>&nbsp;conn,
+<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.909">cleanup</a>(<a href="../../../../../org/apache/hadoop/hbase/client/AsyncClusterConnection.html" title="interface in org.apache.hadoop.hbase.client">AsyncClusterConnection</a>&nbsp;conn,
                      <a href="../../../../../org/apache/hadoop/hbase/TableName.html" title="class in org.apache.hadoop.hbase">TableName</a>&nbsp;tableName,
                      <a href="https://docs.oracle.com/javase/8/docs/api/java/util/Deque.html?is-external=true" title="class or interface in java.util">Deque</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>&gt;&nbsp;queue,
                      <a href="https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true" title="class or interface in java.util.concurrent">ExecutorService</a>&nbsp;pool)
@@ -1221,7 +1227,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>doBulkLoad</h4>
-<pre>private&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true" title="class or interface in java.util">Map</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>,<a href="https://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true" title="class or interface in java.nio">ByteBuffer</a>&gt;&nbsp;<a href=".. [...]
+<pre>private&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true" title="class or interface in java.util">Map</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>,<a href="https://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true" title="class or interface in java.nio">ByteBuffer</a>&gt;&nbsp;<a href=".. [...]
                                                                 <a href="../../../../../org/apache/hadoop/hbase/TableName.html" title="class in org.apache.hadoop.hbase">TableName</a>&nbsp;tableName,
                                                                 <a href="https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true" title="class or interface in java.util">Map</a>&lt;byte[],<a href="https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true" title="class or interface in java.util">List</a>&lt;org.apache.hadoop.fs.Path&gt;&gt;&nbsp;map,
                                                                 boolean&nbsp;silence,
@@ -1246,7 +1252,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>doBulkLoad</h4>
-<pre>private&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true" title="class or interface in java.util">Map</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>,<a href="https://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true" title="class or interface in java.nio">ByteBuffer</a>&gt;&nbsp;<a href=".. [...]
+<pre>private&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true" title="class or interface in java.util">Map</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>,<a href="https://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true" title="class or interface in java.nio">ByteBuffer</a>&gt;&nbsp;<a href=".. [...]
                                                                 <a href="../../../../../org/apache/hadoop/hbase/TableName.html" title="class in org.apache.hadoop.hbase">TableName</a>&nbsp;tableName,
                                                                 org.apache.hadoop.fs.Path&nbsp;hfofDir,
                                                                 boolean&nbsp;silence,
@@ -1272,7 +1278,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>bulkLoad</h4>
-<pre>public&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true" title="class or interface in java.util">Map</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>,<a href="https://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true" title="class or interface in java.nio">ByteBuffer</a>&gt;&nbsp;<a href="../ [...]
+<pre>public&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true" title="class or interface in java.util">Map</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>,<a href="https://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true" title="class or interface in java.nio">ByteBuffer</a>&gt;&nbsp;<a href="../ [...]
                                                              <a href="https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true" title="class or interface in java.util">Map</a>&lt;byte[],<a href="https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true" title="class or interface in java.util">List</a>&lt;org.apache.hadoop.fs.Path&gt;&gt;&nbsp;family2Files)
                                                       throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
 <div class="block"><span class="descfrmTypeLabel">Description copied from interface:&nbsp;<code><a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.html#bulkLoad-org.apache.hadoop.hbase.TableName-java.util.Map-">BulkLoadHFiles</a></code></span></div>
@@ -1295,7 +1301,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>bulkLoad</h4>
-<pre>public&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true" title="class or interface in java.util">Map</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>,<a href="https://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true" title="class or interface in java.nio">ByteBuffer</a>&gt;&nbsp;<a href="../ [...]
+<pre>public&nbsp;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true" title="class or interface in java.util">Map</a>&lt;<a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apache.hadoop.hbase.tool">BulkLoadHFiles.LoadQueueItem</a>,<a href="https://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true" title="class or interface in java.nio">ByteBuffer</a>&gt;&nbsp;<a href="../ [...]
                                                              org.apache.hadoop.fs.Path&nbsp;dir)
                                                       throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
 <div class="block"><span class="descfrmTypeLabel">Description copied from interface:&nbsp;<code><a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.html#bulkLoad-org.apache.hadoop.hbase.TableName-org.apache.hadoop.fs.Path-">BulkLoadHFiles</a></code></span></div>
@@ -1319,7 +1325,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>tableExists</h4>
-<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.1028">tableExists</a>(<a href="../../../../../org/apache/hadoop/hbase/client/AsyncClusterConnection.html" title="interface in org.apache.hadoop.hbase.client">AsyncClusterConnection</a>&nbsp;conn,
+<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.1034">tableExists</a>(<a href="../../../../../org/apache/hadoop/hbase/client/AsyncClusterConnection.html" title="interface in org.apache.hadoop.hbase.client">AsyncClusterConnection</a>&nbsp;conn,
                          <a href="../../../../../org/apache/hadoop/hbase/TableName.html" title="class in org.apache.hadoop.hbase">TableName</a>&nbsp;tableName)
                   throws <a href="https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">IOException</a></pre>
 <dl>
@@ -1335,7 +1341,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>throwAndLogTableNotFoundException</h4>
-<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.1034">throwAndLogTableNotFoundException</a>(<a href="../../../../../org/apache/hadoop/hbase/TableName.html" title="class in org.apache.hadoop.hbase">TableName</a>&nbsp;tn)
+<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.1040">throwAndLogTableNotFoundException</a>(<a href="../../../../../org/apache/hadoop/hbase/TableName.html" title="class in org.apache.hadoop.hbase">TableName</a>&nbsp;tn)
                                         throws <a href="../../../../../org/apache/hadoop/hbase/TableNotFoundException.html" title="class in org.apache.hadoop.hbase">TableNotFoundException</a></pre>
 <dl>
 <dt><span class="throwsLabel">Throws:</span></dt>
@@ -1349,7 +1355,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>setBulkToken</h4>
-<pre>public&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.1040">setBulkToken</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&nbsp;bulkToken)</pre>
+<pre>public&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.1046">setBulkToken</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&nbsp;bulkToken)</pre>
 </li>
 </ul>
 <a name="setClusterIds-java.util.List-">
@@ -1358,7 +1364,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>setClusterIds</h4>
-<pre>public&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.1044">setClusterIds</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true" title="class or interface in java.util">List</a>&lt;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&gt;&nbsp;clusterIds)</pre>
+<pre>public&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.1050">setClusterIds</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true" title="class or interface in java.util">List</a>&lt;<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>&gt;&nbsp;clusterIds)</pre>
 </li>
 </ul>
 <a name="usage--">
@@ -1367,7 +1373,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>usage</h4>
-<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.1048">usage</a>()</pre>
+<pre>private&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.1054">usage</a>()</pre>
 </li>
 </ul>
 <a name="run-java.lang.String:A-">
@@ -1376,7 +1382,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>run</h4>
-<pre>public&nbsp;int&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.1065">run</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>[]&nbsp;args)
+<pre>public&nbsp;int&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.1071">run</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>[]&nbsp;args)
         throws <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true" title="class or interface in java.lang">Exception</a></pre>
 <dl>
 <dt><span class="overrideSpecifyLabel">Specified by:</span></dt>
@@ -1392,7 +1398,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>main</h4>
-<pre>public static&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.1092">main</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>[]&nbsp;args)
+<pre>public static&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.1098">main</a>(<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true" title="class or interface in java.lang">String</a>[]&nbsp;args)
                  throws <a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true" title="class or interface in java.lang">Exception</a></pre>
 <dl>
 <dt><span class="throwsLabel">Throws:</span></dt>
@@ -1406,7 +1412,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockList">
 <li class="blockList">
 <h4>disableReplication</h4>
-<pre>public&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.1099">disableReplication</a>()</pre>
+<pre>public&nbsp;void&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.1105">disableReplication</a>()</pre>
 <div class="block"><span class="descfrmTypeLabel">Description copied from interface:&nbsp;<code><a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.html#disableReplication--">BulkLoadHFiles</a></code></span></div>
 <div class="block">Disables replication for all bulkloads done via this instance,
  when bulkload replication is configured.</div>
@@ -1422,7 +1428,7 @@ implements <a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.h
 <ul class="blockListLast">
 <li class="blockList">
 <h4>isReplicationDisabled</h4>
-<pre>public&nbsp;boolean&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.1104">isReplicationDisabled</a>()</pre>
+<pre>public&nbsp;boolean&nbsp;<a href="../../../../../src-html/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#line.1110">isReplicationDisabled</a>()</pre>
 <dl>
 <dt><span class="overrideSpecifyLabel">Specified by:</span></dt>
 <dd><code><a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.html#isReplicationDisabled--">isReplicationDisabled</a></code>&nbsp;in interface&nbsp;<code><a href="../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.html" title="interface in org.apache.hadoop.hbase.tool">BulkLoadHFiles</a></code></dd>
diff --git a/devapidocs/org/apache/hadoop/hbase/tool/class-use/BulkLoadHFiles.LoadQueueItem.html b/devapidocs/org/apache/hadoop/hbase/tool/class-use/BulkLoadHFiles.LoadQueueItem.html
index 4de8c39..e101677 100644
--- a/devapidocs/org/apache/hadoop/hbase/tool/class-use/BulkLoadHFiles.LoadQueueItem.html
+++ b/devapidocs/org/apache/hadoop/hbase/tool/class-use/BulkLoadHFiles.LoadQueueItem.html
@@ -209,7 +209,7 @@
               byte[]&nbsp;splitKey)</code>&nbsp;</td>
 </tr>
 <tr class="rowColor">
-<td class="colFirst"><code>private <a href="https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true" title="class or interface in java.util.concurrent">CompletableFuture</a>&lt;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true" title="class or interface in java.util">Collection</a>&lt;<a href="../../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apac [...]
+<td class="colFirst"><code>protected <a href="https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true" title="class or interface in java.util.concurrent">CompletableFuture</a>&lt;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true" title="class or interface in java.util">Collection</a>&lt;<a href="../../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.ap [...]
 <td class="colLast"><span class="typeNameLabel">BulkLoadHFilesTool.</span><code><span class="memberNameLink"><a href="../../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#tryAtomicRegionLoad-org.apache.hadoop.hbase.client.AsyncClusterConnection-org.apache.hadoop.hbase.TableName-boolean-byte:A-java.util.Collection-">tryAtomicRegionLoad</a></span>(<a href="../../../../../../org/apache/hadoop/hbase/client/AsyncClusterConnection.html" title="interface in org.apache.hadoop.h [...]
                    <a href="../../../../../../org/apache/hadoop/hbase/TableName.html" title="class in org.apache.hadoop.hbase">TableName</a>&nbsp;tableName,
                    boolean&nbsp;copyFiles,
@@ -378,7 +378,7 @@
 </td>
 </tr>
 <tr class="altColor">
-<td class="colFirst"><code>private <a href="https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true" title="class or interface in java.util.concurrent">CompletableFuture</a>&lt;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true" title="class or interface in java.util">Collection</a>&lt;<a href="../../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.apac [...]
+<td class="colFirst"><code>protected <a href="https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true" title="class or interface in java.util.concurrent">CompletableFuture</a>&lt;<a href="https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true" title="class or interface in java.util">Collection</a>&lt;<a href="../../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFiles.LoadQueueItem.html" title="class in org.ap [...]
 <td class="colLast"><span class="typeNameLabel">BulkLoadHFilesTool.</span><code><span class="memberNameLink"><a href="../../../../../../org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.html#tryAtomicRegionLoad-org.apache.hadoop.hbase.client.AsyncClusterConnection-org.apache.hadoop.hbase.TableName-boolean-byte:A-java.util.Collection-">tryAtomicRegionLoad</a></span>(<a href="../../../../../../org/apache/hadoop/hbase/client/AsyncClusterConnection.html" title="interface in org.apache.hadoop.h [...]
                    <a href="../../../../../../org/apache/hadoop/hbase/TableName.html" title="class in org.apache.hadoop.hbase">TableName</a>&nbsp;tableName,
                    boolean&nbsp;copyFiles,
diff --git a/devapidocs/org/apache/hadoop/hbase/util/package-tree.html b/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
index 15be00e..49bd42e 100644
--- a/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
@@ -550,15 +550,15 @@
 <ul>
 <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a>&lt;E&gt; (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a>&lt;T&gt;, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true [...]
 <ul>
+<li type="circle">org.apache.hadoop.hbase.util.<a href="../../../../../org/apache/hadoop/hbase/util/IdReadWriteLockWithObjectPool.ReferenceType.html" title="enum in org.apache.hadoop.hbase.util"><span class="typeNameLink">IdReadWriteLockWithObjectPool.ReferenceType</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.util.<a href="../../../../../org/apache/hadoop/hbase/util/PrettyPrinter.Unit.html" title="enum in org.apache.hadoop.hbase.util"><span class="typeNameLink">PrettyPrinter.Unit</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.util.<a href="../../../../../org/apache/hadoop/hbase/util/Bytes.LexicographicalComparerHolder.UnsafeComparer.html" title="enum in org.apache.hadoop.hbase.util"><span class="typeNameLink">Bytes.LexicographicalComparerHolder.UnsafeComparer</span></a> (implements org.apache.hadoop.hbase.util.<a href="../../../../../org/apache/hadoop/hbase/util/Bytes.Comparer.html" title="interface in org.apache.hadoop.hbase.util">Bytes.Comparer</a>&lt;T&gt;)</li>
 <li type="circle">org.apache.hadoop.hbase.util.<a href="../../../../../org/apache/hadoop/hbase/util/DNS.ServerType.html" title="enum in org.apache.hadoop.hbase.util"><span class="typeNameLink">DNS.ServerType</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.util.<a href="../../../../../org/apache/hadoop/hbase/util/HbckErrorReporter.ERROR_CODE.html" title="enum in org.apache.hadoop.hbase.util"><span class="typeNameLink">HbckErrorReporter.ERROR_CODE</span></a></li>
+<li type="circle">org.apache.hadoop.hbase.util.<a href="../../../../../org/apache/hadoop/hbase/util/PoolMap.PoolType.html" title="enum in org.apache.hadoop.hbase.util"><span class="typeNameLink">PoolMap.PoolType</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.util.<a href="../../../../../org/apache/hadoop/hbase/util/Bytes.LexicographicalComparerHolder.PureJavaComparer.html" title="enum in org.apache.hadoop.hbase.util"><span class="typeNameLink">Bytes.LexicographicalComparerHolder.PureJavaComparer</span></a> (implements org.apache.hadoop.hbase.util.<a href="../../../../../org/apache/hadoop/hbase/util/Bytes.Comparer.html" title="interface in org.apache.hadoop.hbase.util">Bytes.Comparer</a>&lt;T&gt;)</li>
 <li type="circle">org.apache.hadoop.hbase.util.<a href="../../../../../org/apache/hadoop/hbase/util/Order.html" title="enum in org.apache.hadoop.hbase.util"><span class="typeNameLink">Order</span></a></li>
 <li type="circle">org.apache.hadoop.hbase.util.<a href="../../../../../org/apache/hadoop/hbase/util/ChecksumType.html" title="enum in org.apache.hadoop.hbase.util"><span class="typeNameLink">ChecksumType</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.util.<a href="../../../../../org/apache/hadoop/hbase/util/PoolMap.PoolType.html" title="enum in org.apache.hadoop.hbase.util"><span class="typeNameLink">PoolMap.PoolType</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.util.<a href="../../../../../org/apache/hadoop/hbase/util/PrettyPrinter.Unit.html" title="enum in org.apache.hadoop.hbase.util"><span class="typeNameLink">PrettyPrinter.Unit</span></a></li>
-<li type="circle">org.apache.hadoop.hbase.util.<a href="../../../../../org/apache/hadoop/hbase/util/Bytes.LexicographicalComparerHolder.UnsafeComparer.html" title="enum in org.apache.hadoop.hbase.util"><span class="typeNameLink">Bytes.LexicographicalComparerHolder.UnsafeComparer</span></a> (implements org.apache.hadoop.hbase.util.<a href="../../../../../org/apache/hadoop/hbase/util/Bytes.Comparer.html" title="interface in org.apache.hadoop.hbase.util">Bytes.Comparer</a>&lt;T&gt;)</li>
-<li type="circle">org.apache.hadoop.hbase.util.<a href="../../../../../org/apache/hadoop/hbase/util/IdReadWriteLockWithObjectPool.ReferenceType.html" title="enum in org.apache.hadoop.hbase.util"><span class="typeNameLink">IdReadWriteLockWithObjectPool.ReferenceType</span></a></li>
 </ul>
 </li>
 </ul>
diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/Version.html b/devapidocs/src-html/org/apache/hadoop/hbase/Version.html
index 6e1e25d..0acd8fb 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/Version.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/Version.html
@@ -18,9 +18,9 @@
 <span class="sourceLineNo">010</span>  justification="Intentional; to be modified in test")<a name="line.10"></a>
 <span class="sourceLineNo">011</span>public class Version {<a name="line.11"></a>
 <span class="sourceLineNo">012</span>  public static final String version = new String("3.0.0-SNAPSHOT");<a name="line.12"></a>
-<span class="sourceLineNo">013</span>  public static final String revision = "4f9eecbe61934bb6a28f0d6dde7c94ff04b649ed";<a name="line.13"></a>
+<span class="sourceLineNo">013</span>  public static final String revision = "c44bb2e99aa4a52280df74d46ff556cdfd14dac7";<a name="line.13"></a>
 <span class="sourceLineNo">014</span>  public static final String user = "jenkins";<a name="line.14"></a>
-<span class="sourceLineNo">015</span>  public static final String date = "Sat May  2 14:41:34 UTC 2020";<a name="line.15"></a>
+<span class="sourceLineNo">015</span>  public static final String date = "Mon May  4 14:39:58 UTC 2020";<a name="line.15"></a>
 <span class="sourceLineNo">016</span>  public static final String url = "git://jenkins-websites-he-de.apache.org/home/jenkins/jenkins-slave/workspace/hbase_generate_website/hbase";<a name="line.16"></a>
 <span class="sourceLineNo">017</span>  public static final String srcChecksum = "(stdin)=";<a name="line.17"></a>
 <span class="sourceLineNo">018</span>}<a name="line.18"></a>
diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HashTable.HashMapper.html b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HashTable.HashMapper.html
index 55fc1c9..32934c7 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HashTable.HashMapper.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HashTable.HashMapper.html
@@ -81,679 +81,697 @@
 <span class="sourceLineNo">073</span>  final static String MANIFEST_FILE_NAME = "manifest";<a name="line.73"></a>
 <span class="sourceLineNo">074</span>  final static String HASH_DATA_DIR = "hashes";<a name="line.74"></a>
 <span class="sourceLineNo">075</span>  final static String OUTPUT_DATA_FILE_PREFIX = "part-r-";<a name="line.75"></a>
-<span class="sourceLineNo">076</span>  private final static String TMP_MANIFEST_FILE_NAME = "manifest.tmp";<a name="line.76"></a>
-<span class="sourceLineNo">077</span><a name="line.77"></a>
-<span class="sourceLineNo">078</span>  TableHash tableHash = new TableHash();<a name="line.78"></a>
-<span class="sourceLineNo">079</span>  Path destPath;<a name="line.79"></a>
-<span class="sourceLineNo">080</span><a name="line.80"></a>
-<span class="sourceLineNo">081</span>  public HashTable(Configuration conf) {<a name="line.81"></a>
-<span class="sourceLineNo">082</span>    super(conf);<a name="line.82"></a>
-<span class="sourceLineNo">083</span>  }<a name="line.83"></a>
-<span class="sourceLineNo">084</span><a name="line.84"></a>
-<span class="sourceLineNo">085</span>  public static class TableHash {<a name="line.85"></a>
-<span class="sourceLineNo">086</span><a name="line.86"></a>
-<span class="sourceLineNo">087</span>    Path hashDir;<a name="line.87"></a>
-<span class="sourceLineNo">088</span><a name="line.88"></a>
-<span class="sourceLineNo">089</span>    String tableName;<a name="line.89"></a>
-<span class="sourceLineNo">090</span>    String families = null;<a name="line.90"></a>
-<span class="sourceLineNo">091</span>    long batchSize = DEFAULT_BATCH_SIZE;<a name="line.91"></a>
-<span class="sourceLineNo">092</span>    int numHashFiles = 0;<a name="line.92"></a>
-<span class="sourceLineNo">093</span>    byte[] startRow = HConstants.EMPTY_START_ROW;<a name="line.93"></a>
-<span class="sourceLineNo">094</span>    byte[] stopRow = HConstants.EMPTY_END_ROW;<a name="line.94"></a>
-<span class="sourceLineNo">095</span>    int scanBatch = 0;<a name="line.95"></a>
-<span class="sourceLineNo">096</span>    int versions = -1;<a name="line.96"></a>
-<span class="sourceLineNo">097</span>    long startTime = 0;<a name="line.97"></a>
-<span class="sourceLineNo">098</span>    long endTime = 0;<a name="line.98"></a>
-<span class="sourceLineNo">099</span><a name="line.99"></a>
-<span class="sourceLineNo">100</span>    List&lt;ImmutableBytesWritable&gt; partitions;<a name="line.100"></a>
+<span class="sourceLineNo">076</span>  final static String IGNORE_TIMESTAMPS = "ignoreTimestamps";<a name="line.76"></a>
+<span class="sourceLineNo">077</span>  private final static String TMP_MANIFEST_FILE_NAME = "manifest.tmp";<a name="line.77"></a>
+<span class="sourceLineNo">078</span><a name="line.78"></a>
+<span class="sourceLineNo">079</span>  TableHash tableHash = new TableHash();<a name="line.79"></a>
+<span class="sourceLineNo">080</span>  Path destPath;<a name="line.80"></a>
+<span class="sourceLineNo">081</span><a name="line.81"></a>
+<span class="sourceLineNo">082</span>  public HashTable(Configuration conf) {<a name="line.82"></a>
+<span class="sourceLineNo">083</span>    super(conf);<a name="line.83"></a>
+<span class="sourceLineNo">084</span>  }<a name="line.84"></a>
+<span class="sourceLineNo">085</span><a name="line.85"></a>
+<span class="sourceLineNo">086</span>  public static class TableHash {<a name="line.86"></a>
+<span class="sourceLineNo">087</span><a name="line.87"></a>
+<span class="sourceLineNo">088</span>    Path hashDir;<a name="line.88"></a>
+<span class="sourceLineNo">089</span><a name="line.89"></a>
+<span class="sourceLineNo">090</span>    String tableName;<a name="line.90"></a>
+<span class="sourceLineNo">091</span>    String families = null;<a name="line.91"></a>
+<span class="sourceLineNo">092</span>    long batchSize = DEFAULT_BATCH_SIZE;<a name="line.92"></a>
+<span class="sourceLineNo">093</span>    int numHashFiles = 0;<a name="line.93"></a>
+<span class="sourceLineNo">094</span>    byte[] startRow = HConstants.EMPTY_START_ROW;<a name="line.94"></a>
+<span class="sourceLineNo">095</span>    byte[] stopRow = HConstants.EMPTY_END_ROW;<a name="line.95"></a>
+<span class="sourceLineNo">096</span>    int scanBatch = 0;<a name="line.96"></a>
+<span class="sourceLineNo">097</span>    int versions = -1;<a name="line.97"></a>
+<span class="sourceLineNo">098</span>    long startTime = 0;<a name="line.98"></a>
+<span class="sourceLineNo">099</span>    long endTime = 0;<a name="line.99"></a>
+<span class="sourceLineNo">100</span>    boolean ignoreTimestamps;<a name="line.100"></a>
 <span class="sourceLineNo">101</span><a name="line.101"></a>
-<span class="sourceLineNo">102</span>    public static TableHash read(Configuration conf, Path hashDir) throws IOException {<a name="line.102"></a>
-<span class="sourceLineNo">103</span>      TableHash tableHash = new TableHash();<a name="line.103"></a>
-<span class="sourceLineNo">104</span>      FileSystem fs = hashDir.getFileSystem(conf);<a name="line.104"></a>
-<span class="sourceLineNo">105</span>      tableHash.hashDir = hashDir;<a name="line.105"></a>
-<span class="sourceLineNo">106</span>      tableHash.readPropertiesFile(fs, new Path(hashDir, MANIFEST_FILE_NAME));<a name="line.106"></a>
-<span class="sourceLineNo">107</span>      tableHash.readPartitionFile(fs, conf, new Path(hashDir, PARTITIONS_FILE_NAME));<a name="line.107"></a>
-<span class="sourceLineNo">108</span>      return tableHash;<a name="line.108"></a>
-<span class="sourceLineNo">109</span>    }<a name="line.109"></a>
-<span class="sourceLineNo">110</span><a name="line.110"></a>
-<span class="sourceLineNo">111</span>    void writePropertiesFile(FileSystem fs, Path path) throws IOException {<a name="line.111"></a>
-<span class="sourceLineNo">112</span>      Properties p = new Properties();<a name="line.112"></a>
-<span class="sourceLineNo">113</span>      p.setProperty("table", tableName);<a name="line.113"></a>
-<span class="sourceLineNo">114</span>      if (families != null) {<a name="line.114"></a>
-<span class="sourceLineNo">115</span>        p.setProperty("columnFamilies", families);<a name="line.115"></a>
-<span class="sourceLineNo">116</span>      }<a name="line.116"></a>
-<span class="sourceLineNo">117</span>      p.setProperty("targetBatchSize", Long.toString(batchSize));<a name="line.117"></a>
-<span class="sourceLineNo">118</span>      p.setProperty("numHashFiles", Integer.toString(numHashFiles));<a name="line.118"></a>
-<span class="sourceLineNo">119</span>      if (!isTableStartRow(startRow)) {<a name="line.119"></a>
-<span class="sourceLineNo">120</span>        p.setProperty("startRowHex", Bytes.toHex(startRow));<a name="line.120"></a>
-<span class="sourceLineNo">121</span>      }<a name="line.121"></a>
-<span class="sourceLineNo">122</span>      if (!isTableEndRow(stopRow)) {<a name="line.122"></a>
-<span class="sourceLineNo">123</span>        p.setProperty("stopRowHex", Bytes.toHex(stopRow));<a name="line.123"></a>
-<span class="sourceLineNo">124</span>      }<a name="line.124"></a>
-<span class="sourceLineNo">125</span>      if (scanBatch &gt; 0) {<a name="line.125"></a>
-<span class="sourceLineNo">126</span>        p.setProperty("scanBatch", Integer.toString(scanBatch));<a name="line.126"></a>
-<span class="sourceLineNo">127</span>      }<a name="line.127"></a>
-<span class="sourceLineNo">128</span>      if (versions &gt;= 0) {<a name="line.128"></a>
-<span class="sourceLineNo">129</span>        p.setProperty("versions", Integer.toString(versions));<a name="line.129"></a>
-<span class="sourceLineNo">130</span>      }<a name="line.130"></a>
-<span class="sourceLineNo">131</span>      if (startTime != 0) {<a name="line.131"></a>
-<span class="sourceLineNo">132</span>        p.setProperty("startTimestamp", Long.toString(startTime));<a name="line.132"></a>
-<span class="sourceLineNo">133</span>      }<a name="line.133"></a>
-<span class="sourceLineNo">134</span>      if (endTime != 0) {<a name="line.134"></a>
-<span class="sourceLineNo">135</span>        p.setProperty("endTimestamp", Long.toString(endTime));<a name="line.135"></a>
-<span class="sourceLineNo">136</span>      }<a name="line.136"></a>
-<span class="sourceLineNo">137</span><a name="line.137"></a>
-<span class="sourceLineNo">138</span>      try (OutputStreamWriter osw = new OutputStreamWriter(fs.create(path), Charsets.UTF_8)) {<a name="line.138"></a>
-<span class="sourceLineNo">139</span>        p.store(osw, null);<a name="line.139"></a>
-<span class="sourceLineNo">140</span>      }<a name="line.140"></a>
-<span class="sourceLineNo">141</span>    }<a name="line.141"></a>
-<span class="sourceLineNo">142</span><a name="line.142"></a>
-<span class="sourceLineNo">143</span>    void readPropertiesFile(FileSystem fs, Path path) throws IOException {<a name="line.143"></a>
-<span class="sourceLineNo">144</span>      Properties p = new Properties();<a name="line.144"></a>
-<span class="sourceLineNo">145</span>      try (FSDataInputStream in = fs.open(path)) {<a name="line.145"></a>
-<span class="sourceLineNo">146</span>        try (InputStreamReader isr = new InputStreamReader(in, Charsets.UTF_8)) {<a name="line.146"></a>
-<span class="sourceLineNo">147</span>          p.load(isr);<a name="line.147"></a>
-<span class="sourceLineNo">148</span>        }<a name="line.148"></a>
-<span class="sourceLineNo">149</span>      }<a name="line.149"></a>
-<span class="sourceLineNo">150</span>      tableName = p.getProperty("table");<a name="line.150"></a>
-<span class="sourceLineNo">151</span>      families = p.getProperty("columnFamilies");<a name="line.151"></a>
-<span class="sourceLineNo">152</span>      batchSize = Long.parseLong(p.getProperty("targetBatchSize"));<a name="line.152"></a>
-<span class="sourceLineNo">153</span>      numHashFiles = Integer.parseInt(p.getProperty("numHashFiles"));<a name="line.153"></a>
-<span class="sourceLineNo">154</span><a name="line.154"></a>
-<span class="sourceLineNo">155</span>      String startRowHex = p.getProperty("startRowHex");<a name="line.155"></a>
-<span class="sourceLineNo">156</span>      if (startRowHex != null) {<a name="line.156"></a>
-<span class="sourceLineNo">157</span>        startRow = Bytes.fromHex(startRowHex);<a name="line.157"></a>
-<span class="sourceLineNo">158</span>      }<a name="line.158"></a>
-<span class="sourceLineNo">159</span>      String stopRowHex = p.getProperty("stopRowHex");<a name="line.159"></a>
-<span class="sourceLineNo">160</span>      if (stopRowHex != null) {<a name="line.160"></a>
-<span class="sourceLineNo">161</span>        stopRow = Bytes.fromHex(stopRowHex);<a name="line.161"></a>
-<span class="sourceLineNo">162</span>      }<a name="line.162"></a>
-<span class="sourceLineNo">163</span><a name="line.163"></a>
-<span class="sourceLineNo">164</span>      String scanBatchString = p.getProperty("scanBatch");<a name="line.164"></a>
-<span class="sourceLineNo">165</span>      if (scanBatchString != null) {<a name="line.165"></a>
-<span class="sourceLineNo">166</span>        scanBatch = Integer.parseInt(scanBatchString);<a name="line.166"></a>
-<span class="sourceLineNo">167</span>      }<a name="line.167"></a>
-<span class="sourceLineNo">168</span><a name="line.168"></a>
-<span class="sourceLineNo">169</span>      String versionString = p.getProperty("versions");<a name="line.169"></a>
-<span class="sourceLineNo">170</span>      if (versionString != null) {<a name="line.170"></a>
-<span class="sourceLineNo">171</span>        versions = Integer.parseInt(versionString);<a name="line.171"></a>
-<span class="sourceLineNo">172</span>      }<a name="line.172"></a>
-<span class="sourceLineNo">173</span><a name="line.173"></a>
-<span class="sourceLineNo">174</span>      String startTimeString = p.getProperty("startTimestamp");<a name="line.174"></a>
-<span class="sourceLineNo">175</span>      if (startTimeString != null) {<a name="line.175"></a>
-<span class="sourceLineNo">176</span>        startTime = Long.parseLong(startTimeString);<a name="line.176"></a>
-<span class="sourceLineNo">177</span>      }<a name="line.177"></a>
-<span class="sourceLineNo">178</span><a name="line.178"></a>
-<span class="sourceLineNo">179</span>      String endTimeString = p.getProperty("endTimestamp");<a name="line.179"></a>
-<span class="sourceLineNo">180</span>      if (endTimeString != null) {<a name="line.180"></a>
-<span class="sourceLineNo">181</span>        endTime = Long.parseLong(endTimeString);<a name="line.181"></a>
-<span class="sourceLineNo">182</span>      }<a name="line.182"></a>
-<span class="sourceLineNo">183</span>    }<a name="line.183"></a>
-<span class="sourceLineNo">184</span><a name="line.184"></a>
-<span class="sourceLineNo">185</span>    Scan initScan() throws IOException {<a name="line.185"></a>
-<span class="sourceLineNo">186</span>      Scan scan = new Scan();<a name="line.186"></a>
-<span class="sourceLineNo">187</span>      scan.setCacheBlocks(false);<a name="line.187"></a>
-<span class="sourceLineNo">188</span>      if (startTime != 0 || endTime != 0) {<a name="line.188"></a>
-<span class="sourceLineNo">189</span>        scan.setTimeRange(startTime, endTime == 0 ? HConstants.LATEST_TIMESTAMP : endTime);<a name="line.189"></a>
-<span class="sourceLineNo">190</span>      }<a name="line.190"></a>
-<span class="sourceLineNo">191</span>      if (scanBatch &gt; 0) {<a name="line.191"></a>
-<span class="sourceLineNo">192</span>        scan.setBatch(scanBatch);<a name="line.192"></a>
-<span class="sourceLineNo">193</span>      }<a name="line.193"></a>
-<span class="sourceLineNo">194</span>      if (versions &gt;= 0) {<a name="line.194"></a>
-<span class="sourceLineNo">195</span>        scan.readVersions(versions);<a name="line.195"></a>
-<span class="sourceLineNo">196</span>      }<a name="line.196"></a>
-<span class="sourceLineNo">197</span>      if (!isTableStartRow(startRow)) {<a name="line.197"></a>
-<span class="sourceLineNo">198</span>        scan.withStartRow(startRow);<a name="line.198"></a>
-<span class="sourceLineNo">199</span>      }<a name="line.199"></a>
-<span class="sourceLineNo">200</span>      if (!isTableEndRow(stopRow)) {<a name="line.200"></a>
-<span class="sourceLineNo">201</span>        scan.withStopRow(stopRow);<a name="line.201"></a>
-<span class="sourceLineNo">202</span>      }<a name="line.202"></a>
-<span class="sourceLineNo">203</span>      if(families != null) {<a name="line.203"></a>
-<span class="sourceLineNo">204</span>        for(String fam : families.split(",")) {<a name="line.204"></a>
-<span class="sourceLineNo">205</span>          scan.addFamily(Bytes.toBytes(fam));<a name="line.205"></a>
-<span class="sourceLineNo">206</span>        }<a name="line.206"></a>
-<span class="sourceLineNo">207</span>      }<a name="line.207"></a>
-<span class="sourceLineNo">208</span>      return scan;<a name="line.208"></a>
-<span class="sourceLineNo">209</span>    }<a name="line.209"></a>
-<span class="sourceLineNo">210</span><a name="line.210"></a>
-<span class="sourceLineNo">211</span>    /**<a name="line.211"></a>
-<span class="sourceLineNo">212</span>     * Choose partitions between row ranges to hash to a single output file<a name="line.212"></a>
-<span class="sourceLineNo">213</span>     * Selects region boundaries that fall within the scan range, and groups them<a name="line.213"></a>
-<span class="sourceLineNo">214</span>     * into the desired number of partitions.<a name="line.214"></a>
-<span class="sourceLineNo">215</span>     */<a name="line.215"></a>
-<span class="sourceLineNo">216</span>    void selectPartitions(Pair&lt;byte[][], byte[][]&gt; regionStartEndKeys) {<a name="line.216"></a>
-<span class="sourceLineNo">217</span>      List&lt;byte[]&gt; startKeys = new ArrayList&lt;&gt;();<a name="line.217"></a>
-<span class="sourceLineNo">218</span>      for (int i = 0; i &lt; regionStartEndKeys.getFirst().length; i++) {<a name="line.218"></a>
-<span class="sourceLineNo">219</span>        byte[] regionStartKey = regionStartEndKeys.getFirst()[i];<a name="line.219"></a>
-<span class="sourceLineNo">220</span>        byte[] regionEndKey = regionStartEndKeys.getSecond()[i];<a name="line.220"></a>
-<span class="sourceLineNo">221</span><a name="line.221"></a>
-<span class="sourceLineNo">222</span>        // if scan begins after this region, or starts before this region, then drop this region<a name="line.222"></a>
-<span class="sourceLineNo">223</span>        // in other words:<a name="line.223"></a>
-<span class="sourceLineNo">224</span>        //   IF (scan begins before the end of this region<a name="line.224"></a>
-<span class="sourceLineNo">225</span>        //      AND scan ends before the start of this region)<a name="line.225"></a>
-<span class="sourceLineNo">226</span>        //   THEN include this region<a name="line.226"></a>
-<span class="sourceLineNo">227</span>        if ((isTableStartRow(startRow) || isTableEndRow(regionEndKey)<a name="line.227"></a>
-<span class="sourceLineNo">228</span>            || Bytes.compareTo(startRow, regionEndKey) &lt; 0)<a name="line.228"></a>
-<span class="sourceLineNo">229</span>          &amp;&amp; (isTableEndRow(stopRow) || isTableStartRow(regionStartKey)<a name="line.229"></a>
-<span class="sourceLineNo">230</span>            || Bytes.compareTo(stopRow, regionStartKey) &gt; 0)) {<a name="line.230"></a>
-<span class="sourceLineNo">231</span>          startKeys.add(regionStartKey);<a name="line.231"></a>
-<span class="sourceLineNo">232</span>        }<a name="line.232"></a>
-<span class="sourceLineNo">233</span>      }<a name="line.233"></a>
-<span class="sourceLineNo">234</span><a name="line.234"></a>
-<span class="sourceLineNo">235</span>      int numRegions = startKeys.size();<a name="line.235"></a>
-<span class="sourceLineNo">236</span>      if (numHashFiles == 0) {<a name="line.236"></a>
-<span class="sourceLineNo">237</span>        numHashFiles = numRegions / 100;<a name="line.237"></a>
-<span class="sourceLineNo">238</span>      }<a name="line.238"></a>
-<span class="sourceLineNo">239</span>      if (numHashFiles == 0) {<a name="line.239"></a>
-<span class="sourceLineNo">240</span>        numHashFiles = 1;<a name="line.240"></a>
-<span class="sourceLineNo">241</span>      }<a name="line.241"></a>
-<span class="sourceLineNo">242</span>      if (numHashFiles &gt; numRegions) {<a name="line.242"></a>
-<span class="sourceLineNo">243</span>        // can't partition within regions<a name="line.243"></a>
-<span class="sourceLineNo">244</span>        numHashFiles = numRegions;<a name="line.244"></a>
-<span class="sourceLineNo">245</span>      }<a name="line.245"></a>
-<span class="sourceLineNo">246</span><a name="line.246"></a>
-<span class="sourceLineNo">247</span>      // choose a subset of start keys to group regions into ranges<a name="line.247"></a>
-<span class="sourceLineNo">248</span>      partitions = new ArrayList&lt;&gt;(numHashFiles - 1);<a name="line.248"></a>
-<span class="sourceLineNo">249</span>      // skip the first start key as it is not a partition between ranges.<a name="line.249"></a>
-<span class="sourceLineNo">250</span>      for (long i = 1; i &lt; numHashFiles; i++) {<a name="line.250"></a>
-<span class="sourceLineNo">251</span>        int splitIndex = (int) (numRegions * i / numHashFiles);<a name="line.251"></a>
-<span class="sourceLineNo">252</span>        partitions.add(new ImmutableBytesWritable(startKeys.get(splitIndex)));<a name="line.252"></a>
-<span class="sourceLineNo">253</span>      }<a name="line.253"></a>
-<span class="sourceLineNo">254</span>    }<a name="line.254"></a>
-<span class="sourceLineNo">255</span><a name="line.255"></a>
-<span class="sourceLineNo">256</span>    void writePartitionFile(Configuration conf, Path path) throws IOException {<a name="line.256"></a>
-<span class="sourceLineNo">257</span>      FileSystem fs = path.getFileSystem(conf);<a name="line.257"></a>
-<span class="sourceLineNo">258</span>      @SuppressWarnings("deprecation")<a name="line.258"></a>
-<span class="sourceLineNo">259</span>      SequenceFile.Writer writer = SequenceFile.createWriter(<a name="line.259"></a>
-<span class="sourceLineNo">260</span>        fs, conf, path, ImmutableBytesWritable.class, NullWritable.class);<a name="line.260"></a>
-<span class="sourceLineNo">261</span><a name="line.261"></a>
-<span class="sourceLineNo">262</span>      for (int i = 0; i &lt; partitions.size(); i++) {<a name="line.262"></a>
-<span class="sourceLineNo">263</span>        writer.append(partitions.get(i), NullWritable.get());<a name="line.263"></a>
-<span class="sourceLineNo">264</span>      }<a name="line.264"></a>
-<span class="sourceLineNo">265</span>      writer.close();<a name="line.265"></a>
-<span class="sourceLineNo">266</span>    }<a name="line.266"></a>
-<span class="sourceLineNo">267</span><a name="line.267"></a>
-<span class="sourceLineNo">268</span>    private void readPartitionFile(FileSystem fs, Configuration conf, Path path)<a name="line.268"></a>
-<span class="sourceLineNo">269</span>         throws IOException {<a name="line.269"></a>
-<span class="sourceLineNo">270</span>      @SuppressWarnings("deprecation")<a name="line.270"></a>
-<span class="sourceLineNo">271</span>      SequenceFile.Reader reader = new SequenceFile.Reader(fs, path, conf);<a name="line.271"></a>
-<span class="sourceLineNo">272</span>      ImmutableBytesWritable key = new ImmutableBytesWritable();<a name="line.272"></a>
-<span class="sourceLineNo">273</span>      partitions = new ArrayList&lt;&gt;();<a name="line.273"></a>
-<span class="sourceLineNo">274</span>      while (reader.next(key)) {<a name="line.274"></a>
-<span class="sourceLineNo">275</span>        partitions.add(new ImmutableBytesWritable(key.copyBytes()));<a name="line.275"></a>
-<span class="sourceLineNo">276</span>      }<a name="line.276"></a>
-<span class="sourceLineNo">277</span>      reader.close();<a name="line.277"></a>
-<span class="sourceLineNo">278</span><a name="line.278"></a>
-<span class="sourceLineNo">279</span>      if (!Ordering.natural().isOrdered(partitions)) {<a name="line.279"></a>
-<span class="sourceLineNo">280</span>        throw new IOException("Partitions are not ordered!");<a name="line.280"></a>
-<span class="sourceLineNo">281</span>      }<a name="line.281"></a>
-<span class="sourceLineNo">282</span>    }<a name="line.282"></a>
-<span class="sourceLineNo">283</span><a name="line.283"></a>
-<span class="sourceLineNo">284</span>    @Override<a name="line.284"></a>
-<span class="sourceLineNo">285</span>    public String toString() {<a name="line.285"></a>
-<span class="sourceLineNo">286</span>      StringBuilder sb = new StringBuilder();<a name="line.286"></a>
-<span class="sourceLineNo">287</span>      sb.append("tableName=").append(tableName);<a name="line.287"></a>
-<span class="sourceLineNo">288</span>      if (families != null) {<a name="line.288"></a>
-<span class="sourceLineNo">289</span>        sb.append(", families=").append(families);<a name="line.289"></a>
-<span class="sourceLineNo">290</span>      }<a name="line.290"></a>
-<span class="sourceLineNo">291</span>      sb.append(", batchSize=").append(batchSize);<a name="line.291"></a>
-<span class="sourceLineNo">292</span>      sb.append(", numHashFiles=").append(numHashFiles);<a name="line.292"></a>
-<span class="sourceLineNo">293</span>      if (!isTableStartRow(startRow)) {<a name="line.293"></a>
-<span class="sourceLineNo">294</span>        sb.append(", startRowHex=").append(Bytes.toHex(startRow));<a name="line.294"></a>
-<span class="sourceLineNo">295</span>      }<a name="line.295"></a>
-<span class="sourceLineNo">296</span>      if (!isTableEndRow(stopRow)) {<a name="line.296"></a>
-<span class="sourceLineNo">297</span>        sb.append(", stopRowHex=").append(Bytes.toHex(stopRow));<a name="line.297"></a>
-<span class="sourceLineNo">298</span>      }<a name="line.298"></a>
-<span class="sourceLineNo">299</span>      if (scanBatch &gt;= 0) {<a name="line.299"></a>
-<span class="sourceLineNo">300</span>        sb.append(", scanBatch=").append(scanBatch);<a name="line.300"></a>
-<span class="sourceLineNo">301</span>      }<a name="line.301"></a>
-<span class="sourceLineNo">302</span>      if (versions &gt;= 0) {<a name="line.302"></a>
-<span class="sourceLineNo">303</span>        sb.append(", versions=").append(versions);<a name="line.303"></a>
-<span class="sourceLineNo">304</span>      }<a name="line.304"></a>
-<span class="sourceLineNo">305</span>      if (startTime != 0) {<a name="line.305"></a>
-<span class="sourceLineNo">306</span>        sb.append("startTime=").append(startTime);<a name="line.306"></a>
-<span class="sourceLineNo">307</span>      }<a name="line.307"></a>
-<span class="sourceLineNo">308</span>      if (endTime != 0) {<a name="line.308"></a>
-<span class="sourceLineNo">309</span>        sb.append("endTime=").append(endTime);<a name="line.309"></a>
-<span class="sourceLineNo">310</span>      }<a name="line.310"></a>
-<span class="sourceLineNo">311</span>      return sb.toString();<a name="line.311"></a>
-<span class="sourceLineNo">312</span>    }<a name="line.312"></a>
-<span class="sourceLineNo">313</span><a name="line.313"></a>
-<span class="sourceLineNo">314</span>    static String getDataFileName(int hashFileIndex) {<a name="line.314"></a>
-<span class="sourceLineNo">315</span>      return String.format(HashTable.OUTPUT_DATA_FILE_PREFIX + "%05d", hashFileIndex);<a name="line.315"></a>
-<span class="sourceLineNo">316</span>    }<a name="line.316"></a>
-<span class="sourceLineNo">317</span><a name="line.317"></a>
-<span class="sourceLineNo">318</span>    /**<a name="line.318"></a>
-<span class="sourceLineNo">319</span>     * Open a TableHash.Reader starting at the first hash at or after the given key.<a name="line.319"></a>
-<span class="sourceLineNo">320</span>     * @throws IOException<a name="line.320"></a>
-<span class="sourceLineNo">321</span>     */<a name="line.321"></a>
-<span class="sourceLineNo">322</span>    public Reader newReader(Configuration conf, ImmutableBytesWritable startKey)<a name="line.322"></a>
-<span class="sourceLineNo">323</span>        throws IOException {<a name="line.323"></a>
-<span class="sourceLineNo">324</span>      return new Reader(conf, startKey);<a name="line.324"></a>
-<span class="sourceLineNo">325</span>    }<a name="line.325"></a>
-<span class="sourceLineNo">326</span><a name="line.326"></a>
-<span class="sourceLineNo">327</span>    public class Reader implements java.io.Closeable {<a name="line.327"></a>
-<span class="sourceLineNo">328</span>      private final Configuration conf;<a name="line.328"></a>
-<span class="sourceLineNo">329</span><a name="line.329"></a>
-<span class="sourceLineNo">330</span>      private int hashFileIndex;<a name="line.330"></a>
-<span class="sourceLineNo">331</span>      private MapFile.Reader mapFileReader;<a name="line.331"></a>
-<span class="sourceLineNo">332</span><a name="line.332"></a>
-<span class="sourceLineNo">333</span>      private boolean cachedNext;<a name="line.333"></a>
-<span class="sourceLineNo">334</span>      private ImmutableBytesWritable key;<a name="line.334"></a>
-<span class="sourceLineNo">335</span>      private ImmutableBytesWritable hash;<a name="line.335"></a>
-<span class="sourceLineNo">336</span><a name="line.336"></a>
-<span class="sourceLineNo">337</span>      Reader(Configuration conf, ImmutableBytesWritable startKey) throws IOException {<a name="line.337"></a>
-<span class="sourceLineNo">338</span>        this.conf = conf;<a name="line.338"></a>
-<span class="sourceLineNo">339</span>        int partitionIndex = Collections.binarySearch(partitions, startKey);<a name="line.339"></a>
-<span class="sourceLineNo">340</span>        if (partitionIndex &gt;= 0) {<a name="line.340"></a>
-<span class="sourceLineNo">341</span>          // if the key is equal to a partition, then go the file after that partition<a name="line.341"></a>
-<span class="sourceLineNo">342</span>          hashFileIndex = partitionIndex+1;<a name="line.342"></a>
-<span class="sourceLineNo">343</span>        } else {<a name="line.343"></a>
-<span class="sourceLineNo">344</span>          // if the key is between partitions, then go to the file between those partitions<a name="line.344"></a>
-<span class="sourceLineNo">345</span>          hashFileIndex = -1-partitionIndex;<a name="line.345"></a>
-<span class="sourceLineNo">346</span>        }<a name="line.346"></a>
-<span class="sourceLineNo">347</span>        openHashFile();<a name="line.347"></a>
-<span class="sourceLineNo">348</span><a name="line.348"></a>
-<span class="sourceLineNo">349</span>        // MapFile's don't make it easy to seek() so that the subsequent next() returns<a name="line.349"></a>
-<span class="sourceLineNo">350</span>        // the desired key/value pair.  So we cache it for the first call of next().<a name="line.350"></a>
-<span class="sourceLineNo">351</span>        hash = new ImmutableBytesWritable();<a name="line.351"></a>
-<span class="sourceLineNo">352</span>        key = (ImmutableBytesWritable) mapFileReader.getClosest(startKey, hash);<a name="line.352"></a>
-<span class="sourceLineNo">353</span>        if (key == null) {<a name="line.353"></a>
-<span class="sourceLineNo">354</span>          cachedNext = false;<a name="line.354"></a>
-<span class="sourceLineNo">355</span>          hash = null;<a name="line.355"></a>
-<span class="sourceLineNo">356</span>        } else {<a name="line.356"></a>
-<span class="sourceLineNo">357</span>          cachedNext = true;<a name="line.357"></a>
-<span class="sourceLineNo">358</span>        }<a name="line.358"></a>
-<span class="sourceLineNo">359</span>      }<a name="line.359"></a>
-<span class="sourceLineNo">360</span><a name="line.360"></a>
-<span class="sourceLineNo">361</span>      /**<a name="line.361"></a>
-<span class="sourceLineNo">362</span>       * Read the next key/hash pair.<a name="line.362"></a>
-<span class="sourceLineNo">363</span>       * Returns true if such a pair exists and false when at the end of the data.<a name="line.363"></a>
-<span class="sourceLineNo">364</span>       */<a name="line.364"></a>
-<span class="sourceLineNo">365</span>      public boolean next() throws IOException {<a name="line.365"></a>
-<span class="sourceLineNo">366</span>        if (cachedNext) {<a name="line.366"></a>
-<span class="sourceLineNo">367</span>          cachedNext = false;<a name="line.367"></a>
-<span class="sourceLineNo">368</span>          return true;<a name="line.368"></a>
-<span class="sourceLineNo">369</span>        }<a name="line.369"></a>
-<span class="sourceLineNo">370</span>        key = new ImmutableBytesWritable();<a name="line.370"></a>
-<span class="sourceLineNo">371</span>        hash = new ImmutableBytesWritable();<a name="line.371"></a>
-<span class="sourceLineNo">372</span>        while (true) {<a name="line.372"></a>
-<span class="sourceLineNo">373</span>          boolean hasNext = mapFileReader.next(key, hash);<a name="line.373"></a>
-<span class="sourceLineNo">374</span>          if (hasNext) {<a name="line.374"></a>
-<span class="sourceLineNo">375</span>            return true;<a name="line.375"></a>
-<span class="sourceLineNo">376</span>          }<a name="line.376"></a>
-<span class="sourceLineNo">377</span>          hashFileIndex++;<a name="line.377"></a>
-<span class="sourceLineNo">378</span>          if (hashFileIndex &lt; TableHash.this.numHashFiles) {<a name="line.378"></a>
-<span class="sourceLineNo">379</span>            mapFileReader.close();<a name="line.379"></a>
-<span class="sourceLineNo">380</span>            openHashFile();<a name="line.380"></a>
-<span class="sourceLineNo">381</span>          } else {<a name="line.381"></a>
-<span class="sourceLineNo">382</span>            key = null;<a name="line.382"></a>
-<span class="sourceLineNo">383</span>            hash = null;<a name="line.383"></a>
-<span class="sourceLineNo">384</span>            return false;<a name="line.384"></a>
-<span class="sourceLineNo">385</span>          }<a name="line.385"></a>
-<span class="sourceLineNo">386</span>        }<a name="line.386"></a>
-<span class="sourceLineNo">387</span>      }<a name="line.387"></a>
-<span class="sourceLineNo">388</span><a name="line.388"></a>
-<span class="sourceLineNo">389</span>      /**<a name="line.389"></a>
-<span class="sourceLineNo">390</span>       * Get the current key<a name="line.390"></a>
-<span class="sourceLineNo">391</span>       * @return the current key or null if there is no current key<a name="line.391"></a>
-<span class="sourceLineNo">392</span>       */<a name="line.392"></a>
-<span class="sourceLineNo">393</span>      public ImmutableBytesWritable getCurrentKey() {<a name="line.393"></a>
-<span class="sourceLineNo">394</span>        return key;<a name="line.394"></a>
-<span class="sourceLineNo">395</span>      }<a name="line.395"></a>
-<span class="sourceLineNo">396</span><a name="line.396"></a>
-<span class="sourceLineNo">397</span>      /**<a name="line.397"></a>
-<span class="sourceLineNo">398</span>       * Get the current hash<a name="line.398"></a>
-<span class="sourceLineNo">399</span>       * @return the current hash or null if there is no current hash<a name="line.399"></a>
-<span class="sourceLineNo">400</span>       */<a name="line.400"></a>
-<span class="sourceLineNo">401</span>      public ImmutableBytesWritable getCurrentHash() {<a name="line.401"></a>
-<span class="sourceLineNo">402</span>        return hash;<a name="line.402"></a>
-<span class="sourceLineNo">403</span>      }<a name="line.403"></a>
-<span class="sourceLineNo">404</span><a name="line.404"></a>
-<span class="sourceLineNo">405</span>      private void openHashFile() throws IOException {<a name="line.405"></a>
-<span class="sourceLineNo">406</span>        if (mapFileReader != null) {<a name="line.406"></a>
-<span class="sourceLineNo">407</span>          mapFileReader.close();<a name="line.407"></a>
-<span class="sourceLineNo">408</span>        }<a name="line.408"></a>
-<span class="sourceLineNo">409</span>        Path dataDir = new Path(TableHash.this.hashDir, HASH_DATA_DIR);<a name="line.409"></a>
-<span class="sourceLineNo">410</span>        Path dataFile = new Path(dataDir, getDataFileName(hashFileIndex));<a name="line.410"></a>
-<span class="sourceLineNo">411</span>        mapFileReader = new MapFile.Reader(dataFile, conf);<a name="line.411"></a>
-<span class="sourceLineNo">412</span>      }<a name="line.412"></a>
-<span class="sourceLineNo">413</span><a name="line.413"></a>
-<span class="sourceLineNo">414</span>      @Override<a name="line.414"></a>
-<span class="sourceLineNo">415</span>      public void close() throws IOException {<a name="line.415"></a>
-<span class="sourceLineNo">416</span>        mapFileReader.close();<a name="line.416"></a>
-<span class="sourceLineNo">417</span>      }<a name="line.417"></a>
-<span class="sourceLineNo">418</span>    }<a name="line.418"></a>
-<span class="sourceLineNo">419</span>  }<a name="line.419"></a>
-<span class="sourceLineNo">420</span><a name="line.420"></a>
-<span class="sourceLineNo">421</span>  static boolean isTableStartRow(byte[] row) {<a name="line.421"></a>
-<span class="sourceLineNo">422</span>    return Bytes.equals(HConstants.EMPTY_START_ROW, row);<a name="line.422"></a>
-<span class="sourceLineNo">423</span>  }<a name="line.423"></a>
-<span class="sourceLineNo">424</span><a name="line.424"></a>
-<span class="sourceLineNo">425</span>  static boolean isTableEndRow(byte[] row) {<a name="line.425"></a>
-<span class="sourceLineNo">426</span>    return Bytes.equals(HConstants.EMPTY_END_ROW, row);<a name="line.426"></a>
-<span class="sourceLineNo">427</span>  }<a name="line.427"></a>
-<span class="sourceLineNo">428</span><a name="line.428"></a>
-<span class="sourceLineNo">429</span>  public Job createSubmittableJob(String[] args) throws IOException {<a name="line.429"></a>
-<span class="sourceLineNo">430</span>    Path partitionsPath = new Path(destPath, PARTITIONS_FILE_NAME);<a name="line.430"></a>
-<span class="sourceLineNo">431</span>    generatePartitions(partitionsPath);<a name="line.431"></a>
-<span class="sourceLineNo">432</span><a name="line.432"></a>
-<span class="sourceLineNo">433</span>    Job job = Job.getInstance(getConf(),<a name="line.433"></a>
-<span class="sourceLineNo">434</span>          getConf().get("mapreduce.job.name", "hashTable_" + tableHash.tableName));<a name="line.434"></a>
-<span class="sourceLineNo">435</span>    Configuration jobConf = job.getConfiguration();<a name="line.435"></a>
-<span class="sourceLineNo">436</span>    jobConf.setLong(HASH_BATCH_SIZE_CONF_KEY, tableHash.batchSize);<a name="line.436"></a>
-<span class="sourceLineNo">437</span>    job.setJarByClass(HashTable.class);<a name="line.437"></a>
-<span class="sourceLineNo">438</span><a name="line.438"></a>
-<span class="sourceLineNo">439</span>    TableMapReduceUtil.initTableMapperJob(tableHash.tableName, tableHash.initScan(),<a name="line.439"></a>
-<span class="sourceLineNo">440</span>        HashMapper.class, ImmutableBytesWritable.class, ImmutableBytesWritable.class, job);<a name="line.440"></a>
+<span class="sourceLineNo">102</span>    List&lt;ImmutableBytesWritable&gt; partitions;<a name="line.102"></a>
+<span class="sourceLineNo">103</span><a name="line.103"></a>
+<span class="sourceLineNo">104</span>    public static TableHash read(Configuration conf, Path hashDir) throws IOException {<a name="line.104"></a>
+<span class="sourceLineNo">105</span>      TableHash tableHash = new TableHash();<a name="line.105"></a>
+<span class="sourceLineNo">106</span>      FileSystem fs = hashDir.getFileSystem(conf);<a name="line.106"></a>
+<span class="sourceLineNo">107</span>      tableHash.hashDir = hashDir;<a name="line.107"></a>
+<span class="sourceLineNo">108</span>      tableHash.readPropertiesFile(fs, new Path(hashDir, MANIFEST_FILE_NAME));<a name="line.108"></a>
+<span class="sourceLineNo">109</span>      tableHash.readPartitionFile(fs, conf, new Path(hashDir, PARTITIONS_FILE_NAME));<a name="line.109"></a>
+<span class="sourceLineNo">110</span>      return tableHash;<a name="line.110"></a>
+<span class="sourceLineNo">111</span>    }<a name="line.111"></a>
+<span class="sourceLineNo">112</span><a name="line.112"></a>
+<span class="sourceLineNo">113</span>    void writePropertiesFile(FileSystem fs, Path path) throws IOException {<a name="line.113"></a>
+<span class="sourceLineNo">114</span>      Properties p = new Properties();<a name="line.114"></a>
+<span class="sourceLineNo">115</span>      p.setProperty("table", tableName);<a name="line.115"></a>
+<span class="sourceLineNo">116</span>      if (families != null) {<a name="line.116"></a>
+<span class="sourceLineNo">117</span>        p.setProperty("columnFamilies", families);<a name="line.117"></a>
+<span class="sourceLineNo">118</span>      }<a name="line.118"></a>
+<span class="sourceLineNo">119</span>      p.setProperty("targetBatchSize", Long.toString(batchSize));<a name="line.119"></a>
+<span class="sourceLineNo">120</span>      p.setProperty("numHashFiles", Integer.toString(numHashFiles));<a name="line.120"></a>
+<span class="sourceLineNo">121</span>      if (!isTableStartRow(startRow)) {<a name="line.121"></a>
+<span class="sourceLineNo">122</span>        p.setProperty("startRowHex", Bytes.toHex(startRow));<a name="line.122"></a>
+<span class="sourceLineNo">123</span>      }<a name="line.123"></a>
+<span class="sourceLineNo">124</span>      if (!isTableEndRow(stopRow)) {<a name="line.124"></a>
+<span class="sourceLineNo">125</span>        p.setProperty("stopRowHex", Bytes.toHex(stopRow));<a name="line.125"></a>
+<span class="sourceLineNo">126</span>      }<a name="line.126"></a>
+<span class="sourceLineNo">127</span>      if (scanBatch &gt; 0) {<a name="line.127"></a>
+<span class="sourceLineNo">128</span>        p.setProperty("scanBatch", Integer.toString(scanBatch));<a name="line.128"></a>
+<span class="sourceLineNo">129</span>      }<a name="line.129"></a>
+<span class="sourceLineNo">130</span>      if (versions &gt;= 0) {<a name="line.130"></a>
+<span class="sourceLineNo">131</span>        p.setProperty("versions", Integer.toString(versions));<a name="line.131"></a>
+<span class="sourceLineNo">132</span>      }<a name="line.132"></a>
+<span class="sourceLineNo">133</span>      if (startTime != 0) {<a name="line.133"></a>
+<span class="sourceLineNo">134</span>        p.setProperty("startTimestamp", Long.toString(startTime));<a name="line.134"></a>
+<span class="sourceLineNo">135</span>      }<a name="line.135"></a>
+<span class="sourceLineNo">136</span>      if (endTime != 0) {<a name="line.136"></a>
+<span class="sourceLineNo">137</span>        p.setProperty("endTimestamp", Long.toString(endTime));<a name="line.137"></a>
+<span class="sourceLineNo">138</span>      }<a name="line.138"></a>
+<span class="sourceLineNo">139</span><a name="line.139"></a>
+<span class="sourceLineNo">140</span>      try (OutputStreamWriter osw = new OutputStreamWriter(fs.create(path), Charsets.UTF_8)) {<a name="line.140"></a>
+<span class="sourceLineNo">141</span>        p.store(osw, null);<a name="line.141"></a>
+<span class="sourceLineNo">142</span>      }<a name="line.142"></a>
+<span class="sourceLineNo">143</span>    }<a name="line.143"></a>
+<span class="sourceLineNo">144</span><a name="line.144"></a>
+<span class="sourceLineNo">145</span>    void readPropertiesFile(FileSystem fs, Path path) throws IOException {<a name="line.145"></a>
+<span class="sourceLineNo">146</span>      Properties p = new Properties();<a name="line.146"></a>
+<span class="sourceLineNo">147</span>      try (FSDataInputStream in = fs.open(path)) {<a name="line.147"></a>
+<span class="sourceLineNo">148</span>        try (InputStreamReader isr = new InputStreamReader(in, Charsets.UTF_8)) {<a name="line.148"></a>
+<span class="sourceLineNo">149</span>          p.load(isr);<a name="line.149"></a>
+<span class="sourceLineNo">150</span>        }<a name="line.150"></a>
+<span class="sourceLineNo">151</span>      }<a name="line.151"></a>
+<span class="sourceLineNo">152</span>      tableName = p.getProperty("table");<a name="line.152"></a>
+<span class="sourceLineNo">153</span>      families = p.getProperty("columnFamilies");<a name="line.153"></a>
+<span class="sourceLineNo">154</span>      batchSize = Long.parseLong(p.getProperty("targetBatchSize"));<a name="line.154"></a>
+<span class="sourceLineNo">155</span>      numHashFiles = Integer.parseInt(p.getProperty("numHashFiles"));<a name="line.155"></a>
+<span class="sourceLineNo">156</span><a name="line.156"></a>
+<span class="sourceLineNo">157</span>      String startRowHex = p.getProperty("startRowHex");<a name="line.157"></a>
+<span class="sourceLineNo">158</span>      if (startRowHex != null) {<a name="line.158"></a>
+<span class="sourceLineNo">159</span>        startRow = Bytes.fromHex(startRowHex);<a name="line.159"></a>
+<span class="sourceLineNo">160</span>      }<a name="line.160"></a>
+<span class="sourceLineNo">161</span>      String stopRowHex = p.getProperty("stopRowHex");<a name="line.161"></a>
+<span class="sourceLineNo">162</span>      if (stopRowHex != null) {<a name="line.162"></a>
+<span class="sourceLineNo">163</span>        stopRow = Bytes.fromHex(stopRowHex);<a name="line.163"></a>
+<span class="sourceLineNo">164</span>      }<a name="line.164"></a>
+<span class="sourceLineNo">165</span><a name="line.165"></a>
+<span class="sourceLineNo">166</span>      String scanBatchString = p.getProperty("scanBatch");<a name="line.166"></a>
+<span class="sourceLineNo">167</span>      if (scanBatchString != null) {<a name="line.167"></a>
+<span class="sourceLineNo">168</span>        scanBatch = Integer.parseInt(scanBatchString);<a name="line.168"></a>
+<span class="sourceLineNo">169</span>      }<a name="line.169"></a>
+<span class="sourceLineNo">170</span><a name="line.170"></a>
+<span class="sourceLineNo">171</span>      String versionString = p.getProperty("versions");<a name="line.171"></a>
+<span class="sourceLineNo">172</span>      if (versionString != null) {<a name="line.172"></a>
+<span class="sourceLineNo">173</span>        versions = Integer.parseInt(versionString);<a name="line.173"></a>
+<span class="sourceLineNo">174</span>      }<a name="line.174"></a>
+<span class="sourceLineNo">175</span><a name="line.175"></a>
+<span class="sourceLineNo">176</span>      String startTimeString = p.getProperty("startTimestamp");<a name="line.176"></a>
+<span class="sourceLineNo">177</span>      if (startTimeString != null) {<a name="line.177"></a>
+<span class="sourceLineNo">178</span>        startTime = Long.parseLong(startTimeString);<a name="line.178"></a>
+<span class="sourceLineNo">179</span>      }<a name="line.179"></a>
+<span class="sourceLineNo">180</span><a name="line.180"></a>
+<span class="sourceLineNo">181</span>      String endTimeString = p.getProperty("endTimestamp");<a name="line.181"></a>
+<span class="sourceLineNo">182</span>      if (endTimeString != null) {<a name="line.182"></a>
+<span class="sourceLineNo">183</span>        endTime = Long.parseLong(endTimeString);<a name="line.183"></a>
+<span class="sourceLineNo">184</span>      }<a name="line.184"></a>
+<span class="sourceLineNo">185</span>    }<a name="line.185"></a>
+<span class="sourceLineNo">186</span><a name="line.186"></a>
+<span class="sourceLineNo">187</span>    Scan initScan() throws IOException {<a name="line.187"></a>
+<span class="sourceLineNo">188</span>      Scan scan = new Scan();<a name="line.188"></a>
+<span class="sourceLineNo">189</span>      scan.setCacheBlocks(false);<a name="line.189"></a>
+<span class="sourceLineNo">190</span>      if (startTime != 0 || endTime != 0) {<a name="line.190"></a>
+<span class="sourceLineNo">191</span>        scan.setTimeRange(startTime, endTime == 0 ? HConstants.LATEST_TIMESTAMP : endTime);<a name="line.191"></a>
+<span class="sourceLineNo">192</span>      }<a name="line.192"></a>
+<span class="sourceLineNo">193</span>      if (scanBatch &gt; 0) {<a name="line.193"></a>
+<span class="sourceLineNo">194</span>        scan.setBatch(scanBatch);<a name="line.194"></a>
+<span class="sourceLineNo">195</span>      }<a name="line.195"></a>
+<span class="sourceLineNo">196</span>      if (versions &gt;= 0) {<a name="line.196"></a>
+<span class="sourceLineNo">197</span>        scan.readVersions(versions);<a name="line.197"></a>
+<span class="sourceLineNo">198</span>      }<a name="line.198"></a>
+<span class="sourceLineNo">199</span>      if (!isTableStartRow(startRow)) {<a name="line.199"></a>
+<span class="sourceLineNo">200</span>        scan.withStartRow(startRow);<a name="line.200"></a>
+<span class="sourceLineNo">201</span>      }<a name="line.201"></a>
+<span class="sourceLineNo">202</span>      if (!isTableEndRow(stopRow)) {<a name="line.202"></a>
+<span class="sourceLineNo">203</span>        scan.withStopRow(stopRow);<a name="line.203"></a>
+<span class="sourceLineNo">204</span>      }<a name="line.204"></a>
+<span class="sourceLineNo">205</span>      if(families != null) {<a name="line.205"></a>
+<span class="sourceLineNo">206</span>        for(String fam : families.split(",")) {<a name="line.206"></a>
+<span class="sourceLineNo">207</span>          scan.addFamily(Bytes.toBytes(fam));<a name="line.207"></a>
+<span class="sourceLineNo">208</span>        }<a name="line.208"></a>
+<span class="sourceLineNo">209</span>      }<a name="line.209"></a>
+<span class="sourceLineNo">210</span>      return scan;<a name="line.210"></a>
+<span class="sourceLineNo">211</span>    }<a name="line.211"></a>
+<span class="sourceLineNo">212</span><a name="line.212"></a>
+<span class="sourceLineNo">213</span>    /**<a name="line.213"></a>
+<span class="sourceLineNo">214</span>     * Choose partitions between row ranges to hash to a single output file<a name="line.214"></a>
+<span class="sourceLineNo">215</span>     * Selects region boundaries that fall within the scan range, and groups them<a name="line.215"></a>
+<span class="sourceLineNo">216</span>     * into the desired number of partitions.<a name="line.216"></a>
+<span class="sourceLineNo">217</span>     */<a name="line.217"></a>
+<span class="sourceLineNo">218</span>    void selectPartitions(Pair&lt;byte[][], byte[][]&gt; regionStartEndKeys) {<a name="line.218"></a>
+<span class="sourceLineNo">219</span>      List&lt;byte[]&gt; startKeys = new ArrayList&lt;&gt;();<a name="line.219"></a>
+<span class="sourceLineNo">220</span>      for (int i = 0; i &lt; regionStartEndKeys.getFirst().length; i++) {<a name="line.220"></a>
+<span class="sourceLineNo">221</span>        byte[] regionStartKey = regionStartEndKeys.getFirst()[i];<a name="line.221"></a>
+<span class="sourceLineNo">222</span>        byte[] regionEndKey = regionStartEndKeys.getSecond()[i];<a name="line.222"></a>
+<span class="sourceLineNo">223</span><a name="line.223"></a>
+<span class="sourceLineNo">224</span>        // if scan begins after this region, or starts before this region, then drop this region<a name="line.224"></a>
+<span class="sourceLineNo">225</span>        // in other words:<a name="line.225"></a>
+<span class="sourceLineNo">226</span>        //   IF (scan begins before the end of this region<a name="line.226"></a>
+<span class="sourceLineNo">227</span>        //      AND scan ends before the start of this region)<a name="line.227"></a>
+<span class="sourceLineNo">228</span>        //   THEN include this region<a name="line.228"></a>
+<span class="sourceLineNo">229</span>        if ((isTableStartRow(startRow) || isTableEndRow(regionEndKey)<a name="line.229"></a>
+<span class="sourceLineNo">230</span>            || Bytes.compareTo(startRow, regionEndKey) &lt; 0)<a name="line.230"></a>
+<span class="sourceLineNo">231</span>          &amp;&amp; (isTableEndRow(stopRow) || isTableStartRow(regionStartKey)<a name="line.231"></a>
+<span class="sourceLineNo">232</span>            || Bytes.compareTo(stopRow, regionStartKey) &gt; 0)) {<a name="line.232"></a>
+<span class="sourceLineNo">233</span>          startKeys.add(regionStartKey);<a name="line.233"></a>
+<span class="sourceLineNo">234</span>        }<a name="line.234"></a>
+<span class="sourceLineNo">235</span>      }<a name="line.235"></a>
+<span class="sourceLineNo">236</span><a name="line.236"></a>
+<span class="sourceLineNo">237</span>      int numRegions = startKeys.size();<a name="line.237"></a>
+<span class="sourceLineNo">238</span>      if (numHashFiles == 0) {<a name="line.238"></a>
+<span class="sourceLineNo">239</span>        numHashFiles = numRegions / 100;<a name="line.239"></a>
+<span class="sourceLineNo">240</span>      }<a name="line.240"></a>
+<span class="sourceLineNo">241</span>      if (numHashFiles == 0) {<a name="line.241"></a>
+<span class="sourceLineNo">242</span>        numHashFiles = 1;<a name="line.242"></a>
+<span class="sourceLineNo">243</span>      }<a name="line.243"></a>
+<span class="sourceLineNo">244</span>      if (numHashFiles &gt; numRegions) {<a name="line.244"></a>
+<span class="sourceLineNo">245</span>        // can't partition within regions<a name="line.245"></a>
+<span class="sourceLineNo">246</span>        numHashFiles = numRegions;<a name="line.246"></a>
+<span class="sourceLineNo">247</span>      }<a name="line.247"></a>
+<span class="sourceLineNo">248</span><a name="line.248"></a>
+<span class="sourceLineNo">249</span>      // choose a subset of start keys to group regions into ranges<a name="line.249"></a>
+<span class="sourceLineNo">250</span>      partitions = new ArrayList&lt;&gt;(numHashFiles - 1);<a name="line.250"></a>
+<span class="sourceLineNo">251</span>      // skip the first start key as it is not a partition between ranges.<a name="line.251"></a>
+<span class="sourceLineNo">252</span>      for (long i = 1; i &lt; numHashFiles; i++) {<a name="line.252"></a>
+<span class="sourceLineNo">253</span>        int splitIndex = (int) (numRegions * i / numHashFiles);<a name="line.253"></a>
+<span class="sourceLineNo">254</span>        partitions.add(new ImmutableBytesWritable(startKeys.get(splitIndex)));<a name="line.254"></a>
+<span class="sourceLineNo">255</span>      }<a name="line.255"></a>
+<span class="sourceLineNo">256</span>    }<a name="line.256"></a>
+<span class="sourceLineNo">257</span><a name="line.257"></a>
+<span class="sourceLineNo">258</span>    void writePartitionFile(Configuration conf, Path path) throws IOException {<a name="line.258"></a>
+<span class="sourceLineNo">259</span>      FileSystem fs = path.getFileSystem(conf);<a name="line.259"></a>
+<span class="sourceLineNo">260</span>      @SuppressWarnings("deprecation")<a name="line.260"></a>
+<span class="sourceLineNo">261</span>      SequenceFile.Writer writer = SequenceFile.createWriter(<a name="line.261"></a>
+<span class="sourceLineNo">262</span>        fs, conf, path, ImmutableBytesWritable.class, NullWritable.class);<a name="line.262"></a>
+<span class="sourceLineNo">263</span><a name="line.263"></a>
+<span class="sourceLineNo">264</span>      for (int i = 0; i &lt; partitions.size(); i++) {<a name="line.264"></a>
+<span class="sourceLineNo">265</span>        writer.append(partitions.get(i), NullWritable.get());<a name="line.265"></a>
+<span class="sourceLineNo">266</span>      }<a name="line.266"></a>
+<span class="sourceLineNo">267</span>      writer.close();<a name="line.267"></a>
+<span class="sourceLineNo">268</span>    }<a name="line.268"></a>
+<span class="sourceLineNo">269</span><a name="line.269"></a>
+<span class="sourceLineNo">270</span>    private void readPartitionFile(FileSystem fs, Configuration conf, Path path)<a name="line.270"></a>
+<span class="sourceLineNo">271</span>         throws IOException {<a name="line.271"></a>
+<span class="sourceLineNo">272</span>      @SuppressWarnings("deprecation")<a name="line.272"></a>
+<span class="sourceLineNo">273</span>      SequenceFile.Reader reader = new SequenceFile.Reader(fs, path, conf);<a name="line.273"></a>
+<span class="sourceLineNo">274</span>      ImmutableBytesWritable key = new ImmutableBytesWritable();<a name="line.274"></a>
+<span class="sourceLineNo">275</span>      partitions = new ArrayList&lt;&gt;();<a name="line.275"></a>
+<span class="sourceLineNo">276</span>      while (reader.next(key)) {<a name="line.276"></a>
+<span class="sourceLineNo">277</span>        partitions.add(new ImmutableBytesWritable(key.copyBytes()));<a name="line.277"></a>
+<span class="sourceLineNo">278</span>      }<a name="line.278"></a>
+<span class="sourceLineNo">279</span>      reader.close();<a name="line.279"></a>
+<span class="sourceLineNo">280</span><a name="line.280"></a>
+<span class="sourceLineNo">281</span>      if (!Ordering.natural().isOrdered(partitions)) {<a name="line.281"></a>
+<span class="sourceLineNo">282</span>        throw new IOException("Partitions are not ordered!");<a name="line.282"></a>
+<span class="sourceLineNo">283</span>      }<a name="line.283"></a>
+<span class="sourceLineNo">284</span>    }<a name="line.284"></a>
+<span class="sourceLineNo">285</span><a name="line.285"></a>
+<span class="sourceLineNo">286</span>    @Override<a name="line.286"></a>
+<span class="sourceLineNo">287</span>    public String toString() {<a name="line.287"></a>
+<span class="sourceLineNo">288</span>      StringBuilder sb = new StringBuilder();<a name="line.288"></a>
+<span class="sourceLineNo">289</span>      sb.append("tableName=").append(tableName);<a name="line.289"></a>
+<span class="sourceLineNo">290</span>      if (families != null) {<a name="line.290"></a>
+<span class="sourceLineNo">291</span>        sb.append(", families=").append(families);<a name="line.291"></a>
+<span class="sourceLineNo">292</span>      }<a name="line.292"></a>
+<span class="sourceLineNo">293</span>      sb.append(", batchSize=").append(batchSize);<a name="line.293"></a>
+<span class="sourceLineNo">294</span>      sb.append(", numHashFiles=").append(numHashFiles);<a name="line.294"></a>
+<span class="sourceLineNo">295</span>      if (!isTableStartRow(startRow)) {<a name="line.295"></a>
+<span class="sourceLineNo">296</span>        sb.append(", startRowHex=").append(Bytes.toHex(startRow));<a name="line.296"></a>
+<span class="sourceLineNo">297</span>      }<a name="line.297"></a>
+<span class="sourceLineNo">298</span>      if (!isTableEndRow(stopRow)) {<a name="line.298"></a>
+<span class="sourceLineNo">299</span>        sb.append(", stopRowHex=").append(Bytes.toHex(stopRow));<a name="line.299"></a>
+<span class="sourceLineNo">300</span>      }<a name="line.300"></a>
+<span class="sourceLineNo">301</span>      if (scanBatch &gt;= 0) {<a name="line.301"></a>
+<span class="sourceLineNo">302</span>        sb.append(", scanBatch=").append(scanBatch);<a name="line.302"></a>
+<span class="sourceLineNo">303</span>      }<a name="line.303"></a>
+<span class="sourceLineNo">304</span>      if (versions &gt;= 0) {<a name="line.304"></a>
+<span class="sourceLineNo">305</span>        sb.append(", versions=").append(versions);<a name="line.305"></a>
+<span class="sourceLineNo">306</span>      }<a name="line.306"></a>
+<span class="sourceLineNo">307</span>      if (startTime != 0) {<a name="line.307"></a>
+<span class="sourceLineNo">308</span>        sb.append("startTime=").append(startTime);<a name="line.308"></a>
+<span class="sourceLineNo">309</span>      }<a name="line.309"></a>
+<span class="sourceLineNo">310</span>      if (endTime != 0) {<a name="line.310"></a>
+<span class="sourceLineNo">311</span>        sb.append("endTime=").append(endTime);<a name="line.311"></a>
+<span class="sourceLineNo">312</span>      }<a name="line.312"></a>
+<span class="sourceLineNo">313</span>      return sb.toString();<a name="line.313"></a>
+<span class="sourceLineNo">314</span>    }<a name="line.314"></a>
+<span class="sourceLineNo">315</span><a name="line.315"></a>
+<span class="sourceLineNo">316</span>    static String getDataFileName(int hashFileIndex) {<a name="line.316"></a>
+<span class="sourceLineNo">317</span>      return String.format(HashTable.OUTPUT_DATA_FILE_PREFIX + "%05d", hashFileIndex);<a name="line.317"></a>
+<span class="sourceLineNo">318</span>    }<a name="line.318"></a>
+<span class="sourceLineNo">319</span><a name="line.319"></a>
+<span class="sourceLineNo">320</span>    /**<a name="line.320"></a>
+<span class="sourceLineNo">321</span>     * Open a TableHash.Reader starting at the first hash at or after the given key.<a name="line.321"></a>
+<span class="sourceLineNo">322</span>     * @throws IOException<a name="line.322"></a>
+<span class="sourceLineNo">323</span>     */<a name="line.323"></a>
+<span class="sourceLineNo">324</span>    public Reader newReader(Configuration conf, ImmutableBytesWritable startKey)<a name="line.324"></a>
+<span class="sourceLineNo">325</span>        throws IOException {<a name="line.325"></a>
+<span class="sourceLineNo">326</span>      return new Reader(conf, startKey);<a name="line.326"></a>
+<span class="sourceLineNo">327</span>    }<a name="line.327"></a>
+<span class="sourceLineNo">328</span><a name="line.328"></a>
+<span class="sourceLineNo">329</span>    public class Reader implements java.io.Closeable {<a name="line.329"></a>
+<span class="sourceLineNo">330</span>      private final Configuration conf;<a name="line.330"></a>
+<span class="sourceLineNo">331</span><a name="line.331"></a>
+<span class="sourceLineNo">332</span>      private int hashFileIndex;<a name="line.332"></a>
+<span class="sourceLineNo">333</span>      private MapFile.Reader mapFileReader;<a name="line.333"></a>
+<span class="sourceLineNo">334</span><a name="line.334"></a>
+<span class="sourceLineNo">335</span>      private boolean cachedNext;<a name="line.335"></a>
+<span class="sourceLineNo">336</span>      private ImmutableBytesWritable key;<a name="line.336"></a>
+<span class="sourceLineNo">337</span>      private ImmutableBytesWritable hash;<a name="line.337"></a>
+<span class="sourceLineNo">338</span><a name="line.338"></a>
+<span class="sourceLineNo">339</span>      Reader(Configuration conf, ImmutableBytesWritable startKey) throws IOException {<a name="line.339"></a>
+<span class="sourceLineNo">340</span>        this.conf = conf;<a name="line.340"></a>
+<span class="sourceLineNo">341</span>        int partitionIndex = Collections.binarySearch(partitions, startKey);<a name="line.341"></a>
+<span class="sourceLineNo">342</span>        if (partitionIndex &gt;= 0) {<a name="line.342"></a>
+<span class="sourceLineNo">343</span>          // if the key is equal to a partition, then go the file after that partition<a name="line.343"></a>
+<span class="sourceLineNo">344</span>          hashFileIndex = partitionIndex+1;<a name="line.344"></a>
+<span class="sourceLineNo">345</span>        } else {<a name="line.345"></a>
+<span class="sourceLineNo">346</span>          // if the key is between partitions, then go to the file between those partitions<a name="line.346"></a>
+<span class="sourceLineNo">347</span>          hashFileIndex = -1-partitionIndex;<a name="line.347"></a>
+<span class="sourceLineNo">348</span>        }<a name="line.348"></a>
+<span class="sourceLineNo">349</span>        openHashFile();<a name="line.349"></a>
+<span class="sourceLineNo">350</span><a name="line.350"></a>
+<span class="sourceLineNo">351</span>        // MapFile's don't make it easy to seek() so that the subsequent next() returns<a name="line.351"></a>
+<span class="sourceLineNo">352</span>        // the desired key/value pair.  So we cache it for the first call of next().<a name="line.352"></a>
+<span class="sourceLineNo">353</span>        hash = new ImmutableBytesWritable();<a name="line.353"></a>
+<span class="sourceLineNo">354</span>        key = (ImmutableBytesWritable) mapFileReader.getClosest(startKey, hash);<a name="line.354"></a>
+<span class="sourceLineNo">355</span>        if (key == null) {<a name="line.355"></a>
+<span class="sourceLineNo">356</span>          cachedNext = false;<a name="line.356"></a>
+<span class="sourceLineNo">357</span>          hash = null;<a name="line.357"></a>
+<span class="sourceLineNo">358</span>        } else {<a name="line.358"></a>
+<span class="sourceLineNo">359</span>          cachedNext = true;<a name="line.359"></a>
+<span class="sourceLineNo">360</span>        }<a name="line.360"></a>
+<span class="sourceLineNo">361</span>      }<a name="line.361"></a>
+<span class="sourceLineNo">362</span><a name="line.362"></a>
+<span class="sourceLineNo">363</span>      /**<a name="line.363"></a>
+<span class="sourceLineNo">364</span>       * Read the next key/hash pair.<a name="line.364"></a>
+<span class="sourceLineNo">365</span>       * Returns true if such a pair exists and false when at the end of the data.<a name="line.365"></a>
+<span class="sourceLineNo">366</span>       */<a name="line.366"></a>
+<span class="sourceLineNo">367</span>      public boolean next() throws IOException {<a name="line.367"></a>
+<span class="sourceLineNo">368</span>        if (cachedNext) {<a name="line.368"></a>
+<span class="sourceLineNo">369</span>          cachedNext = false;<a name="line.369"></a>
+<span class="sourceLineNo">370</span>          return true;<a name="line.370"></a>
+<span class="sourceLineNo">371</span>        }<a name="line.371"></a>
+<span class="sourceLineNo">372</span>        key = new ImmutableBytesWritable();<a name="line.372"></a>
+<span class="sourceLineNo">373</span>        hash = new ImmutableBytesWritable();<a name="line.373"></a>
+<span class="sourceLineNo">374</span>        while (true) {<a name="line.374"></a>
+<span class="sourceLineNo">375</span>          boolean hasNext = mapFileReader.next(key, hash);<a name="line.375"></a>
+<span class="sourceLineNo">376</span>          if (hasNext) {<a name="line.376"></a>
+<span class="sourceLineNo">377</span>            return true;<a name="line.377"></a>
+<span class="sourceLineNo">378</span>          }<a name="line.378"></a>
+<span class="sourceLineNo">379</span>          hashFileIndex++;<a name="line.379"></a>
+<span class="sourceLineNo">380</span>          if (hashFileIndex &lt; TableHash.this.numHashFiles) {<a name="line.380"></a>
+<span class="sourceLineNo">381</span>            mapFileReader.close();<a name="line.381"></a>
+<span class="sourceLineNo">382</span>            openHashFile();<a name="line.382"></a>
+<span class="sourceLineNo">383</span>          } else {<a name="line.383"></a>
+<span class="sourceLineNo">384</span>            key = null;<a name="line.384"></a>
+<span class="sourceLineNo">385</span>            hash = null;<a name="line.385"></a>
+<span class="sourceLineNo">386</span>            return false;<a name="line.386"></a>
+<span class="sourceLineNo">387</span>          }<a name="line.387"></a>
+<span class="sourceLineNo">388</span>        }<a name="line.388"></a>
+<span class="sourceLineNo">389</span>      }<a name="line.389"></a>
+<span class="sourceLineNo">390</span><a name="line.390"></a>
+<span class="sourceLineNo">391</span>      /**<a name="line.391"></a>
+<span class="sourceLineNo">392</span>       * Get the current key<a name="line.392"></a>
+<span class="sourceLineNo">393</span>       * @return the current key or null if there is no current key<a name="line.393"></a>
+<span class="sourceLineNo">394</span>       */<a name="line.394"></a>
+<span class="sourceLineNo">395</span>      public ImmutableBytesWritable getCurrentKey() {<a name="line.395"></a>
+<span class="sourceLineNo">396</span>        return key;<a name="line.396"></a>
+<span class="sourceLineNo">397</span>      }<a name="line.397"></a>
+<span class="sourceLineNo">398</span><a name="line.398"></a>
+<span class="sourceLineNo">399</span>      /**<a name="line.399"></a>
+<span class="sourceLineNo">400</span>       * Get the current hash<a name="line.400"></a>
+<span class="sourceLineNo">401</span>       * @return the current hash or null if there is no current hash<a name="line.401"></a>
+<span class="sourceLineNo">402</span>       */<a name="line.402"></a>
+<span class="sourceLineNo">403</span>      public ImmutableBytesWritable getCurrentHash() {<a name="line.403"></a>
+<span class="sourceLineNo">404</span>        return hash;<a name="line.404"></a>
+<span class="sourceLineNo">405</span>      }<a name="line.405"></a>
+<span class="sourceLineNo">406</span><a name="line.406"></a>
+<span class="sourceLineNo">407</span>      private void openHashFile() throws IOException {<a name="line.407"></a>
+<span class="sourceLineNo">408</span>        if (mapFileReader != null) {<a name="line.408"></a>
+<span class="sourceLineNo">409</span>          mapFileReader.close();<a name="line.409"></a>
+<span class="sourceLineNo">410</span>        }<a name="line.410"></a>
+<span class="sourceLineNo">411</span>        Path dataDir = new Path(TableHash.this.hashDir, HASH_DATA_DIR);<a name="line.411"></a>
+<span class="sourceLineNo">412</span>        Path dataFile = new Path(dataDir, getDataFileName(hashFileIndex));<a name="line.412"></a>
+<span class="sourceLineNo">413</span>        mapFileReader = new MapFile.Reader(dataFile, conf);<a name="line.413"></a>
+<span class="sourceLineNo">414</span>      }<a name="line.414"></a>
+<span class="sourceLineNo">415</span><a name="line.415"></a>
+<span class="sourceLineNo">416</span>      @Override<a name="line.416"></a>
+<span class="sourceLineNo">417</span>      public void close() throws IOException {<a name="line.417"></a>
+<span class="sourceLineNo">418</span>        mapFileReader.close();<a name="line.418"></a>
+<span class="sourceLineNo">419</span>      }<a name="line.419"></a>
+<span class="sourceLineNo">420</span>    }<a name="line.420"></a>
+<span class="sourceLineNo">421</span>  }<a name="line.421"></a>
+<span class="sourceLineNo">422</span><a name="line.422"></a>
+<span class="sourceLineNo">423</span>  static boolean isTableStartRow(byte[] row) {<a name="line.423"></a>
+<span class="sourceLineNo">424</span>    return Bytes.equals(HConstants.EMPTY_START_ROW, row);<a name="line.424"></a>
+<span class="sourceLineNo">425</span>  }<a name="line.425"></a>
+<span class="sourceLineNo">426</span><a name="line.426"></a>
+<span class="sourceLineNo">427</span>  static boolean isTableEndRow(byte[] row) {<a name="line.427"></a>
+<span class="sourceLineNo">428</span>    return Bytes.equals(HConstants.EMPTY_END_ROW, row);<a name="line.428"></a>
+<span class="sourceLineNo">429</span>  }<a name="line.429"></a>
+<span class="sourceLineNo">430</span><a name="line.430"></a>
+<span class="sourceLineNo">431</span>  public Job createSubmittableJob(String[] args) throws IOException {<a name="line.431"></a>
+<span class="sourceLineNo">432</span>    Path partitionsPath = new Path(destPath, PARTITIONS_FILE_NAME);<a name="line.432"></a>
+<span class="sourceLineNo">433</span>    generatePartitions(partitionsPath);<a name="line.433"></a>
+<span class="sourceLineNo">434</span><a name="line.434"></a>
+<span class="sourceLineNo">435</span>    Job job = Job.getInstance(getConf(),<a name="line.435"></a>
+<span class="sourceLineNo">436</span>          getConf().get("mapreduce.job.name", "hashTable_" + tableHash.tableName));<a name="line.436"></a>
+<span class="sourceLineNo">437</span>    Configuration jobConf = job.getConfiguration();<a name="line.437"></a>
+<span class="sourceLineNo">438</span>    jobConf.setLong(HASH_BATCH_SIZE_CONF_KEY, tableHash.batchSize);<a name="line.438"></a>
+<span class="sourceLineNo">439</span>    jobConf.setBoolean(IGNORE_TIMESTAMPS, tableHash.ignoreTimestamps);<a name="line.439"></a>
+<span class="sourceLineNo">440</span>    job.setJarByClass(HashTable.class);<a name="line.440"></a>
 <span class="sourceLineNo">441</span><a name="line.441"></a>
-<span class="sourceLineNo">442</span>    // use a TotalOrderPartitioner and reducers to group region output into hash files<a name="line.442"></a>
-<span class="sourceLineNo">443</span>    job.setPartitionerClass(TotalOrderPartitioner.class);<a name="line.443"></a>
-<span class="sourceLineNo">444</span>    TotalOrderPartitioner.setPartitionFile(jobConf, partitionsPath);<a name="line.444"></a>
-<span class="sourceLineNo">445</span>    job.setReducerClass(Reducer.class);  // identity reducer<a name="line.445"></a>
-<span class="sourceLineNo">446</span>    job.setNumReduceTasks(tableHash.numHashFiles);<a name="line.446"></a>
-<span class="sourceLineNo">447</span>    job.setOutputKeyClass(ImmutableBytesWritable.class);<a name="line.447"></a>
-<span class="sourceLineNo">448</span>    job.setOutputValueClass(ImmutableBytesWritable.class);<a name="line.448"></a>
-<span class="sourceLineNo">449</span>    job.setOutputFormatClass(MapFileOutputFormat.class);<a name="line.449"></a>
-<span class="sourceLineNo">450</span>    FileOutputFormat.setOutputPath(job, new Path(destPath, HASH_DATA_DIR));<a name="line.450"></a>
-<span class="sourceLineNo">451</span><a name="line.451"></a>
-<span class="sourceLineNo">452</span>    return job;<a name="line.452"></a>
-<span class="sourceLineNo">453</span>  }<a name="line.453"></a>
+<span class="sourceLineNo">442</span>    TableMapReduceUtil.initTableMapperJob(tableHash.tableName, tableHash.initScan(),<a name="line.442"></a>
+<span class="sourceLineNo">443</span>        HashMapper.class, ImmutableBytesWritable.class, ImmutableBytesWritable.class, job);<a name="line.443"></a>
+<span class="sourceLineNo">444</span><a name="line.444"></a>
+<span class="sourceLineNo">445</span>    // use a TotalOrderPartitioner and reducers to group region output into hash files<a name="line.445"></a>
+<span class="sourceLineNo">446</span>    job.setPartitionerClass(TotalOrderPartitioner.class);<a name="line.446"></a>
+<span class="sourceLineNo">447</span>    TotalOrderPartitioner.setPartitionFile(jobConf, partitionsPath);<a name="line.447"></a>
+<span class="sourceLineNo">448</span>    job.setReducerClass(Reducer.class);  // identity reducer<a name="line.448"></a>
+<span class="sourceLineNo">449</span>    job.setNumReduceTasks(tableHash.numHashFiles);<a name="line.449"></a>
+<span class="sourceLineNo">450</span>    job.setOutputKeyClass(ImmutableBytesWritable.class);<a name="line.450"></a>
+<span class="sourceLineNo">451</span>    job.setOutputValueClass(ImmutableBytesWritable.class);<a name="line.451"></a>
+<span class="sourceLineNo">452</span>    job.setOutputFormatClass(MapFileOutputFormat.class);<a name="line.452"></a>
+<span class="sourceLineNo">453</span>    FileOutputFormat.setOutputPath(job, new Path(destPath, HASH_DATA_DIR));<a name="line.453"></a>
 <span class="sourceLineNo">454</span><a name="line.454"></a>
-<span class="sourceLineNo">455</span>  private void generatePartitions(Path partitionsPath) throws IOException {<a name="line.455"></a>
-<span class="sourceLineNo">456</span>    Connection connection = ConnectionFactory.createConnection(getConf());<a name="line.456"></a>
-<span class="sourceLineNo">457</span>    Pair&lt;byte[][], byte[][]&gt; regionKeys<a name="line.457"></a>
-<span class="sourceLineNo">458</span>      = connection.getRegionLocator(TableName.valueOf(tableHash.tableName)).getStartEndKeys();<a name="line.458"></a>
-<span class="sourceLineNo">459</span>    connection.close();<a name="line.459"></a>
-<span class="sourceLineNo">460</span><a name="line.460"></a>
-<span class="sourceLineNo">461</span>    tableHash.selectPartitions(regionKeys);<a name="line.461"></a>
-<span class="sourceLineNo">462</span>    LOG.info("Writing " + tableHash.partitions.size() + " partition keys to " + partitionsPath);<a name="line.462"></a>
+<span class="sourceLineNo">455</span>    return job;<a name="line.455"></a>
+<span class="sourceLineNo">456</span>  }<a name="line.456"></a>
+<span class="sourceLineNo">457</span><a name="line.457"></a>
+<span class="sourceLineNo">458</span>  private void generatePartitions(Path partitionsPath) throws IOException {<a name="line.458"></a>
+<span class="sourceLineNo">459</span>    Connection connection = ConnectionFactory.createConnection(getConf());<a name="line.459"></a>
+<span class="sourceLineNo">460</span>    Pair&lt;byte[][], byte[][]&gt; regionKeys<a name="line.460"></a>
+<span class="sourceLineNo">461</span>      = connection.getRegionLocator(TableName.valueOf(tableHash.tableName)).getStartEndKeys();<a name="line.461"></a>
+<span class="sourceLineNo">462</span>    connection.close();<a name="line.462"></a>
 <span class="sourceLineNo">463</span><a name="line.463"></a>
-<span class="sourceLineNo">464</span>    tableHash.writePartitionFile(getConf(), partitionsPath);<a name="line.464"></a>
-<span class="sourceLineNo">465</span>  }<a name="line.465"></a>
+<span class="sourceLineNo">464</span>    tableHash.selectPartitions(regionKeys);<a name="line.464"></a>
+<span class="sourceLineNo">465</span>    LOG.info("Writing " + tableHash.partitions.size() + " partition keys to " + partitionsPath);<a name="line.465"></a>
 <span class="sourceLineNo">466</span><a name="line.466"></a>
-<span class="sourceLineNo">467</span>  static class ResultHasher {<a name="line.467"></a>
-<span class="sourceLineNo">468</span>    private MessageDigest digest;<a name="line.468"></a>
+<span class="sourceLineNo">467</span>    tableHash.writePartitionFile(getConf(), partitionsPath);<a name="line.467"></a>
+<span class="sourceLineNo">468</span>  }<a name="line.468"></a>
 <span class="sourceLineNo">469</span><a name="line.469"></a>
-<span class="sourceLineNo">470</span>    private boolean batchStarted = false;<a name="line.470"></a>
-<span class="sourceLineNo">471</span>    private ImmutableBytesWritable batchStartKey;<a name="line.471"></a>
-<span class="sourceLineNo">472</span>    private ImmutableBytesWritable batchHash;<a name="line.472"></a>
-<span class="sourceLineNo">473</span>    private long batchSize = 0;<a name="line.473"></a>
-<span class="sourceLineNo">474</span><a name="line.474"></a>
-<span class="sourceLineNo">475</span><a name="line.475"></a>
-<span class="sourceLineNo">476</span>    public ResultHasher() {<a name="line.476"></a>
-<span class="sourceLineNo">477</span>      try {<a name="line.477"></a>
-<span class="sourceLineNo">478</span>        digest = MessageDigest.getInstance("MD5");<a name="line.478"></a>
-<span class="sourceLineNo">479</span>      } catch (NoSuchAlgorithmException e) {<a name="line.479"></a>
-<span class="sourceLineNo">480</span>        Throwables.propagate(e);<a name="line.480"></a>
-<span class="sourceLineNo">481</span>      }<a name="line.481"></a>
-<span class="sourceLineNo">482</span>    }<a name="line.482"></a>
-<span class="sourceLineNo">483</span><a name="line.483"></a>
-<span class="sourceLineNo">484</span>    public void startBatch(ImmutableBytesWritable row) {<a name="line.484"></a>
-<span class="sourceLineNo">485</span>      if (batchStarted) {<a name="line.485"></a>
-<span class="sourceLineNo">486</span>        throw new RuntimeException("Cannot start new batch without finishing existing one.");<a name="line.486"></a>
-<span class="sourceLineNo">487</span>      }<a name="line.487"></a>
-<span class="sourceLineNo">488</span>      batchStarted = true;<a name="line.488"></a>
-<span class="sourceLineNo">489</span>      batchSize = 0;<a name="line.489"></a>
-<span class="sourceLineNo">490</span>      batchStartKey = row;<a name="line.490"></a>
-<span class="sourceLineNo">491</span>      batchHash = null;<a name="line.491"></a>
-<span class="sourceLineNo">492</span>    }<a name="line.492"></a>
-<span class="sourceLineNo">493</span><a name="line.493"></a>
-<span class="sourceLineNo">494</span>    public void hashResult(Result result) {<a name="line.494"></a>
-<span class="sourceLineNo">495</span>      if (!batchStarted) {<a name="line.495"></a>
-<span class="sourceLineNo">496</span>        throw new RuntimeException("Cannot add to batch that has not been started.");<a name="line.496"></a>
-<span class="sourceLineNo">497</span>      }<a name="line.497"></a>
-<span class="sourceLineNo">498</span>      for (Cell cell : result.rawCells()) {<a name="line.498"></a>
-<span class="sourceLineNo">499</span>        int rowLength = cell.getRowLength();<a name="line.499"></a>
-<span class="sourceLineNo">500</span>        int familyLength = cell.getFamilyLength();<a name="line.500"></a>
-<span class="sourceLineNo">501</span>        int qualifierLength = cell.getQualifierLength();<a name="line.501"></a>
-<span class="sourceLineNo">502</span>        int valueLength = cell.getValueLength();<a name="line.502"></a>
-<span class="sourceLineNo">503</span>        digest.update(cell.getRowArray(), cell.getRowOffset(), rowLength);<a name="line.503"></a>
-<span class="sourceLineNo">504</span>        digest.update(cell.getFamilyArray(), cell.getFamilyOffset(), familyLength);<a name="line.504"></a>
-<span class="sourceLineNo">505</span>        digest.update(cell.getQualifierArray(), cell.getQualifierOffset(), qualifierLength);<a name="line.505"></a>
-<span class="sourceLineNo">506</span>        long ts = cell.getTimestamp();<a name="line.506"></a>
-<span class="sourceLineNo">507</span>        for (int i = 8; i &gt; 0; i--) {<a name="line.507"></a>
-<span class="sourceLineNo">508</span>          digest.update((byte) ts);<a name="line.508"></a>
-<span class="sourceLineNo">509</span>          ts &gt;&gt;&gt;= 8;<a name="line.509"></a>
-<span class="sourceLineNo">510</span>        }<a name="line.510"></a>
-<span class="sourceLineNo">511</span>        digest.update(cell.getValueArray(), cell.getValueOffset(), valueLength);<a name="line.511"></a>
-<span class="sourceLineNo">512</span><a name="line.512"></a>
-<span class="sourceLineNo">513</span>        batchSize += rowLength + familyLength + qualifierLength + 8 + valueLength;<a name="line.513"></a>
-<span class="sourceLineNo">514</span>      }<a name="line.514"></a>
-<span class="sourceLineNo">515</span>    }<a name="line.515"></a>
-<span class="sourceLineNo">516</span><a name="line.516"></a>
-<span class="sourceLineNo">517</span>    public void finishBatch() {<a name="line.517"></a>
-<span class="sourceLineNo">518</span>      if (!batchStarted) {<a name="line.518"></a>
-<span class="sourceLineNo">519</span>        throw new RuntimeException("Cannot finish batch that has not started.");<a name="line.519"></a>
-<span class="sourceLineNo">520</span>      }<a name="line.520"></a>
-<span class="sourceLineNo">521</span>      batchStarted = false;<a name="line.521"></a>
-<span class="sourceLineNo">522</span>      batchHash = new ImmutableBytesWritable(digest.digest());<a name="line.522"></a>
-<span class="sourceLineNo">523</span>    }<a name="line.523"></a>
-<span class="sourceLineNo">524</span><a name="line.524"></a>
-<span class="sourceLineNo">525</span>    public boolean isBatchStarted() {<a name="line.525"></a>
-<span class="sourceLineNo">526</span>      return batchStarted;<a name="line.526"></a>
-<span class="sourceLineNo">527</span>    }<a name="line.527"></a>
-<span class="sourceLineNo">528</span><a name="line.528"></a>
-<span class="sourceLineNo">529</span>    public ImmutableBytesWritable getBatchStartKey() {<a name="line.529"></a>
-<span class="sourceLineNo">530</span>      return batchStartKey;<a name="line.530"></a>
-<span class="sourceLineNo">531</span>    }<a name="line.531"></a>
-<span class="sourceLineNo">532</span><a name="line.532"></a>
-<span class="sourceLineNo">533</span>    public ImmutableBytesWritable getBatchHash() {<a name="line.533"></a>
-<span class="sourceLineNo">534</span>      return batchHash;<a name="line.534"></a>
-<span class="sourceLineNo">535</span>    }<a name="line.535"></a>
-<span class="sourceLineNo">536</span><a name="line.536"></a>
-<span class="sourceLineNo">537</span>    public long getBatchSize() {<a name="line.537"></a>
-<span class="sourceLineNo">538</span>      return batchSize;<a name="line.538"></a>
-<span class="sourceLineNo">539</span>    }<a name="line.539"></a>
-<span class="sourceLineNo">540</span>  }<a name="line.540"></a>
-<span class="sourceLineNo">541</span><a name="line.541"></a>
-<span class="sourceLineNo">542</span>  public static class HashMapper<a name="line.542"></a>
-<span class="sourceLineNo">543</span>    extends TableMapper&lt;ImmutableBytesWritable, ImmutableBytesWritable&gt; {<a name="line.543"></a>
-<span class="sourceLineNo">544</span><a name="line.544"></a>
-<span class="sourceLineNo">545</span>    private ResultHasher hasher;<a name="line.545"></a>
-<span class="sourceLineNo">546</span>    private long targetBatchSize;<a name="line.546"></a>
-<span class="sourceLineNo">547</span><a name="line.547"></a>
-<span class="sourceLineNo">548</span>    private ImmutableBytesWritable currentRow;<a name="line.548"></a>
-<span class="sourceLineNo">549</span><a name="line.549"></a>
-<span class="sourceLineNo">550</span>    @Override<a name="line.550"></a>
-<span class="sourceLineNo">551</span>    protected void setup(Context context) throws IOException, InterruptedException {<a name="line.551"></a>
-<span class="sourceLineNo">552</span>      targetBatchSize = context.getConfiguration()<a name="line.552"></a>
-<span class="sourceLineNo">553</span>          .getLong(HASH_BATCH_SIZE_CONF_KEY, DEFAULT_BATCH_SIZE);<a name="line.553"></a>
-<span class="sourceLineNo">554</span>      hasher = new ResultHasher();<a name="line.554"></a>
-<span class="sourceLineNo">555</span><a name="line.555"></a>
-<span class="sourceLineNo">556</span>      TableSplit split = (TableSplit) context.getInputSplit();<a name="line.556"></a>
-<span class="sourceLineNo">557</span>      hasher.startBatch(new ImmutableBytesWritable(split.getStartRow()));<a name="line.557"></a>
-<span class="sourceLineNo">558</span>    }<a name="line.558"></a>
-<span class="sourceLineNo">559</span><a name="line.559"></a>
-<span class="sourceLineNo">560</span>    @Override<a name="line.560"></a>
-<span class="sourceLineNo">561</span>    protected void map(ImmutableBytesWritable key, Result value, Context context)<a name="line.561"></a>
-<span class="sourceLineNo">562</span>        throws IOException, InterruptedException {<a name="line.562"></a>
-<span class="sourceLineNo">563</span><a name="line.563"></a>
-<span class="sourceLineNo">564</span>      if (currentRow == null || !currentRow.equals(key)) {<a name="line.564"></a>
-<span class="sourceLineNo">565</span>        currentRow = new ImmutableBytesWritable(key); // not immutable<a name="line.565"></a>
-<span class="sourceLineNo">566</span><a name="line.566"></a>
-<span class="sourceLineNo">567</span>        if (hasher.getBatchSize() &gt;= targetBatchSize) {<a name="line.567"></a>
-<span class="sourceLineNo">568</span>          hasher.finishBatch();<a name="line.568"></a>
-<span class="sourceLineNo">569</span>          context.write(hasher.getBatchStartKey(), hasher.getBatchHash());<a name="line.569"></a>
-<span class="sourceLineNo">570</span>          hasher.startBatch(currentRow);<a name="line.570"></a>
-<span class="sourceLineNo">571</span>        }<a name="line.571"></a>
-<span class="sourceLineNo">572</span>      }<a name="line.572"></a>
-<span class="sourceLineNo">573</span><a name="line.573"></a>
-<span class="sourceLineNo">574</span>      hasher.hashResult(value);<a name="line.574"></a>
-<span class="sourceLineNo">575</span>    }<a name="line.575"></a>
-<span class="sourceLineNo">576</span><a name="line.576"></a>
-<span class="sourceLineNo">577</span>    @Override<a name="line.577"></a>
-<span class="sourceLineNo">578</span>    protected void cleanup(Context context) throws IOException, InterruptedException {<a name="line.578"></a>
-<span class="sourceLineNo">579</span>      hasher.finishBatch();<a name="line.579"></a>
-<span class="sourceLineNo">580</span>      context.write(hasher.getBatchStartKey(), hasher.getBatchHash());<a name="line.580"></a>
-<span class="sourceLineNo">581</span>    }<a name="line.581"></a>
-<span class="sourceLineNo">582</span>  }<a name="line.582"></a>
-<span class="sourceLineNo">583</span><a name="line.583"></a>
-<span class="sourceLineNo">584</span>  private void writeTempManifestFile() throws IOException {<a name="line.584"></a>
-<span class="sourceLineNo">585</span>    Path tempManifestPath = new Path(destPath, TMP_MANIFEST_FILE_NAME);<a name="line.585"></a>
-<span class="sourceLineNo">586</span>    FileSystem fs = tempManifestPath.getFileSystem(getConf());<a name="line.586"></a>
-<span class="sourceLineNo">587</span>    tableHash.writePropertiesFile(fs, tempManifestPath);<a name="line.587"></a>
-<span class="sourceLineNo">588</span>  }<a name="line.588"></a>
-<span class="sourceLineNo">589</span><a name="line.589"></a>
-<span class="sourceLineNo">590</span>  private void completeManifest() throws IOException {<a name="line.590"></a>
-<span class="sourceLineNo">591</span>    Path tempManifestPath = new Path(destPath, TMP_MANIFEST_FILE_NAME);<a name="line.591"></a>
-<span class="sourceLineNo">592</span>    Path manifestPath = new Path(destPath, MANIFEST_FILE_NAME);<a name="line.592"></a>
-<span class="sourceLineNo">593</span>    FileSystem fs = tempManifestPath.getFileSystem(getConf());<a name="line.593"></a>
-<span class="sourceLineNo">594</span>    fs.rename(tempManifestPath, manifestPath);<a name="line.594"></a>
-<span class="sourceLineNo">595</span>  }<a name="line.595"></a>
-<span class="sourceLineNo">596</span><a name="line.596"></a>
-<span class="sourceLineNo">597</span>  private static final int NUM_ARGS = 2;<a name="line.597"></a>
-<span class="sourceLineNo">598</span>  private static void printUsage(final String errorMsg) {<a name="line.598"></a>
-<span class="sourceLineNo">599</span>    if (errorMsg != null &amp;&amp; errorMsg.length() &gt; 0) {<a name="line.599"></a>
-<span class="sourceLineNo">600</span>      System.err.println("ERROR: " + errorMsg);<a name="line.600"></a>
-<span class="sourceLineNo">601</span>      System.err.println();<a name="line.601"></a>
-<span class="sourceLineNo">602</span>    }<a name="line.602"></a>
-<span class="sourceLineNo">603</span>    System.err.println("Usage: HashTable [options] &lt;tablename&gt; &lt;outputpath&gt;");<a name="line.603"></a>
-<span class="sourceLineNo">604</span>    System.err.println();<a name="line.604"></a>
-<span class="sourceLineNo">605</span>    System.err.println("Options:");<a name="line.605"></a>
-<span class="sourceLineNo">606</span>    System.err.println(" batchsize     the target amount of bytes to hash in each batch");<a name="line.606"></a>
-<span class="sourceLineNo">607</span>    System.err.println("               rows are added to the batch until this size is reached");<a name="line.607"></a>
-<span class="sourceLineNo">608</span>    System.err.println("               (defaults to " + DEFAULT_BATCH_SIZE + " bytes)");<a name="line.608"></a>
-<span class="sourceLineNo">609</span>    System.err.println(" numhashfiles  the number of hash files to create");<a name="line.609"></a>
-<span class="sourceLineNo">610</span>    System.err.println("               if set to fewer than number of regions then");<a name="line.610"></a>
-<span class="sourceLineNo">611</span>    System.err.println("               the job will create this number of reducers");<a name="line.611"></a>
-<span class="sourceLineNo">612</span>    System.err.println("               (defaults to 1/100 of regions -- at least 1)");<a name="line.612"></a>
-<span class="sourceLineNo">613</span>    System.err.println(" startrow      the start row");<a name="line.613"></a>
-<span class="sourceLineNo">614</span>    System.err.println(" stoprow       the stop row");<a name="line.614"></a>
-<span class="sourceLineNo">615</span>    System.err.println(" starttime     beginning of the time range (unixtime in millis)");<a name="line.615"></a>
-<span class="sourceLineNo">616</span>    System.err.println("               without endtime means from starttime to forever");<a name="line.616"></a>
-<span class="sourceLineNo">617</span>    System.err.println(" endtime       end of the time range.  Ignored if no starttime specified.");<a name="line.617"></a>
-<span class="sourceLineNo">618</span>    System.err.println(" scanbatch     scanner batch size to support intra row scans");<a name="line.618"></a>
-<span class="sourceLineNo">619</span>    System.err.println(" versions      number of cell versions to include");<a name="line.619"></a>
-<span class="sourceLineNo">620</span>    System.err.println(" families      comma-separated list of families to include");<a name="line.620"></a>
-<span class="sourceLineNo">621</span>    System.err.println();<a name="line.621"></a>
-<span class="sourceLineNo">622</span>    System.err.println("Args:");<a name="line.622"></a>
-<span class="sourceLineNo">623</span>    System.err.println(" tablename     Name of the table to hash");<a name="line.623"></a>
-<span class="sourceLineNo">624</span>    System.err.println(" outputpath    Filesystem path to put the output data");<a name="line.624"></a>
-<span class="sourceLineNo">625</span>    System.err.println();<a name="line.625"></a>
-<span class="sourceLineNo">626</span>    System.err.println("Examples:");<a name="line.626"></a>
-<span class="sourceLineNo">627</span>    System.err.println(" To hash 'TestTable' in 32kB batches for a 1 hour window into 50 files:");<a name="line.627"></a>
-<span class="sourceLineNo">628</span>    System.err.println(" $ hbase " +<a name="line.628"></a>
-<span class="sourceLineNo">629</span>        "org.apache.hadoop.hbase.mapreduce.HashTable --batchsize=32000 --numhashfiles=50"<a name="line.629"></a>
-<span class="sourceLineNo">630</span>        + " --starttime=1265875194289 --endtime=1265878794289 --families=cf2,cf3"<a name="line.630"></a>
-<span class="sourceLineNo">631</span>        + " TestTable /hashes/testTable");<a name="line.631"></a>
-<span class="sourceLineNo">632</span>  }<a name="line.632"></a>
-<span class="sourceLineNo">633</span><a name="line.633"></a>
-<span class="sourceLineNo">634</span>  private boolean doCommandLine(final String[] args) {<a name="line.634"></a>
-<span class="sourceLineNo">635</span>    if (args.length &lt; NUM_ARGS) {<a name="line.635"></a>
-<span class="sourceLineNo">636</span>      printUsage(null);<a name="line.636"></a>
-<span class="sourceLineNo">637</span>      return false;<a name="line.637"></a>
-<span class="sourceLineNo">638</span>    }<a name="line.638"></a>
-<span class="sourceLineNo">639</span>    try {<a name="line.639"></a>
-<span class="sourceLineNo">640</span><a name="line.640"></a>
-<span class="sourceLineNo">641</span>      tableHash.tableName = args[args.length-2];<a name="line.641"></a>
-<span class="sourceLineNo">642</span>      destPath = new Path(args[args.length-1]);<a name="line.642"></a>
-<span class="sourceLineNo">643</span><a name="line.643"></a>
-<span class="sourceLineNo">644</span>      for (int i = 0; i &lt; args.length - NUM_ARGS; i++) {<a name="line.644"></a>
-<span class="sourceLineNo">645</span>        String cmd = args[i];<a name="line.645"></a>
-<span class="sourceLineNo">646</span>        if (cmd.equals("-h") || cmd.startsWith("--h")) {<a name="line.646"></a>
-<span class="sourceLineNo">647</span>          printUsage(null);<a name="line.647"></a>
-<span class="sourceLineNo">648</span>          return false;<a name="line.648"></a>
-<span class="sourceLineNo">649</span>        }<a name="line.649"></a>
-<span class="sourceLineNo">650</span><a name="line.650"></a>
-<span class="sourceLineNo">651</span>        final String batchSizeArgKey = "--batchsize=";<a name="line.651"></a>
-<span class="sourceLineNo">652</span>        if (cmd.startsWith(batchSizeArgKey)) {<a name="line.652"></a>
-<span class="sourceLineNo">653</span>          tableHash.batchSize = Long.parseLong(cmd.substring(batchSizeArgKey.length()));<a name="line.653"></a>
-<span class="sourceLineNo">654</span>          continue;<a name="line.654"></a>
-<span class="sourceLineNo">655</span>        }<a name="line.655"></a>
-<span class="sourceLineNo">656</span><a name="line.656"></a>
-<span class="sourceLineNo">657</span>        final String numHashFilesArgKey = "--numhashfiles=";<a name="line.657"></a>
-<span class="sourceLineNo">658</span>        if (cmd.startsWith(numHashFilesArgKey)) {<a name="line.658"></a>
-<span class="sourceLineNo">659</span>          tableHash.numHashFiles = Integer.parseInt(cmd.substring(numHashFilesArgKey.length()));<a name="line.659"></a>
-<span class="sourceLineNo">660</span>          continue;<a name="line.660"></a>
-<span class="sourceLineNo">661</span>        }<a name="line.661"></a>
-<span class="sourceLineNo">662</span><a name="line.662"></a>
-<span class="sourceLineNo">663</span>        final String startRowArgKey = "--startrow=";<a name="line.663"></a>
-<span class="sourceLineNo">664</span>        if (cmd.startsWith(startRowArgKey)) {<a name="line.664"></a>
-<span class="sourceLineNo">665</span>          tableHash.startRow = Bytes.fromHex(cmd.substring(startRowArgKey.length()));<a name="line.665"></a>
-<span class="sourceLineNo">666</span>          continue;<a name="line.666"></a>
-<span class="sourceLineNo">667</span>        }<a name="line.667"></a>
-<span class="sourceLineNo">668</span><a name="line.668"></a>
-<span class="sourceLineNo">669</span>        final String stopRowArgKey = "--stoprow=";<a name="line.669"></a>
-<span class="sourceLineNo">670</span>        if (cmd.startsWith(stopRowArgKey)) {<a name="line.670"></a>
-<span class="sourceLineNo">671</span>          tableHash.stopRow = Bytes.fromHex(cmd.substring(stopRowArgKey.length()));<a name="line.671"></a>
-<span class="sourceLineNo">672</span>          continue;<a name="line.672"></a>
-<span class="sourceLineNo">673</span>        }<a name="line.673"></a>
-<span class="sourceLineNo">674</span><a name="line.674"></a>
-<span class="sourceLineNo">675</span>        final String startTimeArgKey = "--starttime=";<a name="line.675"></a>
-<span class="sourceLineNo">676</span>        if (cmd.startsWith(startTimeArgKey)) {<a name="line.676"></a>
-<span class="sourceLineNo">677</span>          tableHash.startTime = Long.parseLong(cmd.substring(startTimeArgKey.length()));<a name="line.677"></a>
-<span class="sourceLineNo">678</span>          continue;<a name="line.678"></a>
-<span class="sourceLineNo">679</span>        }<a name="line.679"></a>
-<span class="sourceLineNo">680</span><a name="line.680"></a>
-<span class="sourceLineNo">681</span>        final String endTimeArgKey = "--endtime=";<a name="line.681"></a>
-<span class="sourceLineNo">682</span>        if (cmd.startsWith(endTimeArgKey)) {<a name="line.682"></a>
-<span class="sourceLineNo">683</span>          tableHash.endTime = Long.parseLong(cmd.substring(endTimeArgKey.length()));<a name="line.683"></a>
-<span class="sourceLineNo">684</span>          continue;<a name="line.684"></a>
-<span class="sourceLineNo">685</span>        }<a name="line.685"></a>
-<span class="sourceLineNo">686</span><a name="line.686"></a>
-<span class="sourceLineNo">687</span>        final String scanBatchArgKey = "--scanbatch=";<a name="line.687"></a>
-<span class="sourceLineNo">688</span>        if (cmd.startsWith(scanBatchArgKey)) {<a name="line.688"></a>
-<span class="sourceLineNo">689</span>          tableHash.scanBatch = Integer.parseInt(cmd.substring(scanBatchArgKey.length()));<a name="line.689"></a>
-<span class="sourceLineNo">690</span>          continue;<a name="line.690"></a>
-<span class="sourceLineNo">691</span>        }<a name="line.691"></a>
-<span class="sourceLineNo">692</span><a name="line.692"></a>
-<span class="sourceLineNo">693</span>        final String versionsArgKey = "--versions=";<a name="line.693"></a>
-<span class="sourceLineNo">694</span>        if (cmd.startsWith(versionsArgKey)) {<a name="line.694"></a>
-<span class="sourceLineNo">695</span>          tableHash.versions = Integer.parseInt(cmd.substring(versionsArgKey.length()));<a name="line.695"></a>
-<span class="sourceLineNo">696</span>          continue;<a name="line.696"></a>
-<span class="sourceLineNo">697</span>        }<a name="line.697"></a>
-<span class="sourceLineNo">698</span><a name="line.698"></a>
-<span class="sourceLineNo">699</span>        final String familiesArgKey = "--families=";<a name="line.699"></a>
-<span class="sourceLineNo">700</span>        if (cmd.startsWith(familiesArgKey)) {<a name="line.700"></a>
-<span class="sourceLineNo">701</span>          tableHash.families = cmd.substring(familiesArgKey.length());<a name="line.701"></a>
-<span class="sourceLineNo">702</span>          continue;<a name="line.702"></a>
-<span class="sourceLineNo">703</span>        }<a name="line.703"></a>
-<span class="sourceLineNo">704</span><a name="line.704"></a>
-<span class="sourceLineNo">705</span>        printUsage("Invalid argument '" + cmd + "'");<a name="line.705"></a>
-<span class="sourceLineNo">706</span>        return false;<a name="line.706"></a>
-<span class="sourceLineNo">707</span>      }<a name="line.707"></a>
-<span class="sourceLineNo">708</span>      if ((tableHash.startTime != 0 || tableHash.endTime != 0)<a name="line.708"></a>
-<span class="sourceLineNo">709</span>          &amp;&amp; (tableHash.startTime &gt;= tableHash.endTime)) {<a name="line.709"></a>
-<span class="sourceLineNo">710</span>        printUsage("Invalid time range filter: starttime="<a name="line.710"></a>
-<span class="sourceLineNo">711</span>            + tableHash.startTime + " &gt;=  endtime=" + tableHash.endTime);<a name="line.711"></a>
-<span class="sourceLineNo">712</span>        return false;<a name="line.712"></a>
-<span class="sourceLineNo">713</span>      }<a name="line.713"></a>
-<span class="sourceLineNo">714</span><a name="line.714"></a>
-<span class="sourceLineNo">715</span>    } catch (Exception e) {<a name="line.715"></a>
-<span class="sourceLineNo">716</span>      LOG.error("Failed to parse commandLine arguments", e);<a name="line.716"></a>
-<span class="sourceLineNo">717</span>      printUsage("Can't start because " + e.getMessage());<a name="line.717"></a>
-<span class="sourceLineNo">718</span>      return false;<a name="line.718"></a>
-<span class="sourceLineNo">719</span>    }<a name="line.719"></a>
-<span class="sourceLineNo">720</span>    return true;<a name="line.720"></a>
-<span class="sourceLineNo">721</span>  }<a name="line.721"></a>
+<span class="sourceLineNo">470</span>  static class ResultHasher {<a name="line.470"></a>
+<span class="sourceLineNo">471</span>    private MessageDigest digest;<a name="line.471"></a>
+<span class="sourceLineNo">472</span><a name="line.472"></a>
+<span class="sourceLineNo">473</span>    private boolean batchStarted = false;<a name="line.473"></a>
+<span class="sourceLineNo">474</span>    private ImmutableBytesWritable batchStartKey;<a name="line.474"></a>
+<span class="sourceLineNo">475</span>    private ImmutableBytesWritable batchHash;<a name="line.475"></a>
+<span class="sourceLineNo">476</span>    private long batchSize = 0;<a name="line.476"></a>
+<span class="sourceLineNo">477</span>    boolean ignoreTimestamps;<a name="line.477"></a>
+<span class="sourceLineNo">478</span><a name="line.478"></a>
+<span class="sourceLineNo">479</span><a name="line.479"></a>
+<span class="sourceLineNo">480</span>    public ResultHasher() {<a name="line.480"></a>
+<span class="sourceLineNo">481</span>      try {<a name="line.481"></a>
+<span class="sourceLineNo">482</span>        digest = MessageDigest.getInstance("MD5");<a name="line.482"></a>
+<span class="sourceLineNo">483</span>      } catch (NoSuchAlgorithmException e) {<a name="line.483"></a>
+<span class="sourceLineNo">484</span>        Throwables.propagate(e);<a name="line.484"></a>
+<span class="sourceLineNo">485</span>      }<a name="line.485"></a>
+<span class="sourceLineNo">486</span>    }<a name="line.486"></a>
+<span class="sourceLineNo">487</span><a name="line.487"></a>
+<span class="sourceLineNo">488</span>    public void startBatch(ImmutableBytesWritable row) {<a name="line.488"></a>
+<span class="sourceLineNo">489</span>      if (batchStarted) {<a name="line.489"></a>
+<span class="sourceLineNo">490</span>        throw new RuntimeException("Cannot start new batch without finishing existing one.");<a name="line.490"></a>
+<span class="sourceLineNo">491</span>      }<a name="line.491"></a>
+<span class="sourceLineNo">492</span>      batchStarted = true;<a name="line.492"></a>
+<span class="sourceLineNo">493</span>      batchSize = 0;<a name="line.493"></a>
+<span class="sourceLineNo">494</span>      batchStartKey = row;<a name="line.494"></a>
+<span class="sourceLineNo">495</span>      batchHash = null;<a name="line.495"></a>
+<span class="sourceLineNo">496</span>    }<a name="line.496"></a>
+<span class="sourceLineNo">497</span><a name="line.497"></a>
+<span class="sourceLineNo">498</span>    public void hashResult(Result result) {<a name="line.498"></a>
+<span class="sourceLineNo">499</span>      if (!batchStarted) {<a name="line.499"></a>
+<span class="sourceLineNo">500</span>        throw new RuntimeException("Cannot add to batch that has not been started.");<a name="line.500"></a>
+<span class="sourceLineNo">501</span>      }<a name="line.501"></a>
+<span class="sourceLineNo">502</span>      for (Cell cell : result.rawCells()) {<a name="line.502"></a>
+<span class="sourceLineNo">503</span>        int rowLength = cell.getRowLength();<a name="line.503"></a>
+<span class="sourceLineNo">504</span>        int familyLength = cell.getFamilyLength();<a name="line.504"></a>
+<span class="sourceLineNo">505</span>        int qualifierLength = cell.getQualifierLength();<a name="line.505"></a>
+<span class="sourceLineNo">506</span>        int valueLength = cell.getValueLength();<a name="line.506"></a>
+<span class="sourceLineNo">507</span>        digest.update(cell.getRowArray(), cell.getRowOffset(), rowLength);<a name="line.507"></a>
+<span class="sourceLineNo">508</span>        digest.update(cell.getFamilyArray(), cell.getFamilyOffset(), familyLength);<a name="line.508"></a>
+<span class="sourceLineNo">509</span>        digest.update(cell.getQualifierArray(), cell.getQualifierOffset(), qualifierLength);<a name="line.509"></a>
+<span class="sourceLineNo">510</span><a name="line.510"></a>
+<span class="sourceLineNo">511</span>        if (!ignoreTimestamps) {<a name="line.511"></a>
+<span class="sourceLineNo">512</span>          long ts = cell.getTimestamp();<a name="line.512"></a>
+<span class="sourceLineNo">513</span>          for (int i = 8; i &gt; 0; i--) {<a name="line.513"></a>
+<span class="sourceLineNo">514</span>            digest.update((byte) ts);<a name="line.514"></a>
+<span class="sourceLineNo">515</span>            ts &gt;&gt;&gt;= 8;<a name="line.515"></a>
+<span class="sourceLineNo">516</span>          }<a name="line.516"></a>
+<span class="sourceLineNo">517</span>        }<a name="line.517"></a>
+<span class="sourceLineNo">518</span>        digest.update(cell.getValueArray(), cell.getValueOffset(), valueLength);<a name="line.518"></a>
+<span class="sourceLineNo">519</span><a name="line.519"></a>
+<span class="sourceLineNo">520</span>        batchSize += rowLength + familyLength + qualifierLength + 8 + valueLength;<a name="line.520"></a>
+<span class="sourceLineNo">521</span>      }<a name="line.521"></a>
+<span class="sourceLineNo">522</span>    }<a name="line.522"></a>
+<span class="sourceLineNo">523</span><a name="line.523"></a>
+<span class="sourceLineNo">524</span>    public void finishBatch() {<a name="line.524"></a>
+<span class="sourceLineNo">525</span>      if (!batchStarted) {<a name="line.525"></a>
+<span class="sourceLineNo">526</span>        throw new RuntimeException("Cannot finish batch that has not started.");<a name="line.526"></a>
+<span class="sourceLineNo">527</span>      }<a name="line.527"></a>
+<span class="sourceLineNo">528</span>      batchStarted = false;<a name="line.528"></a>
+<span class="sourceLineNo">529</span>      batchHash = new ImmutableBytesWritable(digest.digest());<a name="line.529"></a>
+<span class="sourceLineNo">530</span>    }<a name="line.530"></a>
+<span class="sourceLineNo">531</span><a name="line.531"></a>
+<span class="sourceLineNo">532</span>    public boolean isBatchStarted() {<a name="line.532"></a>
+<span class="sourceLineNo">533</span>      return batchStarted;<a name="line.533"></a>
+<span class="sourceLineNo">534</span>    }<a name="line.534"></a>
+<span class="sourceLineNo">535</span><a name="line.535"></a>
+<span class="sourceLineNo">536</span>    public ImmutableBytesWritable getBatchStartKey() {<a name="line.536"></a>
+<span class="sourceLineNo">537</span>      return batchStartKey;<a name="line.537"></a>
+<span class="sourceLineNo">538</span>    }<a name="line.538"></a>
+<span class="sourceLineNo">539</span><a name="line.539"></a>
+<span class="sourceLineNo">540</span>    public ImmutableBytesWritable getBatchHash() {<a name="line.540"></a>
+<span class="sourceLineNo">541</span>      return batchHash;<a name="line.541"></a>
+<span class="sourceLineNo">542</span>    }<a name="line.542"></a>
+<span class="sourceLineNo">543</span><a name="line.543"></a>
+<span class="sourceLineNo">544</span>    public long getBatchSize() {<a name="line.544"></a>
+<span class="sourceLineNo">545</span>      return batchSize;<a name="line.545"></a>
+<span class="sourceLineNo">546</span>    }<a name="line.546"></a>
+<span class="sourceLineNo">547</span>  }<a name="line.547"></a>
+<span class="sourceLineNo">548</span><a name="line.548"></a>
+<span class="sourceLineNo">549</span>  public static class HashMapper<a name="line.549"></a>
+<span class="sourceLineNo">550</span>    extends TableMapper&lt;ImmutableBytesWritable, ImmutableBytesWritable&gt; {<a name="line.550"></a>
+<span class="sourceLineNo">551</span><a name="line.551"></a>
+<span class="sourceLineNo">552</span>    private ResultHasher hasher;<a name="line.552"></a>
+<span class="sourceLineNo">553</span>    private long targetBatchSize;<a name="line.553"></a>
+<span class="sourceLineNo">554</span><a name="line.554"></a>
+<span class="sourceLineNo">555</span>    private ImmutableBytesWritable currentRow;<a name="line.555"></a>
+<span class="sourceLineNo">556</span><a name="line.556"></a>
+<span class="sourceLineNo">557</span>    @Override<a name="line.557"></a>
+<span class="sourceLineNo">558</span>    protected void setup(Context context) throws IOException, InterruptedException {<a name="line.558"></a>
+<span class="sourceLineNo">559</span>      targetBatchSize = context.getConfiguration()<a name="line.559"></a>
+<span class="sourceLineNo">560</span>          .getLong(HASH_BATCH_SIZE_CONF_KEY, DEFAULT_BATCH_SIZE);<a name="line.560"></a>
+<span class="sourceLineNo">561</span>      hasher = new ResultHasher();<a name="line.561"></a>
+<span class="sourceLineNo">562</span>      hasher.ignoreTimestamps = context.getConfiguration().<a name="line.562"></a>
+<span class="sourceLineNo">563</span>        getBoolean(IGNORE_TIMESTAMPS, false);<a name="line.563"></a>
+<span class="sourceLineNo">564</span>      TableSplit split = (TableSplit) context.getInputSplit();<a name="line.564"></a>
+<span class="sourceLineNo">565</span>      hasher.startBatch(new ImmutableBytesWritable(split.getStartRow()));<a name="line.565"></a>
+<span class="sourceLineNo">566</span>    }<a name="line.566"></a>
+<span class="sourceLineNo">567</span><a name="line.567"></a>
+<span class="sourceLineNo">568</span>    @Override<a name="line.568"></a>
+<span class="sourceLineNo">569</span>    protected void map(ImmutableBytesWritable key, Result value, Context context)<a name="line.569"></a>
+<span class="sourceLineNo">570</span>        throws IOException, InterruptedException {<a name="line.570"></a>
+<span class="sourceLineNo">571</span><a name="line.571"></a>
+<span class="sourceLineNo">572</span>      if (currentRow == null || !currentRow.equals(key)) {<a name="line.572"></a>
+<span class="sourceLineNo">573</span>        currentRow = new ImmutableBytesWritable(key); // not immutable<a name="line.573"></a>
+<span class="sourceLineNo">574</span><a name="line.574"></a>
+<span class="sourceLineNo">575</span>        if (hasher.getBatchSize() &gt;= targetBatchSize) {<a name="line.575"></a>
+<span class="sourceLineNo">576</span>          hasher.finishBatch();<a name="line.576"></a>
+<span class="sourceLineNo">577</span>          context.write(hasher.getBatchStartKey(), hasher.getBatchHash());<a name="line.577"></a>
+<span class="sourceLineNo">578</span>          hasher.startBatch(currentRow);<a name="line.578"></a>
+<span class="sourceLineNo">579</span>        }<a name="line.579"></a>
+<span class="sourceLineNo">580</span>      }<a name="line.580"></a>
+<span class="sourceLineNo">581</span><a name="line.581"></a>
+<span class="sourceLineNo">582</span>      hasher.hashResult(value);<a name="line.582"></a>
+<span class="sourceLineNo">583</span>    }<a name="line.583"></a>
+<span class="sourceLineNo">584</span><a name="line.584"></a>
+<span class="sourceLineNo">585</span>    @Override<a name="line.585"></a>
+<span class="sourceLineNo">586</span>    protected void cleanup(Context context) throws IOException, InterruptedException {<a name="line.586"></a>
+<span class="sourceLineNo">587</span>      hasher.finishBatch();<a name="line.587"></a>
+<span class="sourceLineNo">588</span>      context.write(hasher.getBatchStartKey(), hasher.getBatchHash());<a name="line.588"></a>
+<span class="sourceLineNo">589</span>    }<a name="line.589"></a>
+<span class="sourceLineNo">590</span>  }<a name="line.590"></a>
+<span class="sourceLineNo">591</span><a name="line.591"></a>
+<span class="sourceLineNo">592</span>  private void writeTempManifestFile() throws IOException {<a name="line.592"></a>
+<span class="sourceLineNo">593</span>    Path tempManifestPath = new Path(destPath, TMP_MANIFEST_FILE_NAME);<a name="line.593"></a>
+<span class="sourceLineNo">594</span>    FileSystem fs = tempManifestPath.getFileSystem(getConf());<a name="line.594"></a>
+<span class="sourceLineNo">595</span>    tableHash.writePropertiesFile(fs, tempManifestPath);<a name="line.595"></a>
+<span class="sourceLineNo">596</span>  }<a name="line.596"></a>
+<span class="sourceLineNo">597</span><a name="line.597"></a>
+<span class="sourceLineNo">598</span>  private void completeManifest() throws IOException {<a name="line.598"></a>
+<span class="sourceLineNo">599</span>    Path tempManifestPath = new Path(destPath, TMP_MANIFEST_FILE_NAME);<a name="line.599"></a>
+<span class="sourceLineNo">600</span>    Path manifestPath = new Path(destPath, MANIFEST_FILE_NAME);<a name="line.600"></a>
+<span class="sourceLineNo">601</span>    FileSystem fs = tempManifestPath.getFileSystem(getConf());<a name="line.601"></a>
+<span class="sourceLineNo">602</span>    fs.rename(tempManifestPath, manifestPath);<a name="line.602"></a>
+<span class="sourceLineNo">603</span>  }<a name="line.603"></a>
+<span class="sourceLineNo">604</span><a name="line.604"></a>
+<span class="sourceLineNo">605</span>  private static final int NUM_ARGS = 2;<a name="line.605"></a>
+<span class="sourceLineNo">606</span>  private static void printUsage(final String errorMsg) {<a name="line.606"></a>
+<span class="sourceLineNo">607</span>    if (errorMsg != null &amp;&amp; errorMsg.length() &gt; 0) {<a name="line.607"></a>
+<span class="sourceLineNo">608</span>      System.err.println("ERROR: " + errorMsg);<a name="line.608"></a>
+<span class="sourceLineNo">609</span>      System.err.println();<a name="line.609"></a>
+<span class="sourceLineNo">610</span>    }<a name="line.610"></a>
+<span class="sourceLineNo">611</span>    System.err.println("Usage: HashTable [options] &lt;tablename&gt; &lt;outputpath&gt;");<a name="line.611"></a>
+<span class="sourceLineNo">612</span>    System.err.println();<a name="line.612"></a>
+<span class="sourceLineNo">613</span>    System.err.println("Options:");<a name="line.613"></a>
+<span class="sourceLineNo">614</span>    System.err.println(" batchsize         the target amount of bytes to hash in each batch");<a name="line.614"></a>
+<span class="sourceLineNo">615</span>    System.err.println("                   rows are added to the batch until this size is reached");<a name="line.615"></a>
+<span class="sourceLineNo">616</span>    System.err.println("                   (defaults to " + DEFAULT_BATCH_SIZE + " bytes)");<a name="line.616"></a>
+<span class="sourceLineNo">617</span>    System.err.println(" numhashfiles      the number of hash files to create");<a name="line.617"></a>
+<span class="sourceLineNo">618</span>    System.err.println("                   if set to fewer than number of regions then");<a name="line.618"></a>
+<span class="sourceLineNo">619</span>    System.err.println("                   the job will create this number of reducers");<a name="line.619"></a>
+<span class="sourceLineNo">620</span>    System.err.println("                   (defaults to 1/100 of regions -- at least 1)");<a name="line.620"></a>
+<span class="sourceLineNo">621</span>    System.err.println(" startrow          the start row");<a name="line.621"></a>
+<span class="sourceLineNo">622</span>    System.err.println(" stoprow           the stop row");<a name="line.622"></a>
+<span class="sourceLineNo">623</span>    System.err.println(" starttime         beginning of the time range (unixtime in millis)");<a name="line.623"></a>
+<span class="sourceLineNo">624</span>    System.err.println("                   without endtime means from starttime to forever");<a name="line.624"></a>
+<span class="sourceLineNo">625</span>    System.err.println(" endtime           end of the time range.");<a name="line.625"></a>
+<span class="sourceLineNo">626</span>    System.err.println("                   Ignored if no starttime specified.");<a name="line.626"></a>
+<span class="sourceLineNo">627</span>    System.err.println(" scanbatch         scanner batch size to support intra row scans");<a name="line.627"></a>
+<span class="sourceLineNo">628</span>    System.err.println(" versions          number of cell versions to include");<a name="line.628"></a>
+<span class="sourceLineNo">629</span>    System.err.println(" families          comma-separated list of families to include");<a name="line.629"></a>
+<span class="sourceLineNo">630</span>    System.err.println(" ignoreTimestamps  if true, ignores cell timestamps");<a name="line.630"></a>
+<span class="sourceLineNo">631</span>    System.err.println("                   when calculating hashes");<a name="line.631"></a>
+<span class="sourceLineNo">632</span>    System.err.println();<a name="line.632"></a>
+<span class="sourceLineNo">633</span>    System.err.println("Args:");<a name="line.633"></a>
+<span class="sourceLineNo">634</span>    System.err.println(" tablename     Name of the table to hash");<a name="line.634"></a>
+<span class="sourceLineNo">635</span>    System.err.println(" outputpath    Filesystem path to put the output data");<a name="line.635"></a>
+<span class="sourceLineNo">636</span>    System.err.println();<a name="line.636"></a>
+<span class="sourceLineNo">637</span>    System.err.println("Examples:");<a name="line.637"></a>
+<span class="sourceLineNo">638</span>    System.err.println(" To hash 'TestTable' in 32kB batches for a 1 hour window into 50 files:");<a name="line.638"></a>
+<span class="sourceLineNo">639</span>    System.err.println(" $ hbase " +<a name="line.639"></a>
+<span class="sourceLineNo">640</span>        "org.apache.hadoop.hbase.mapreduce.HashTable --batchsize=32000 --numhashfiles=50"<a name="line.640"></a>
+<span class="sourceLineNo">641</span>        + " --starttime=1265875194289 --endtime=1265878794289 --families=cf2,cf3"<a name="line.641"></a>
+<span class="sourceLineNo">642</span>        + " TestTable /hashes/testTable");<a name="line.642"></a>
+<span class="sourceLineNo">643</span>  }<a name="line.643"></a>
+<span class="sourceLineNo">644</span><a name="line.644"></a>
+<span class="sourceLineNo">645</span>  private boolean doCommandLine(final String[] args) {<a name="line.645"></a>
+<span class="sourceLineNo">646</span>    if (args.length &lt; NUM_ARGS) {<a name="line.646"></a>
+<span class="sourceLineNo">647</span>      printUsage(null);<a name="line.647"></a>
+<span class="sourceLineNo">648</span>      return false;<a name="line.648"></a>
+<span class="sourceLineNo">649</span>    }<a name="line.649"></a>
+<span class="sourceLineNo">650</span>    try {<a name="line.650"></a>
+<span class="sourceLineNo">651</span><a name="line.651"></a>
+<span class="sourceLineNo">652</span>      tableHash.tableName = args[args.length-2];<a name="line.652"></a>
+<span class="sourceLineNo">653</span>      destPath = new Path(args[args.length-1]);<a name="line.653"></a>
+<span class="sourceLineNo">654</span><a name="line.654"></a>
+<span class="sourceLineNo">655</span>      for (int i = 0; i &lt; args.length - NUM_ARGS; i++) {<a name="line.655"></a>
+<span class="sourceLineNo">656</span>        String cmd = args[i];<a name="line.656"></a>
+<span class="sourceLineNo">657</span>        if (cmd.equals("-h") || cmd.startsWith("--h")) {<a name="line.657"></a>
+<span class="sourceLineNo">658</span>          printUsage(null);<a name="line.658"></a>
+<span class="sourceLineNo">659</span>          return false;<a name="line.659"></a>
+<span class="sourceLineNo">660</span>        }<a name="line.660"></a>
+<span class="sourceLineNo">661</span><a name="line.661"></a>
+<span class="sourceLineNo">662</span>        final String batchSizeArgKey = "--batchsize=";<a name="line.662"></a>
+<span class="sourceLineNo">663</span>        if (cmd.startsWith(batchSizeArgKey)) {<a name="line.663"></a>
+<span class="sourceLineNo">664</span>          tableHash.batchSize = Long.parseLong(cmd.substring(batchSizeArgKey.length()));<a name="line.664"></a>
+<span class="sourceLineNo">665</span>          continue;<a name="line.665"></a>
+<span class="sourceLineNo">666</span>        }<a name="line.666"></a>
+<span class="sourceLineNo">667</span><a name="line.667"></a>
+<span class="sourceLineNo">668</span>        final String numHashFilesArgKey = "--numhashfiles=";<a name="line.668"></a>
+<span class="sourceLineNo">669</span>        if (cmd.startsWith(numHashFilesArgKey)) {<a name="line.669"></a>
+<span class="sourceLineNo">670</span>          tableHash.numHashFiles = Integer.parseInt(cmd.substring(numHashFilesArgKey.length()));<a name="line.670"></a>
+<span class="sourceLineNo">671</span>          continue;<a name="line.671"></a>
+<span class="sourceLineNo">672</span>        }<a name="line.672"></a>
+<span class="sourceLineNo">673</span><a name="line.673"></a>
+<span class="sourceLineNo">674</span>        final String startRowArgKey = "--startrow=";<a name="line.674"></a>
+<span class="sourceLineNo">675</span>        if (cmd.startsWith(startRowArgKey)) {<a name="line.675"></a>
+<span class="sourceLineNo">676</span>          tableHash.startRow = Bytes.fromHex(cmd.substring(startRowArgKey.length()));<a name="line.676"></a>
+<span class="sourceLineNo">677</span>          continue;<a name="line.677"></a>
+<span class="sourceLineNo">678</span>        }<a name="line.678"></a>
+<span class="sourceLineNo">679</span><a name="line.679"></a>
+<span class="sourceLineNo">680</span>        final String stopRowArgKey = "--stoprow=";<a name="line.680"></a>
+<span class="sourceLineNo">681</span>        if (cmd.startsWith(stopRowArgKey)) {<a name="line.681"></a>
+<span class="sourceLineNo">682</span>          tableHash.stopRow = Bytes.fromHex(cmd.substring(stopRowArgKey.length()));<a name="line.682"></a>
+<span class="sourceLineNo">683</span>          continue;<a name="line.683"></a>
+<span class="sourceLineNo">684</span>        }<a name="line.684"></a>
+<span class="sourceLineNo">685</span><a name="line.685"></a>
+<span class="sourceLineNo">686</span>        final String startTimeArgKey = "--starttime=";<a name="line.686"></a>
+<span class="sourceLineNo">687</span>        if (cmd.startsWith(startTimeArgKey)) {<a name="line.687"></a>
+<span class="sourceLineNo">688</span>          tableHash.startTime = Long.parseLong(cmd.substring(startTimeArgKey.length()));<a name="line.688"></a>
+<span class="sourceLineNo">689</span>          continue;<a name="line.689"></a>
+<span class="sourceLineNo">690</span>        }<a name="line.690"></a>
+<span class="sourceLineNo">691</span><a name="line.691"></a>
+<span class="sourceLineNo">692</span>        final String endTimeArgKey = "--endtime=";<a name="line.692"></a>
+<span class="sourceLineNo">693</span>        if (cmd.startsWith(endTimeArgKey)) {<a name="line.693"></a>
+<span class="sourceLineNo">694</span>          tableHash.endTime = Long.parseLong(cmd.substring(endTimeArgKey.length()));<a name="line.694"></a>
+<span class="sourceLineNo">695</span>          continue;<a name="line.695"></a>
+<span class="sourceLineNo">696</span>        }<a name="line.696"></a>
+<span class="sourceLineNo">697</span><a name="line.697"></a>
+<span class="sourceLineNo">698</span>        final String scanBatchArgKey = "--scanbatch=";<a name="line.698"></a>
+<span class="sourceLineNo">699</span>        if (cmd.startsWith(scanBatchArgKey)) {<a name="line.699"></a>
+<span class="sourceLineNo">700</span>          tableHash.scanBatch = Integer.parseInt(cmd.substring(scanBatchArgKey.length()));<a name="line.700"></a>
+<span class="sourceLineNo">701</span>          continue;<a name="line.701"></a>
+<span class="sourceLineNo">702</span>        }<a name="line.702"></a>
+<span class="sourceLineNo">703</span><a name="line.703"></a>
+<span class="sourceLineNo">704</span>        final String versionsArgKey = "--versions=";<a name="line.704"></a>
+<span class="sourceLineNo">705</span>        if (cmd.startsWith(versionsArgKey)) {<a name="line.705"></a>
+<span class="sourceLineNo">706</span>          tableHash.versions = Integer.parseInt(cmd.substring(versionsArgKey.length()));<a name="line.706"></a>
+<span class="sourceLineNo">707</span>          continue;<a name="line.707"></a>
+<span class="sourceLineNo">708</span>        }<a name="line.708"></a>
+<span class="sourceLineNo">709</span><a name="line.709"></a>
+<span class="sourceLineNo">710</span>        final String familiesArgKey = "--families=";<a name="line.710"></a>
+<span class="sourceLineNo">711</span>        if (cmd.startsWith(familiesArgKey)) {<a name="line.711"></a>
+<span class="sourceLineNo">712</span>          tableHash.families = cmd.substring(familiesArgKey.length());<a name="line.712"></a>
+<span class="sourceLineNo">713</span>          continue;<a name="line.713"></a>
+<span class="sourceLineNo">714</span>        }<a name="line.714"></a>
+<span class="sourceLineNo">715</span><a name="line.715"></a>
+<span class="sourceLineNo">716</span>        final String ignoreTimestampsKey = "--ignoreTimestamps=";<a name="line.716"></a>
+<span class="sourceLineNo">717</span>        if (cmd.startsWith(ignoreTimestampsKey)) {<a name="line.717"></a>
+<span class="sourceLineNo">718</span>          tableHash.ignoreTimestamps = Boolean.<a name="line.718"></a>
+<span class="sourceLineNo">719</span>            parseBoolean(cmd.substring(ignoreTimestampsKey.length()));<a name="line.719"></a>
+<span class="sourceLineNo">720</span>          continue;<a name="line.720"></a>
+<span class="sourceLineNo">721</span>        }<a name="line.721"></a>
 <span class="sourceLineNo">722</span><a name="line.722"></a>
-<span class="sourceLineNo">723</span>  /**<a name="line.723"></a>
-<span class="sourceLineNo">724</span>   * Main entry point.<a name="line.724"></a>
-<span class="sourceLineNo">725</span>   */<a name="line.725"></a>
-<span class="sourceLineNo">726</span>  public static void main(String[] args) throws Exception {<a name="line.726"></a>
-<span class="sourceLineNo">727</span>    int ret = ToolRunner.run(new HashTable(HBaseConfiguration.create()), args);<a name="line.727"></a>
-<span class="sourceLineNo">728</span>    System.exit(ret);<a name="line.728"></a>
-<span class="sourceLineNo">729</span>  }<a name="line.729"></a>
-<span class="sourceLineNo">730</span><a name="line.730"></a>
-<span class="sourceLineNo">731</span>  @Override<a name="line.731"></a>
-<span class="sourceLineNo">732</span>  public int run(String[] args) throws Exception {<a name="line.732"></a>
-<span class="sourceLineNo">733</span>    String[] otherArgs = new GenericOptionsParser(getConf(), args).getRemainingArgs();<a name="line.733"></a>
-<span class="sourceLineNo">734</span>    if (!doCommandLine(otherArgs)) {<a name="line.734"></a>
-<span class="sourceLineNo">735</span>      return 1;<a name="line.735"></a>
-<span class="sourceLineNo">736</span>    }<a name="line.736"></a>
-<span class="sourceLineNo">737</span><a name="line.737"></a>
-<span class="sourceLineNo">738</span>    Job job = createSubmittableJob(otherArgs);<a name="line.738"></a>
-<span class="sourceLineNo">739</span>    writeTempManifestFile();<a name="line.739"></a>
-<span class="sourceLineNo">740</span>    if (!job.waitForCompletion(true)) {<a name="line.740"></a>
-<span class="sourceLineNo">741</span>      LOG.info("Map-reduce job failed!");<a name="line.741"></a>
-<span class="sourceLineNo">742</span>      return 1;<a name="line.742"></a>
-<span class="sourceLineNo">743</span>    }<a name="line.743"></a>
-<span class="sourceLineNo">744</span>    completeManifest();<a name="line.744"></a>
-<span class="sourceLineNo">745</span>    return 0;<a name="line.745"></a>
-<span class="sourceLineNo">746</span>  }<a name="line.746"></a>
-<span class="sourceLineNo">747</span><a name="line.747"></a>
-<span class="sourceLineNo">748</span>}<a name="line.748"></a>
+<span class="sourceLineNo">723</span>        printUsage("Invalid argument '" + cmd + "'");<a name="line.723"></a>
+<span class="sourceLineNo">724</span>        return false;<a name="line.724"></a>
+<span class="sourceLineNo">725</span>      }<a name="line.725"></a>
+<span class="sourceLineNo">726</span>      if ((tableHash.startTime != 0 || tableHash.endTime != 0)<a name="line.726"></a>
+<span class="sourceLineNo">727</span>          &amp;&amp; (tableHash.startTime &gt;= tableHash.endTime)) {<a name="line.727"></a>
+<span class="sourceLineNo">728</span>        printUsage("Invalid time range filter: starttime="<a name="line.728"></a>
+<span class="sourceLineNo">729</span>            + tableHash.startTime + " &gt;=  endtime=" + tableHash.endTime);<a name="line.729"></a>
+<span class="sourceLineNo">730</span>        return false;<a name="line.730"></a>
+<span class="sourceLineNo">731</span>      }<a name="line.731"></a>
+<span class="sourceLineNo">732</span><a name="line.732"></a>
+<span class="sourceLineNo">733</span>    } catch (Exception e) {<a name="line.733"></a>
+<span class="sourceLineNo">734</span>      LOG.error("Failed to parse commandLine arguments", e);<a name="line.734"></a>
+<span class="sourceLineNo">735</span>      printUsage("Can't start because " + e.getMessage());<a name="line.735"></a>
+<span class="sourceLineNo">736</span>      return false;<a name="line.736"></a>
+<span class="sourceLineNo">737</span>    }<a name="line.737"></a>
+<span class="sourceLineNo">738</span>    return true;<a name="line.738"></a>
+<span class="sourceLineNo">739</span>  }<a name="line.739"></a>
+<span class="sourceLineNo">740</span><a name="line.740"></a>
+<span class="sourceLineNo">741</span>  /**<a name="line.741"></a>
+<span class="sourceLineNo">742</span>   * Main entry point.<a name="line.742"></a>
+<span class="sourceLineNo">743</span>   */<a name="line.743"></a>
+<span class="sourceLineNo">744</span>  public static void main(String[] args) throws Exception {<a name="line.744"></a>
+<span class="sourceLineNo">745</span>    int ret = ToolRunner.run(new HashTable(HBaseConfiguration.create()), args);<a name="line.745"></a>
+<span class="sourceLineNo">746</span>    System.exit(ret);<a name="line.746"></a>
+<span class="sourceLineNo">747</span>  }<a name="line.747"></a>
+<span class="sourceLineNo">748</span><a name="line.748"></a>
+<span class="sourceLineNo">749</span>  @Override<a name="line.749"></a>
+<span class="sourceLineNo">750</span>  public int run(String[] args) throws Exception {<a name="line.750"></a>
+<span class="sourceLineNo">751</span>    String[] otherArgs = new GenericOptionsParser(getConf(), args).getRemainingArgs();<a name="line.751"></a>
+<span class="sourceLineNo">752</span>    if (!doCommandLine(otherArgs)) {<a name="line.752"></a>
+<span class="sourceLineNo">753</span>      return 1;<a name="line.753"></a>
+<span class="sourceLineNo">754</span>    }<a name="line.754"></a>
+<span class="sourceLineNo">755</span><a name="line.755"></a>
+<span class="sourceLineNo">756</span>    Job job = createSubmittableJob(otherArgs);<a name="line.756"></a>
+<span class="sourceLineNo">757</span>    writeTempManifestFile();<a name="line.757"></a>
+<span class="sourceLineNo">758</span>    if (!job.waitForCompletion(true)) {<a name="line.758"></a>
+<span class="sourceLineNo">759</span>      LOG.info("Map-reduce job failed!");<a name="line.759"></a>
+<span class="sourceLineNo">760</span>      return 1;<a name="line.760"></a>
+<span class="sourceLineNo">761</span>    }<a name="line.761"></a>
+<span class="sourceLineNo">762</span>    completeManifest();<a name="line.762"></a>
+<span class="sourceLineNo">763</span>    return 0;<a name="line.763"></a>
+<span class="sourceLineNo">764</span>  }<a name="line.764"></a>
+<span class="sourceLineNo">765</span><a name="line.765"></a>
+<span class="sourceLineNo">766</span>}<a name="line.766"></a>
 
 
 
diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html
index 55fc1c9..32934c7 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HashTable.ResultHasher.html
@@ -81,679 +81,697 @@
 <span class="sourceLineNo">073</span>  final static String MANIFEST_FILE_NAME = "manifest";<a name="line.73"></a>
 <span class="sourceLineNo">074</span>  final static String HASH_DATA_DIR = "hashes";<a name="line.74"></a>
 <span class="sourceLineNo">075</span>  final static String OUTPUT_DATA_FILE_PREFIX = "part-r-";<a name="line.75"></a>
-<span class="sourceLineNo">076</span>  private final static String TMP_MANIFEST_FILE_NAME = "manifest.tmp";<a name="line.76"></a>
-<span class="sourceLineNo">077</span><a name="line.77"></a>
-<span class="sourceLineNo">078</span>  TableHash tableHash = new TableHash();<a name="line.78"></a>
-<span class="sourceLineNo">079</span>  Path destPath;<a name="line.79"></a>
-<span class="sourceLineNo">080</span><a name="line.80"></a>
-<span class="sourceLineNo">081</span>  public HashTable(Configuration conf) {<a name="line.81"></a>
-<span class="sourceLineNo">082</span>    super(conf);<a name="line.82"></a>
-<span class="sourceLineNo">083</span>  }<a name="line.83"></a>
-<span class="sourceLineNo">084</span><a name="line.84"></a>
-<span class="sourceLineNo">085</span>  public static class TableHash {<a name="line.85"></a>
-<span class="sourceLineNo">086</span><a name="line.86"></a>
-<span class="sourceLineNo">087</span>    Path hashDir;<a name="line.87"></a>
-<span class="sourceLineNo">088</span><a name="line.88"></a>
-<span class="sourceLineNo">089</span>    String tableName;<a name="line.89"></a>
-<span class="sourceLineNo">090</span>    String families = null;<a name="line.90"></a>
-<span class="sourceLineNo">091</span>    long batchSize = DEFAULT_BATCH_SIZE;<a name="line.91"></a>
-<span class="sourceLineNo">092</span>    int numHashFiles = 0;<a name="line.92"></a>
-<span class="sourceLineNo">093</span>    byte[] startRow = HConstants.EMPTY_START_ROW;<a name="line.93"></a>
-<span class="sourceLineNo">094</span>    byte[] stopRow = HConstants.EMPTY_END_ROW;<a name="line.94"></a>
-<span class="sourceLineNo">095</span>    int scanBatch = 0;<a name="line.95"></a>
-<span class="sourceLineNo">096</span>    int versions = -1;<a name="line.96"></a>
-<span class="sourceLineNo">097</span>    long startTime = 0;<a name="line.97"></a>
-<span class="sourceLineNo">098</span>    long endTime = 0;<a name="line.98"></a>
-<span class="sourceLineNo">099</span><a name="line.99"></a>
-<span class="sourceLineNo">100</span>    List&lt;ImmutableBytesWritable&gt; partitions;<a name="line.100"></a>
+<span class="sourceLineNo">076</span>  final static String IGNORE_TIMESTAMPS = "ignoreTimestamps";<a name="line.76"></a>
+<span class="sourceLineNo">077</span>  private final static String TMP_MANIFEST_FILE_NAME = "manifest.tmp";<a name="line.77"></a>
+<span class="sourceLineNo">078</span><a name="line.78"></a>
+<span class="sourceLineNo">079</span>  TableHash tableHash = new TableHash();<a name="line.79"></a>
+<span class="sourceLineNo">080</span>  Path destPath;<a name="line.80"></a>
+<span class="sourceLineNo">081</span><a name="line.81"></a>
+<span class="sourceLineNo">082</span>  public HashTable(Configuration conf) {<a name="line.82"></a>
+<span class="sourceLineNo">083</span>    super(conf);<a name="line.83"></a>
+<span class="sourceLineNo">084</span>  }<a name="line.84"></a>
+<span class="sourceLineNo">085</span><a name="line.85"></a>
+<span class="sourceLineNo">086</span>  public static class TableHash {<a name="line.86"></a>
+<span class="sourceLineNo">087</span><a name="line.87"></a>
+<span class="sourceLineNo">088</span>    Path hashDir;<a name="line.88"></a>
+<span class="sourceLineNo">089</span><a name="line.89"></a>
+<span class="sourceLineNo">090</span>    String tableName;<a name="line.90"></a>
+<span class="sourceLineNo">091</span>    String families = null;<a name="line.91"></a>
+<span class="sourceLineNo">092</span>    long batchSize = DEFAULT_BATCH_SIZE;<a name="line.92"></a>
+<span class="sourceLineNo">093</span>    int numHashFiles = 0;<a name="line.93"></a>
+<span class="sourceLineNo">094</span>    byte[] startRow = HConstants.EMPTY_START_ROW;<a name="line.94"></a>
+<span class="sourceLineNo">095</span>    byte[] stopRow = HConstants.EMPTY_END_ROW;<a name="line.95"></a>
+<span class="sourceLineNo">096</span>    int scanBatch = 0;<a name="line.96"></a>
+<span class="sourceLineNo">097</span>    int versions = -1;<a name="line.97"></a>
+<span class="sourceLineNo">098</span>    long startTime = 0;<a name="line.98"></a>
+<span class="sourceLineNo">099</span>    long endTime = 0;<a name="line.99"></a>
+<span class="sourceLineNo">100</span>    boolean ignoreTimestamps;<a name="line.100"></a>
 <span class="sourceLineNo">101</span><a name="line.101"></a>
-<span class="sourceLineNo">102</span>    public static TableHash read(Configuration conf, Path hashDir) throws IOException {<a name="line.102"></a>
-<span class="sourceLineNo">103</span>      TableHash tableHash = new TableHash();<a name="line.103"></a>
-<span class="sourceLineNo">104</span>      FileSystem fs = hashDir.getFileSystem(conf);<a name="line.104"></a>
-<span class="sourceLineNo">105</span>      tableHash.hashDir = hashDir;<a name="line.105"></a>
-<span class="sourceLineNo">106</span>      tableHash.readPropertiesFile(fs, new Path(hashDir, MANIFEST_FILE_NAME));<a name="line.106"></a>
-<span class="sourceLineNo">107</span>      tableHash.readPartitionFile(fs, conf, new Path(hashDir, PARTITIONS_FILE_NAME));<a name="line.107"></a>
-<span class="sourceLineNo">108</span>      return tableHash;<a name="line.108"></a>
-<span class="sourceLineNo">109</span>    }<a name="line.109"></a>
-<span class="sourceLineNo">110</span><a name="line.110"></a>
-<span class="sourceLineNo">111</span>    void writePropertiesFile(FileSystem fs, Path path) throws IOException {<a name="line.111"></a>
-<span class="sourceLineNo">112</span>      Properties p = new Properties();<a name="line.112"></a>
-<span class="sourceLineNo">113</span>      p.setProperty("table", tableName);<a name="line.113"></a>
-<span class="sourceLineNo">114</span>      if (families != null) {<a name="line.114"></a>
-<span class="sourceLineNo">115</span>        p.setProperty("columnFamilies", families);<a name="line.115"></a>
-<span class="sourceLineNo">116</span>      }<a name="line.116"></a>
-<span class="sourceLineNo">117</span>      p.setProperty("targetBatchSize", Long.toString(batchSize));<a name="line.117"></a>
-<span class="sourceLineNo">118</span>      p.setProperty("numHashFiles", Integer.toString(numHashFiles));<a name="line.118"></a>
-<span class="sourceLineNo">119</span>      if (!isTableStartRow(startRow)) {<a name="line.119"></a>
-<span class="sourceLineNo">120</span>        p.setProperty("startRowHex", Bytes.toHex(startRow));<a name="line.120"></a>
-<span class="sourceLineNo">121</span>      }<a name="line.121"></a>
-<span class="sourceLineNo">122</span>      if (!isTableEndRow(stopRow)) {<a name="line.122"></a>
-<span class="sourceLineNo">123</span>        p.setProperty("stopRowHex", Bytes.toHex(stopRow));<a name="line.123"></a>
-<span class="sourceLineNo">124</span>      }<a name="line.124"></a>
-<span class="sourceLineNo">125</span>      if (scanBatch &gt; 0) {<a name="line.125"></a>
-<span class="sourceLineNo">126</span>        p.setProperty("scanBatch", Integer.toString(scanBatch));<a name="line.126"></a>
-<span class="sourceLineNo">127</span>      }<a name="line.127"></a>
-<span class="sourceLineNo">128</span>      if (versions &gt;= 0) {<a name="line.128"></a>
-<span class="sourceLineNo">129</span>        p.setProperty("versions", Integer.toString(versions));<a name="line.129"></a>
-<span class="sourceLineNo">130</span>      }<a name="line.130"></a>
-<span class="sourceLineNo">131</span>      if (startTime != 0) {<a name="line.131"></a>
-<span class="sourceLineNo">132</span>        p.setProperty("startTimestamp", Long.toString(startTime));<a name="line.132"></a>
-<span class="sourceLineNo">133</span>      }<a name="line.133"></a>
-<span class="sourceLineNo">134</span>      if (endTime != 0) {<a name="line.134"></a>
-<span class="sourceLineNo">135</span>        p.setProperty("endTimestamp", Long.toString(endTime));<a name="line.135"></a>
-<span class="sourceLineNo">136</span>      }<a name="line.136"></a>
-<span class="sourceLineNo">137</span><a name="line.137"></a>
-<span class="sourceLineNo">138</span>      try (OutputStreamWriter osw = new OutputStreamWriter(fs.create(path), Charsets.UTF_8)) {<a name="line.138"></a>
-<span class="sourceLineNo">139</span>        p.store(osw, null);<a name="line.139"></a>
-<span class="sourceLineNo">140</span>      }<a name="line.140"></a>
-<span class="sourceLineNo">141</span>    }<a name="line.141"></a>
-<span class="sourceLineNo">142</span><a name="line.142"></a>
-<span class="sourceLineNo">143</span>    void readPropertiesFile(FileSystem fs, Path path) throws IOException {<a name="line.143"></a>
-<span class="sourceLineNo">144</span>      Properties p = new Properties();<a name="line.144"></a>
-<span class="sourceLineNo">145</span>      try (FSDataInputStream in = fs.open(path)) {<a name="line.145"></a>
-<span class="sourceLineNo">146</span>        try (InputStreamReader isr = new InputStreamReader(in, Charsets.UTF_8)) {<a name="line.146"></a>
-<span class="sourceLineNo">147</span>          p.load(isr);<a name="line.147"></a>
-<span class="sourceLineNo">148</span>        }<a name="line.148"></a>
-<span class="sourceLineNo">149</span>      }<a name="line.149"></a>
-<span class="sourceLineNo">150</span>      tableName = p.getProperty("table");<a name="line.150"></a>
-<span class="sourceLineNo">151</span>      families = p.getProperty("columnFamilies");<a name="line.151"></a>
-<span class="sourceLineNo">152</span>      batchSize = Long.parseLong(p.getProperty("targetBatchSize"));<a name="line.152"></a>
-<span class="sourceLineNo">153</span>      numHashFiles = Integer.parseInt(p.getProperty("numHashFiles"));<a name="line.153"></a>
-<span class="sourceLineNo">154</span><a name="line.154"></a>
-<span class="sourceLineNo">155</span>      String startRowHex = p.getProperty("startRowHex");<a name="line.155"></a>
-<span class="sourceLineNo">156</span>      if (startRowHex != null) {<a name="line.156"></a>
-<span class="sourceLineNo">157</span>        startRow = Bytes.fromHex(startRowHex);<a name="line.157"></a>
-<span class="sourceLineNo">158</span>      }<a name="line.158"></a>
-<span class="sourceLineNo">159</span>      String stopRowHex = p.getProperty("stopRowHex");<a name="line.159"></a>
-<span class="sourceLineNo">160</span>      if (stopRowHex != null) {<a name="line.160"></a>
-<span class="sourceLineNo">161</span>        stopRow = Bytes.fromHex(stopRowHex);<a name="line.161"></a>
-<span class="sourceLineNo">162</span>      }<a name="line.162"></a>
-<span class="sourceLineNo">163</span><a name="line.163"></a>
-<span class="sourceLineNo">164</span>      String scanBatchString = p.getProperty("scanBatch");<a name="line.164"></a>
-<span class="sourceLineNo">165</span>      if (scanBatchString != null) {<a name="line.165"></a>
-<span class="sourceLineNo">166</span>        scanBatch = Integer.parseInt(scanBatchString);<a name="line.166"></a>
-<span class="sourceLineNo">167</span>      }<a name="line.167"></a>
-<span class="sourceLineNo">168</span><a name="line.168"></a>
-<span class="sourceLineNo">169</span>      String versionString = p.getProperty("versions");<a name="line.169"></a>
-<span class="sourceLineNo">170</span>      if (versionString != null) {<a name="line.170"></a>
-<span class="sourceLineNo">171</span>        versions = Integer.parseInt(versionString);<a name="line.171"></a>
-<span class="sourceLineNo">172</span>      }<a name="line.172"></a>
-<span class="sourceLineNo">173</span><a name="line.173"></a>
-<span class="sourceLineNo">174</span>      String startTimeString = p.getProperty("startTimestamp");<a name="line.174"></a>
-<span class="sourceLineNo">175</span>      if (startTimeString != null) {<a name="line.175"></a>
-<span class="sourceLineNo">176</span>        startTime = Long.parseLong(startTimeString);<a name="line.176"></a>
-<span class="sourceLineNo">177</span>      }<a name="line.177"></a>
-<span class="sourceLineNo">178</span><a name="line.178"></a>
-<span class="sourceLineNo">179</span>      String endTimeString = p.getProperty("endTimestamp");<a name="line.179"></a>
-<span class="sourceLineNo">180</span>      if (endTimeString != null) {<a name="line.180"></a>
-<span class="sourceLineNo">181</span>        endTime = Long.parseLong(endTimeString);<a name="line.181"></a>
-<span class="sourceLineNo">182</span>      }<a name="line.182"></a>
-<span class="sourceLineNo">183</span>    }<a name="line.183"></a>
-<span class="sourceLineNo">184</span><a name="line.184"></a>
-<span class="sourceLineNo">185</span>    Scan initScan() throws IOException {<a name="line.185"></a>
-<span class="sourceLineNo">186</span>      Scan scan = new Scan();<a name="line.186"></a>
-<span class="sourceLineNo">187</span>      scan.setCacheBlocks(false);<a name="line.187"></a>
-<span class="sourceLineNo">188</span>      if (startTime != 0 || endTime != 0) {<a name="line.188"></a>
-<span class="sourceLineNo">189</span>        scan.setTimeRange(startTime, endTime == 0 ? HConstants.LATEST_TIMESTAMP : endTime);<a name="line.189"></a>
-<span class="sourceLineNo">190</span>      }<a name="line.190"></a>
-<span class="sourceLineNo">191</span>      if (scanBatch &gt; 0) {<a name="line.191"></a>
-<span class="sourceLineNo">192</span>        scan.setBatch(scanBatch);<a name="line.192"></a>
-<span class="sourceLineNo">193</span>      }<a name="line.193"></a>
-<span class="sourceLineNo">194</span>      if (versions &gt;= 0) {<a name="line.194"></a>
-<span class="sourceLineNo">195</span>        scan.readVersions(versions);<a name="line.195"></a>
-<span class="sourceLineNo">196</span>      }<a name="line.196"></a>
-<span class="sourceLineNo">197</span>      if (!isTableStartRow(startRow)) {<a name="line.197"></a>
-<span class="sourceLineNo">198</span>        scan.withStartRow(startRow);<a name="line.198"></a>
-<span class="sourceLineNo">199</span>      }<a name="line.199"></a>
-<span class="sourceLineNo">200</span>      if (!isTableEndRow(stopRow)) {<a name="line.200"></a>
-<span class="sourceLineNo">201</span>        scan.withStopRow(stopRow);<a name="line.201"></a>
-<span class="sourceLineNo">202</span>      }<a name="line.202"></a>
-<span class="sourceLineNo">203</span>      if(families != null) {<a name="line.203"></a>
-<span class="sourceLineNo">204</span>        for(String fam : families.split(",")) {<a name="line.204"></a>
-<span class="sourceLineNo">205</span>          scan.addFamily(Bytes.toBytes(fam));<a name="line.205"></a>
-<span class="sourceLineNo">206</span>        }<a name="line.206"></a>
-<span class="sourceLineNo">207</span>      }<a name="line.207"></a>
-<span class="sourceLineNo">208</span>      return scan;<a name="line.208"></a>
-<span class="sourceLineNo">209</span>    }<a name="line.209"></a>
-<span class="sourceLineNo">210</span><a name="line.210"></a>
-<span class="sourceLineNo">211</span>    /**<a name="line.211"></a>
-<span class="sourceLineNo">212</span>     * Choose partitions between row ranges to hash to a single output file<a name="line.212"></a>
-<span class="sourceLineNo">213</span>     * Selects region boundaries that fall within the scan range, and groups them<a name="line.213"></a>
-<span class="sourceLineNo">214</span>     * into the desired number of partitions.<a name="line.214"></a>
-<span class="sourceLineNo">215</span>     */<a name="line.215"></a>
-<span class="sourceLineNo">216</span>    void selectPartitions(Pair&lt;byte[][], byte[][]&gt; regionStartEndKeys) {<a name="line.216"></a>
-<span class="sourceLineNo">217</span>      List&lt;byte[]&gt; startKeys = new ArrayList&lt;&gt;();<a name="line.217"></a>
-<span class="sourceLineNo">218</span>      for (int i = 0; i &lt; regionStartEndKeys.getFirst().length; i++) {<a name="line.218"></a>
-<span class="sourceLineNo">219</span>        byte[] regionStartKey = regionStartEndKeys.getFirst()[i];<a name="line.219"></a>
-<span class="sourceLineNo">220</span>        byte[] regionEndKey = regionStartEndKeys.getSecond()[i];<a name="line.220"></a>
-<span class="sourceLineNo">221</span><a name="line.221"></a>
-<span class="sourceLineNo">222</span>        // if scan begins after this region, or starts before this region, then drop this region<a name="line.222"></a>
-<span class="sourceLineNo">223</span>        // in other words:<a name="line.223"></a>
-<span class="sourceLineNo">224</span>        //   IF (scan begins before the end of this region<a name="line.224"></a>
-<span class="sourceLineNo">225</span>        //      AND scan ends before the start of this region)<a name="line.225"></a>
-<span class="sourceLineNo">226</span>        //   THEN include this region<a name="line.226"></a>
-<span class="sourceLineNo">227</span>        if ((isTableStartRow(startRow) || isTableEndRow(regionEndKey)<a name="line.227"></a>
-<span class="sourceLineNo">228</span>            || Bytes.compareTo(startRow, regionEndKey) &lt; 0)<a name="line.228"></a>
-<span class="sourceLineNo">229</span>          &amp;&amp; (isTableEndRow(stopRow) || isTableStartRow(regionStartKey)<a name="line.229"></a>
-<span class="sourceLineNo">230</span>            || Bytes.compareTo(stopRow, regionStartKey) &gt; 0)) {<a name="line.230"></a>
-<span class="sourceLineNo">231</span>          startKeys.add(regionStartKey);<a name="line.231"></a>
-<span class="sourceLineNo">232</span>        }<a name="line.232"></a>
-<span class="sourceLineNo">233</span>      }<a name="line.233"></a>
-<span class="sourceLineNo">234</span><a name="line.234"></a>
-<span class="sourceLineNo">235</span>      int numRegions = startKeys.size();<a name="line.235"></a>
-<span class="sourceLineNo">236</span>      if (numHashFiles == 0) {<a name="line.236"></a>
-<span class="sourceLineNo">237</span>        numHashFiles = numRegions / 100;<a name="line.237"></a>
-<span class="sourceLineNo">238</span>      }<a name="line.238"></a>
-<span class="sourceLineNo">239</span>      if (numHashFiles == 0) {<a name="line.239"></a>
-<span class="sourceLineNo">240</span>        numHashFiles = 1;<a name="line.240"></a>
-<span class="sourceLineNo">241</span>      }<a name="line.241"></a>
-<span class="sourceLineNo">242</span>      if (numHashFiles &gt; numRegions) {<a name="line.242"></a>
-<span class="sourceLineNo">243</span>        // can't partition within regions<a name="line.243"></a>
-<span class="sourceLineNo">244</span>        numHashFiles = numRegions;<a name="line.244"></a>
-<span class="sourceLineNo">245</span>      }<a name="line.245"></a>
-<span class="sourceLineNo">246</span><a name="line.246"></a>
-<span class="sourceLineNo">247</span>      // choose a subset of start keys to group regions into ranges<a name="line.247"></a>
-<span class="sourceLineNo">248</span>      partitions = new ArrayList&lt;&gt;(numHashFiles - 1);<a name="line.248"></a>
-<span class="sourceLineNo">249</span>      // skip the first start key as it is not a partition between ranges.<a name="line.249"></a>
-<span class="sourceLineNo">250</span>      for (long i = 1; i &lt; numHashFiles; i++) {<a name="line.250"></a>
-<span class="sourceLineNo">251</span>        int splitIndex = (int) (numRegions * i / numHashFiles);<a name="line.251"></a>
-<span class="sourceLineNo">252</span>        partitions.add(new ImmutableBytesWritable(startKeys.get(splitIndex)));<a name="line.252"></a>
-<span class="sourceLineNo">253</span>      }<a name="line.253"></a>
-<span class="sourceLineNo">254</span>    }<a name="line.254"></a>
-<span class="sourceLineNo">255</span><a name="line.255"></a>
-<span class="sourceLineNo">256</span>    void writePartitionFile(Configuration conf, Path path) throws IOException {<a name="line.256"></a>
-<span class="sourceLineNo">257</span>      FileSystem fs = path.getFileSystem(conf);<a name="line.257"></a>
-<span class="sourceLineNo">258</span>      @SuppressWarnings("deprecation")<a name="line.258"></a>
-<span class="sourceLineNo">259</span>      SequenceFile.Writer writer = SequenceFile.createWriter(<a name="line.259"></a>
-<span class="sourceLineNo">260</span>        fs, conf, path, ImmutableBytesWritable.class, NullWritable.class);<a name="line.260"></a>
-<span class="sourceLineNo">261</span><a name="line.261"></a>
-<span class="sourceLineNo">262</span>      for (int i = 0; i &lt; partitions.size(); i++) {<a name="line.262"></a>
-<span class="sourceLineNo">263</span>        writer.append(partitions.get(i), NullWritable.get());<a name="line.263"></a>
-<span class="sourceLineNo">264</span>      }<a name="line.264"></a>
-<span class="sourceLineNo">265</span>      writer.close();<a name="line.265"></a>
-<span class="sourceLineNo">266</span>    }<a name="line.266"></a>
-<span class="sourceLineNo">267</span><a name="line.267"></a>
-<span class="sourceLineNo">268</span>    private void readPartitionFile(FileSystem fs, Configuration conf, Path path)<a name="line.268"></a>
-<span class="sourceLineNo">269</span>         throws IOException {<a name="line.269"></a>
-<span class="sourceLineNo">270</span>      @SuppressWarnings("deprecation")<a name="line.270"></a>
-<span class="sourceLineNo">271</span>      SequenceFile.Reader reader = new SequenceFile.Reader(fs, path, conf);<a name="line.271"></a>
-<span class="sourceLineNo">272</span>      ImmutableBytesWritable key = new ImmutableBytesWritable();<a name="line.272"></a>
-<span class="sourceLineNo">273</span>      partitions = new ArrayList&lt;&gt;();<a name="line.273"></a>
-<span class="sourceLineNo">274</span>      while (reader.next(key)) {<a name="line.274"></a>
-<span class="sourceLineNo">275</span>        partitions.add(new ImmutableBytesWritable(key.copyBytes()));<a name="line.275"></a>
-<span class="sourceLineNo">276</span>      }<a name="line.276"></a>
-<span class="sourceLineNo">277</span>      reader.close();<a name="line.277"></a>
-<span class="sourceLineNo">278</span><a name="line.278"></a>
-<span class="sourceLineNo">279</span>      if (!Ordering.natural().isOrdered(partitions)) {<a name="line.279"></a>
-<span class="sourceLineNo">280</span>        throw new IOException("Partitions are not ordered!");<a name="line.280"></a>
-<span class="sourceLineNo">281</span>      }<a name="line.281"></a>
-<span class="sourceLineNo">282</span>    }<a name="line.282"></a>
-<span class="sourceLineNo">283</span><a name="line.283"></a>
-<span class="sourceLineNo">284</span>    @Override<a name="line.284"></a>
-<span class="sourceLineNo">285</span>    public String toString() {<a name="line.285"></a>
-<span class="sourceLineNo">286</span>      StringBuilder sb = new StringBuilder();<a name="line.286"></a>
-<span class="sourceLineNo">287</span>      sb.append("tableName=").append(tableName);<a name="line.287"></a>
-<span class="sourceLineNo">288</span>      if (families != null) {<a name="line.288"></a>
-<span class="sourceLineNo">289</span>        sb.append(", families=").append(families);<a name="line.289"></a>
-<span class="sourceLineNo">290</span>      }<a name="line.290"></a>
-<span class="sourceLineNo">291</span>      sb.append(", batchSize=").append(batchSize);<a name="line.291"></a>
-<span class="sourceLineNo">292</span>      sb.append(", numHashFiles=").append(numHashFiles);<a name="line.292"></a>
-<span class="sourceLineNo">293</span>      if (!isTableStartRow(startRow)) {<a name="line.293"></a>
-<span class="sourceLineNo">294</span>        sb.append(", startRowHex=").append(Bytes.toHex(startRow));<a name="line.294"></a>
-<span class="sourceLineNo">295</span>      }<a name="line.295"></a>
-<span class="sourceLineNo">296</span>      if (!isTableEndRow(stopRow)) {<a name="line.296"></a>
-<span class="sourceLineNo">297</span>        sb.append(", stopRowHex=").append(Bytes.toHex(stopRow));<a name="line.297"></a>
-<span class="sourceLineNo">298</span>      }<a name="line.298"></a>
-<span class="sourceLineNo">299</span>      if (scanBatch &gt;= 0) {<a name="line.299"></a>
-<span class="sourceLineNo">300</span>        sb.append(", scanBatch=").append(scanBatch);<a name="line.300"></a>
-<span class="sourceLineNo">301</span>      }<a name="line.301"></a>
-<span class="sourceLineNo">302</span>      if (versions &gt;= 0) {<a name="line.302"></a>
-<span class="sourceLineNo">303</span>        sb.append(", versions=").append(versions);<a name="line.303"></a>
-<span class="sourceLineNo">304</span>      }<a name="line.304"></a>
-<span class="sourceLineNo">305</span>      if (startTime != 0) {<a name="line.305"></a>
-<span class="sourceLineNo">306</span>        sb.append("startTime=").append(startTime);<a name="line.306"></a>
-<span class="sourceLineNo">307</span>      }<a name="line.307"></a>
-<span class="sourceLineNo">308</span>      if (endTime != 0) {<a name="line.308"></a>
-<span class="sourceLineNo">309</span>        sb.append("endTime=").append(endTime);<a name="line.309"></a>
-<span class="sourceLineNo">310</span>      }<a name="line.310"></a>
-<span class="sourceLineNo">311</span>      return sb.toString();<a name="line.311"></a>
-<span class="sourceLineNo">312</span>    }<a name="line.312"></a>
-<span class="sourceLineNo">313</span><a name="line.313"></a>
-<span class="sourceLineNo">314</span>    static String getDataFileName(int hashFileIndex) {<a name="line.314"></a>
-<span class="sourceLineNo">315</span>      return String.format(HashTable.OUTPUT_DATA_FILE_PREFIX + "%05d", hashFileIndex);<a name="line.315"></a>
-<span class="sourceLineNo">316</span>    }<a name="line.316"></a>
-<span class="sourceLineNo">317</span><a name="line.317"></a>
-<span class="sourceLineNo">318</span>    /**<a name="line.318"></a>
-<span class="sourceLineNo">319</span>     * Open a TableHash.Reader starting at the first hash at or after the given key.<a name="line.319"></a>
-<span class="sourceLineNo">320</span>     * @throws IOException<a name="line.320"></a>
-<span class="sourceLineNo">321</span>     */<a name="line.321"></a>
-<span class="sourceLineNo">322</span>    public Reader newReader(Configuration conf, ImmutableBytesWritable startKey)<a name="line.322"></a>
-<span class="sourceLineNo">323</span>        throws IOException {<a name="line.323"></a>
-<span class="sourceLineNo">324</span>      return new Reader(conf, startKey);<a name="line.324"></a>
-<span class="sourceLineNo">325</span>    }<a name="line.325"></a>
-<span class="sourceLineNo">326</span><a name="line.326"></a>
-<span class="sourceLineNo">327</span>    public class Reader implements java.io.Closeable {<a name="line.327"></a>
-<span class="sourceLineNo">328</span>      private final Configuration conf;<a name="line.328"></a>
-<span class="sourceLineNo">329</span><a name="line.329"></a>
-<span class="sourceLineNo">330</span>      private int hashFileIndex;<a name="line.330"></a>
-<span class="sourceLineNo">331</span>      private MapFile.Reader mapFileReader;<a name="line.331"></a>
-<span class="sourceLineNo">332</span><a name="line.332"></a>
-<span class="sourceLineNo">333</span>      private boolean cachedNext;<a name="line.333"></a>
-<span class="sourceLineNo">334</span>      private ImmutableBytesWritable key;<a name="line.334"></a>
-<span class="sourceLineNo">335</span>      private ImmutableBytesWritable hash;<a name="line.335"></a>
-<span class="sourceLineNo">336</span><a name="line.336"></a>
-<span class="sourceLineNo">337</span>      Reader(Configuration conf, ImmutableBytesWritable startKey) throws IOException {<a name="line.337"></a>
-<span class="sourceLineNo">338</span>        this.conf = conf;<a name="line.338"></a>
-<span class="sourceLineNo">339</span>        int partitionIndex = Collections.binarySearch(partitions, startKey);<a name="line.339"></a>
-<span class="sourceLineNo">340</span>        if (partitionIndex &gt;= 0) {<a name="line.340"></a>
-<span class="sourceLineNo">341</span>          // if the key is equal to a partition, then go the file after that partition<a name="line.341"></a>
-<span class="sourceLineNo">342</span>          hashFileIndex = partitionIndex+1;<a name="line.342"></a>
-<span class="sourceLineNo">343</span>        } else {<a name="line.343"></a>
-<span class="sourceLineNo">344</span>          // if the key is between partitions, then go to the file between those partitions<a name="line.344"></a>
-<span class="sourceLineNo">345</span>          hashFileIndex = -1-partitionIndex;<a name="line.345"></a>
-<span class="sourceLineNo">346</span>        }<a name="line.346"></a>
-<span class="sourceLineNo">347</span>        openHashFile();<a name="line.347"></a>
-<span class="sourceLineNo">348</span><a name="line.348"></a>
-<span class="sourceLineNo">349</span>        // MapFile's don't make it easy to seek() so that the subsequent next() returns<a name="line.349"></a>
-<span class="sourceLineNo">350</span>        // the desired key/value pair.  So we cache it for the first call of next().<a name="line.350"></a>
-<span class="sourceLineNo">351</span>        hash = new ImmutableBytesWritable();<a name="line.351"></a>
-<span class="sourceLineNo">352</span>        key = (ImmutableBytesWritable) mapFileReader.getClosest(startKey, hash);<a name="line.352"></a>
-<span class="sourceLineNo">353</span>        if (key == null) {<a name="line.353"></a>
-<span class="sourceLineNo">354</span>          cachedNext = false;<a name="line.354"></a>
-<span class="sourceLineNo">355</span>          hash = null;<a name="line.355"></a>
-<span class="sourceLineNo">356</span>        } else {<a name="line.356"></a>
-<span class="sourceLineNo">357</span>          cachedNext = true;<a name="line.357"></a>
-<span class="sourceLineNo">358</span>        }<a name="line.358"></a>
-<span class="sourceLineNo">359</span>      }<a name="line.359"></a>
-<span class="sourceLineNo">360</span><a name="line.360"></a>
-<span class="sourceLineNo">361</span>      /**<a name="line.361"></a>
-<span class="sourceLineNo">362</span>       * Read the next key/hash pair.<a name="line.362"></a>
-<span class="sourceLineNo">363</span>       * Returns true if such a pair exists and false when at the end of the data.<a name="line.363"></a>
-<span class="sourceLineNo">364</span>       */<a name="line.364"></a>
-<span class="sourceLineNo">365</span>      public boolean next() throws IOException {<a name="line.365"></a>
-<span class="sourceLineNo">366</span>        if (cachedNext) {<a name="line.366"></a>
-<span class="sourceLineNo">367</span>          cachedNext = false;<a name="line.367"></a>
-<span class="sourceLineNo">368</span>          return true;<a name="line.368"></a>
-<span class="sourceLineNo">369</span>        }<a name="line.369"></a>
-<span class="sourceLineNo">370</span>        key = new ImmutableBytesWritable();<a name="line.370"></a>
-<span class="sourceLineNo">371</span>        hash = new ImmutableBytesWritable();<a name="line.371"></a>
-<span class="sourceLineNo">372</span>        while (true) {<a name="line.372"></a>
-<span class="sourceLineNo">373</span>          boolean hasNext = mapFileReader.next(key, hash);<a name="line.373"></a>
-<span class="sourceLineNo">374</span>          if (hasNext) {<a name="line.374"></a>
-<span class="sourceLineNo">375</span>            return true;<a name="line.375"></a>
-<span class="sourceLineNo">376</span>          }<a name="line.376"></a>
-<span class="sourceLineNo">377</span>          hashFileIndex++;<a name="line.377"></a>
-<span class="sourceLineNo">378</span>          if (hashFileIndex &lt; TableHash.this.numHashFiles) {<a name="line.378"></a>
-<span class="sourceLineNo">379</span>            mapFileReader.close();<a name="line.379"></a>
-<span class="sourceLineNo">380</span>            openHashFile();<a name="line.380"></a>
-<span class="sourceLineNo">381</span>          } else {<a name="line.381"></a>
-<span class="sourceLineNo">382</span>            key = null;<a name="line.382"></a>
-<span class="sourceLineNo">383</span>            hash = null;<a name="line.383"></a>
-<span class="sourceLineNo">384</span>            return false;<a name="line.384"></a>
-<span class="sourceLineNo">385</span>          }<a name="line.385"></a>
-<span class="sourceLineNo">386</span>        }<a name="line.386"></a>
-<span class="sourceLineNo">387</span>      }<a name="line.387"></a>
-<span class="sourceLineNo">388</span><a name="line.388"></a>
-<span class="sourceLineNo">389</span>      /**<a name="line.389"></a>
-<span class="sourceLineNo">390</span>       * Get the current key<a name="line.390"></a>
-<span class="sourceLineNo">391</span>       * @return the current key or null if there is no current key<a name="line.391"></a>
-<span class="sourceLineNo">392</span>       */<a name="line.392"></a>
-<span class="sourceLineNo">393</span>      public ImmutableBytesWritable getCurrentKey() {<a name="line.393"></a>
-<span class="sourceLineNo">394</span>        return key;<a name="line.394"></a>
-<span class="sourceLineNo">395</span>      }<a name="line.395"></a>
-<span class="sourceLineNo">396</span><a name="line.396"></a>
-<span class="sourceLineNo">397</span>      /**<a name="line.397"></a>
-<span class="sourceLineNo">398</span>       * Get the current hash<a name="line.398"></a>
-<span class="sourceLineNo">399</span>       * @return the current hash or null if there is no current hash<a name="line.399"></a>
-<span class="sourceLineNo">400</span>       */<a name="line.400"></a>
-<span class="sourceLineNo">401</span>      public ImmutableBytesWritable getCurrentHash() {<a name="line.401"></a>
-<span class="sourceLineNo">402</span>        return hash;<a name="line.402"></a>
-<span class="sourceLineNo">403</span>      }<a name="line.403"></a>
-<span class="sourceLineNo">404</span><a name="line.404"></a>
-<span class="sourceLineNo">405</span>      private void openHashFile() throws IOException {<a name="line.405"></a>
-<span class="sourceLineNo">406</span>        if (mapFileReader != null) {<a name="line.406"></a>
-<span class="sourceLineNo">407</span>          mapFileReader.close();<a name="line.407"></a>
-<span class="sourceLineNo">408</span>        }<a name="line.408"></a>
-<span class="sourceLineNo">409</span>        Path dataDir = new Path(TableHash.this.hashDir, HASH_DATA_DIR);<a name="line.409"></a>
-<span class="sourceLineNo">410</span>        Path dataFile = new Path(dataDir, getDataFileName(hashFileIndex));<a name="line.410"></a>
-<span class="sourceLineNo">411</span>        mapFileReader = new MapFile.Reader(dataFile, conf);<a name="line.411"></a>
-<span class="sourceLineNo">412</span>      }<a name="line.412"></a>
-<span class="sourceLineNo">413</span><a name="line.413"></a>
-<span class="sourceLineNo">414</span>      @Override<a name="line.414"></a>
-<span class="sourceLineNo">415</span>      public void close() throws IOException {<a name="line.415"></a>
-<span class="sourceLineNo">416</span>        mapFileReader.close();<a name="line.416"></a>
-<span class="sourceLineNo">417</span>      }<a name="line.417"></a>
-<span class="sourceLineNo">418</span>    }<a name="line.418"></a>
-<span class="sourceLineNo">419</span>  }<a name="line.419"></a>
-<span class="sourceLineNo">420</span><a name="line.420"></a>
-<span class="sourceLineNo">421</span>  static boolean isTableStartRow(byte[] row) {<a name="line.421"></a>
-<span class="sourceLineNo">422</span>    return Bytes.equals(HConstants.EMPTY_START_ROW, row);<a name="line.422"></a>
-<span class="sourceLineNo">423</span>  }<a name="line.423"></a>
-<span class="sourceLineNo">424</span><a name="line.424"></a>
-<span class="sourceLineNo">425</span>  static boolean isTableEndRow(byte[] row) {<a name="line.425"></a>
-<span class="sourceLineNo">426</span>    return Bytes.equals(HConstants.EMPTY_END_ROW, row);<a name="line.426"></a>
-<span class="sourceLineNo">427</span>  }<a name="line.427"></a>
-<span class="sourceLineNo">428</span><a name="line.428"></a>
-<span class="sourceLineNo">429</span>  public Job createSubmittableJob(String[] args) throws IOException {<a name="line.429"></a>
-<span class="sourceLineNo">430</span>    Path partitionsPath = new Path(destPath, PARTITIONS_FILE_NAME);<a name="line.430"></a>
-<span class="sourceLineNo">431</span>    generatePartitions(partitionsPath);<a name="line.431"></a>
-<span class="sourceLineNo">432</span><a name="line.432"></a>
-<span class="sourceLineNo">433</span>    Job job = Job.getInstance(getConf(),<a name="line.433"></a>
-<span class="sourceLineNo">434</span>          getConf().get("mapreduce.job.name", "hashTable_" + tableHash.tableName));<a name="line.434"></a>
-<span class="sourceLineNo">435</span>    Configuration jobConf = job.getConfiguration();<a name="line.435"></a>
-<span class="sourceLineNo">436</span>    jobConf.setLong(HASH_BATCH_SIZE_CONF_KEY, tableHash.batchSize);<a name="line.436"></a>
-<span class="sourceLineNo">437</span>    job.setJarByClass(HashTable.class);<a name="line.437"></a>
-<span class="sourceLineNo">438</span><a name="line.438"></a>
-<span class="sourceLineNo">439</span>    TableMapReduceUtil.initTableMapperJob(tableHash.tableName, tableHash.initScan(),<a name="line.439"></a>
-<span class="sourceLineNo">440</span>        HashMapper.class, ImmutableBytesWritable.class, ImmutableBytesWritable.class, job);<a name="line.440"></a>
+<span class="sourceLineNo">102</span>    List&lt;ImmutableBytesWritable&gt; partitions;<a name="line.102"></a>
+<span class="sourceLineNo">103</span><a name="line.103"></a>
+<span class="sourceLineNo">104</span>    public static TableHash read(Configuration conf, Path hashDir) throws IOException {<a name="line.104"></a>
+<span class="sourceLineNo">105</span>      TableHash tableHash = new TableHash();<a name="line.105"></a>
+<span class="sourceLineNo">106</span>      FileSystem fs = hashDir.getFileSystem(conf);<a name="line.106"></a>
+<span class="sourceLineNo">107</span>      tableHash.hashDir = hashDir;<a name="line.107"></a>
+<span class="sourceLineNo">108</span>      tableHash.readPropertiesFile(fs, new Path(hashDir, MANIFEST_FILE_NAME));<a name="line.108"></a>
+<span class="sourceLineNo">109</span>      tableHash.readPartitionFile(fs, conf, new Path(hashDir, PARTITIONS_FILE_NAME));<a name="line.109"></a>
+<span class="sourceLineNo">110</span>      return tableHash;<a name="line.110"></a>
+<span class="sourceLineNo">111</span>    }<a name="line.111"></a>
+<span class="sourceLineNo">112</span><a name="line.112"></a>
+<span class="sourceLineNo">113</span>    void writePropertiesFile(FileSystem fs, Path path) throws IOException {<a name="line.113"></a>
+<span class="sourceLineNo">114</span>      Properties p = new Properties();<a name="line.114"></a>
+<span class="sourceLineNo">115</span>      p.setProperty("table", tableName);<a name="line.115"></a>
+<span class="sourceLineNo">116</span>      if (families != null) {<a name="line.116"></a>
+<span class="sourceLineNo">117</span>        p.setProperty("columnFamilies", families);<a name="line.117"></a>
+<span class="sourceLineNo">118</span>      }<a name="line.118"></a>
+<span class="sourceLineNo">119</span>      p.setProperty("targetBatchSize", Long.toString(batchSize));<a name="line.119"></a>
+<span class="sourceLineNo">120</span>      p.setProperty("numHashFiles", Integer.toString(numHashFiles));<a name="line.120"></a>
+<span class="sourceLineNo">121</span>      if (!isTableStartRow(startRow)) {<a name="line.121"></a>
+<span class="sourceLineNo">122</span>        p.setProperty("startRowHex", Bytes.toHex(startRow));<a name="line.122"></a>
+<span class="sourceLineNo">123</span>      }<a name="line.123"></a>
+<span class="sourceLineNo">124</span>      if (!isTableEndRow(stopRow)) {<a name="line.124"></a>
+<span class="sourceLineNo">125</span>        p.setProperty("stopRowHex", Bytes.toHex(stopRow));<a name="line.125"></a>
+<span class="sourceLineNo">126</span>      }<a name="line.126"></a>
+<span class="sourceLineNo">127</span>      if (scanBatch &gt; 0) {<a name="line.127"></a>
+<span class="sourceLineNo">128</span>        p.setProperty("scanBatch", Integer.toString(scanBatch));<a name="line.128"></a>
+<span class="sourceLineNo">129</span>      }<a name="line.129"></a>
+<span class="sourceLineNo">130</span>      if (versions &gt;= 0) {<a name="line.130"></a>
+<span class="sourceLineNo">131</span>        p.setProperty("versions", Integer.toString(versions));<a name="line.131"></a>
+<span class="sourceLineNo">132</span>      }<a name="line.132"></a>
+<span class="sourceLineNo">133</span>      if (startTime != 0) {<a name="line.133"></a>
+<span class="sourceLineNo">134</span>        p.setProperty("startTimestamp", Long.toString(startTime));<a name="line.134"></a>
+<span class="sourceLineNo">135</span>      }<a name="line.135"></a>
+<span class="sourceLineNo">136</span>      if (endTime != 0) {<a name="line.136"></a>
+<span class="sourceLineNo">137</span>        p.setProperty("endTimestamp", Long.toString(endTime));<a name="line.137"></a>
+<span class="sourceLineNo">138</span>      }<a name="line.138"></a>
+<span class="sourceLineNo">139</span><a name="line.139"></a>
+<span class="sourceLineNo">140</span>      try (OutputStreamWriter osw = new OutputStreamWriter(fs.create(path), Charsets.UTF_8)) {<a name="line.140"></a>
+<span class="sourceLineNo">141</span>        p.store(osw, null);<a name="line.141"></a>
+<span class="sourceLineNo">142</span>      }<a name="line.142"></a>
+<span class="sourceLineNo">143</span>    }<a name="line.143"></a>
+<span class="sourceLineNo">144</span><a name="line.144"></a>
+<span class="sourceLineNo">145</span>    void readPropertiesFile(FileSystem fs, Path path) throws IOException {<a name="line.145"></a>
+<span class="sourceLineNo">146</span>      Properties p = new Properties();<a name="line.146"></a>
+<span class="sourceLineNo">147</span>      try (FSDataInputStream in = fs.open(path)) {<a name="line.147"></a>
+<span class="sourceLineNo">148</span>        try (InputStreamReader isr = new InputStreamReader(in, Charsets.UTF_8)) {<a name="line.148"></a>
+<span class="sourceLineNo">149</span>          p.load(isr);<a name="line.149"></a>
+<span class="sourceLineNo">150</span>        }<a name="line.150"></a>
+<span class="sourceLineNo">151</span>      }<a name="line.151"></a>
+<span class="sourceLineNo">152</span>      tableName = p.getProperty("table");<a name="line.152"></a>
+<span class="sourceLineNo">153</span>      families = p.getProperty("columnFamilies");<a name="line.153"></a>
+<span class="sourceLineNo">154</span>      batchSize = Long.parseLong(p.getProperty("targetBatchSize"));<a name="line.154"></a>
+<span class="sourceLineNo">155</span>      numHashFiles = Integer.parseInt(p.getProperty("numHashFiles"));<a name="line.155"></a>
+<span class="sourceLineNo">156</span><a name="line.156"></a>
+<span class="sourceLineNo">157</span>      String startRowHex = p.getProperty("startRowHex");<a name="line.157"></a>
+<span class="sourceLineNo">158</span>      if (startRowHex != null) {<a name="line.158"></a>
+<span class="sourceLineNo">159</span>        startRow = Bytes.fromHex(startRowHex);<a name="line.159"></a>
+<span class="sourceLineNo">160</span>      }<a name="line.160"></a>
+<span class="sourceLineNo">161</span>      String stopRowHex = p.getProperty("stopRowHex");<a name="line.161"></a>
+<span class="sourceLineNo">162</span>      if (stopRowHex != null) {<a name="line.162"></a>
+<span class="sourceLineNo">163</span>        stopRow = Bytes.fromHex(stopRowHex);<a name="line.163"></a>
+<span class="sourceLineNo">164</span>      }<a name="line.164"></a>
+<span class="sourceLineNo">165</span><a name="line.165"></a>
+<span class="sourceLineNo">166</span>      String scanBatchString = p.getProperty("scanBatch");<a name="line.166"></a>
+<span class="sourceLineNo">167</span>      if (scanBatchString != null) {<a name="line.167"></a>
+<span class="sourceLineNo">168</span>        scanBatch = Integer.parseInt(scanBatchString);<a name="line.168"></a>
+<span class="sourceLineNo">169</span>      }<a name="line.169"></a>
+<span class="sourceLineNo">170</span><a name="line.170"></a>
+<span class="sourceLineNo">171</span>      String versionString = p.getProperty("versions");<a name="line.171"></a>
+<span class="sourceLineNo">172</span>      if (versionString != null) {<a name="line.172"></a>
+<span class="sourceLineNo">173</span>        versions = Integer.parseInt(versionString);<a name="line.173"></a>
+<span class="sourceLineNo">174</span>      }<a name="line.174"></a>
+<span class="sourceLineNo">175</span><a name="line.175"></a>
+<span class="sourceLineNo">176</span>      String startTimeString = p.getProperty("startTimestamp");<a name="line.176"></a>
+<span class="sourceLineNo">177</span>      if (startTimeString != null) {<a name="line.177"></a>
+<span class="sourceLineNo">178</span>        startTime = Long.parseLong(startTimeString);<a name="line.178"></a>
+<span class="sourceLineNo">179</span>      }<a name="line.179"></a>
+<span class="sourceLineNo">180</span><a name="line.180"></a>
+<span class="sourceLineNo">181</span>      String endTimeString = p.getProperty("endTimestamp");<a name="line.181"></a>
+<span class="sourceLineNo">182</span>      if (endTimeString != null) {<a name="line.182"></a>
+<span class="sourceLineNo">183</span>        endTime = Long.parseLong(endTimeString);<a name="line.183"></a>
+<span class="sourceLineNo">184</span>      }<a name="line.184"></a>
+<span class="sourceLineNo">185</span>    }<a name="line.185"></a>
+<span class="sourceLineNo">186</span><a name="line.186"></a>
+<span class="sourceLineNo">187</span>    Scan initScan() throws IOException {<a name="line.187"></a>
+<span class="sourceLineNo">188</span>      Scan scan = new Scan();<a name="line.188"></a>
+<span class="sourceLineNo">189</span>      scan.setCacheBlocks(false);<a name="line.189"></a>
+<span class="sourceLineNo">190</span>      if (startTime != 0 || endTime != 0) {<a name="line.190"></a>
+<span class="sourceLineNo">191</span>        scan.setTimeRange(startTime, endTime == 0 ? HConstants.LATEST_TIMESTAMP : endTime);<a name="line.191"></a>
+<span class="sourceLineNo">192</span>      }<a name="line.192"></a>
+<span class="sourceLineNo">193</span>      if (scanBatch &gt; 0) {<a name="line.193"></a>
+<span class="sourceLineNo">194</span>        scan.setBatch(scanBatch);<a name="line.194"></a>
+<span class="sourceLineNo">195</span>      }<a name="line.195"></a>
+<span class="sourceLineNo">196</span>      if (versions &gt;= 0) {<a name="line.196"></a>
+<span class="sourceLineNo">197</span>        scan.readVersions(versions);<a name="line.197"></a>
+<span class="sourceLineNo">198</span>      }<a name="line.198"></a>
+<span class="sourceLineNo">199</span>      if (!isTableStartRow(startRow)) {<a name="line.199"></a>
+<span class="sourceLineNo">200</span>        scan.withStartRow(startRow);<a name="line.200"></a>
+<span class="sourceLineNo">201</span>      }<a name="line.201"></a>
+<span class="sourceLineNo">202</span>      if (!isTableEndRow(stopRow)) {<a name="line.202"></a>
+<span class="sourceLineNo">203</span>        scan.withStopRow(stopRow);<a name="line.203"></a>
+<span class="sourceLineNo">204</span>      }<a name="line.204"></a>
+<span class="sourceLineNo">205</span>      if(families != null) {<a name="line.205"></a>
+<span class="sourceLineNo">206</span>        for(String fam : families.split(",")) {<a name="line.206"></a>
+<span class="sourceLineNo">207</span>          scan.addFamily(Bytes.toBytes(fam));<a name="line.207"></a>
+<span class="sourceLineNo">208</span>        }<a name="line.208"></a>
+<span class="sourceLineNo">209</span>      }<a name="line.209"></a>
+<span class="sourceLineNo">210</span>      return scan;<a name="line.210"></a>
+<span class="sourceLineNo">211</span>    }<a name="line.211"></a>
+<span class="sourceLineNo">212</span><a name="line.212"></a>
+<span class="sourceLineNo">213</span>    /**<a name="line.213"></a>
+<span class="sourceLineNo">214</span>     * Choose partitions between row ranges to hash to a single output file<a name="line.214"></a>
+<span class="sourceLineNo">215</span>     * Selects region boundaries that fall within the scan range, and groups them<a name="line.215"></a>
+<span class="sourceLineNo">216</span>     * into the desired number of partitions.<a name="line.216"></a>
+<span class="sourceLineNo">217</span>     */<a name="line.217"></a>
+<span class="sourceLineNo">218</span>    void selectPartitions(Pair&lt;byte[][], byte[][]&gt; regionStartEndKeys) {<a name="line.218"></a>
+<span class="sourceLineNo">219</span>      List&lt;byte[]&gt; startKeys = new ArrayList&lt;&gt;();<a name="line.219"></a>
+<span class="sourceLineNo">220</span>      for (int i = 0; i &lt; regionStartEndKeys.getFirst().length; i++) {<a name="line.220"></a>
+<span class="sourceLineNo">221</span>        byte[] regionStartKey = regionStartEndKeys.getFirst()[i];<a name="line.221"></a>
+<span class="sourceLineNo">222</span>        byte[] regionEndKey = regionStartEndKeys.getSecond()[i];<a name="line.222"></a>
+<span class="sourceLineNo">223</span><a name="line.223"></a>
+<span class="sourceLineNo">224</span>        // if scan begins after this region, or starts before this region, then drop this region<a name="line.224"></a>
+<span class="sourceLineNo">225</span>        // in other words:<a name="line.225"></a>
+<span class="sourceLineNo">226</span>        //   IF (scan begins before the end of this region<a name="line.226"></a>
+<span class="sourceLineNo">227</span>        //      AND scan ends before the start of this region)<a name="line.227"></a>
+<span class="sourceLineNo">228</span>        //   THEN include this region<a name="line.228"></a>
+<span class="sourceLineNo">229</span>        if ((isTableStartRow(startRow) || isTableEndRow(regionEndKey)<a name="line.229"></a>
+<span class="sourceLineNo">230</span>            || Bytes.compareTo(startRow, regionEndKey) &lt; 0)<a name="line.230"></a>
+<span class="sourceLineNo">231</span>          &amp;&amp; (isTableEndRow(stopRow) || isTableStartRow(regionStartKey)<a name="line.231"></a>
+<span class="sourceLineNo">232</span>            || Bytes.compareTo(stopRow, regionStartKey) &gt; 0)) {<a name="line.232"></a>
+<span class="sourceLineNo">233</span>          startKeys.add(regionStartKey);<a name="line.233"></a>
+<span class="sourceLineNo">234</span>        }<a name="line.234"></a>
+<span class="sourceLineNo">235</span>      }<a name="line.235"></a>
+<span class="sourceLineNo">236</span><a name="line.236"></a>
+<span class="sourceLineNo">237</span>      int numRegions = startKeys.size();<a name="line.237"></a>
+<span class="sourceLineNo">238</span>      if (numHashFiles == 0) {<a name="line.238"></a>
+<span class="sourceLineNo">239</span>        numHashFiles = numRegions / 100;<a name="line.239"></a>
+<span class="sourceLineNo">240</span>      }<a name="line.240"></a>
+<span class="sourceLineNo">241</span>      if (numHashFiles == 0) {<a name="line.241"></a>
+<span class="sourceLineNo">242</span>        numHashFiles = 1;<a name="line.242"></a>
+<span class="sourceLineNo">243</span>      }<a name="line.243"></a>
+<span class="sourceLineNo">244</span>      if (numHashFiles &gt; numRegions) {<a name="line.244"></a>
+<span class="sourceLineNo">245</span>        // can't partition within regions<a name="line.245"></a>
+<span class="sourceLineNo">246</span>        numHashFiles = numRegions;<a name="line.246"></a>
+<span class="sourceLineNo">247</span>      }<a name="line.247"></a>
+<span class="sourceLineNo">248</span><a name="line.248"></a>
+<span class="sourceLineNo">249</span>      // choose a subset of start keys to group regions into ranges<a name="line.249"></a>
+<span class="sourceLineNo">250</span>      partitions = new ArrayList&lt;&gt;(numHashFiles - 1);<a name="line.250"></a>
+<span class="sourceLineNo">251</span>      // skip the first start key as it is not a partition between ranges.<a name="line.251"></a>
+<span class="sourceLineNo">252</span>      for (long i = 1; i &lt; numHashFiles; i++) {<a name="line.252"></a>
+<span class="sourceLineNo">253</span>        int splitIndex = (int) (numRegions * i / numHashFiles);<a name="line.253"></a>
+<span class="sourceLineNo">254</span>        partitions.add(new ImmutableBytesWritable(startKeys.get(splitIndex)));<a name="line.254"></a>
+<span class="sourceLineNo">255</span>      }<a name="line.255"></a>
+<span class="sourceLineNo">256</span>    }<a name="line.256"></a>
+<span class="sourceLineNo">257</span><a name="line.257"></a>
+<span class="sourceLineNo">258</span>    void writePartitionFile(Configuration conf, Path path) throws IOException {<a name="line.258"></a>
+<span class="sourceLineNo">259</span>      FileSystem fs = path.getFileSystem(conf);<a name="line.259"></a>
+<span class="sourceLineNo">260</span>      @SuppressWarnings("deprecation")<a name="line.260"></a>
+<span class="sourceLineNo">261</span>      SequenceFile.Writer writer = SequenceFile.createWriter(<a name="line.261"></a>
+<span class="sourceLineNo">262</span>        fs, conf, path, ImmutableBytesWritable.class, NullWritable.class);<a name="line.262"></a>
+<span class="sourceLineNo">263</span><a name="line.263"></a>
+<span class="sourceLineNo">264</span>      for (int i = 0; i &lt; partitions.size(); i++) {<a name="line.264"></a>
+<span class="sourceLineNo">265</span>        writer.append(partitions.get(i), NullWritable.get());<a name="line.265"></a>
+<span class="sourceLineNo">266</span>      }<a name="line.266"></a>
+<span class="sourceLineNo">267</span>      writer.close();<a name="line.267"></a>
+<span class="sourceLineNo">268</span>    }<a name="line.268"></a>
+<span class="sourceLineNo">269</span><a name="line.269"></a>
+<span class="sourceLineNo">270</span>    private void readPartitionFile(FileSystem fs, Configuration conf, Path path)<a name="line.270"></a>
+<span class="sourceLineNo">271</span>         throws IOException {<a name="line.271"></a>
+<span class="sourceLineNo">272</span>      @SuppressWarnings("deprecation")<a name="line.272"></a>
+<span class="sourceLineNo">273</span>      SequenceFile.Reader reader = new SequenceFile.Reader(fs, path, conf);<a name="line.273"></a>
+<span class="sourceLineNo">274</span>      ImmutableBytesWritable key = new ImmutableBytesWritable();<a name="line.274"></a>
+<span class="sourceLineNo">275</span>      partitions = new ArrayList&lt;&gt;();<a name="line.275"></a>
+<span class="sourceLineNo">276</span>      while (reader.next(key)) {<a name="line.276"></a>
+<span class="sourceLineNo">277</span>        partitions.add(new ImmutableBytesWritable(key.copyBytes()));<a name="line.277"></a>
+<span class="sourceLineNo">278</span>      }<a name="line.278"></a>
+<span class="sourceLineNo">279</span>      reader.close();<a name="line.279"></a>
+<span class="sourceLineNo">280</span><a name="line.280"></a>
+<span class="sourceLineNo">281</span>      if (!Ordering.natural().isOrdered(partitions)) {<a name="line.281"></a>
+<span class="sourceLineNo">282</span>        throw new IOException("Partitions are not ordered!");<a name="line.282"></a>
+<span class="sourceLineNo">283</span>      }<a name="line.283"></a>
+<span class="sourceLineNo">284</span>    }<a name="line.284"></a>
+<span class="sourceLineNo">285</span><a name="line.285"></a>
+<span class="sourceLineNo">286</span>    @Override<a name="line.286"></a>
+<span class="sourceLineNo">287</span>    public String toString() {<a name="line.287"></a>
+<span class="sourceLineNo">288</span>      StringBuilder sb = new StringBuilder();<a name="line.288"></a>
+<span class="sourceLineNo">289</span>      sb.append("tableName=").append(tableName);<a name="line.289"></a>
+<span class="sourceLineNo">290</span>      if (families != null) {<a name="line.290"></a>
+<span class="sourceLineNo">291</span>        sb.append(", families=").append(families);<a name="line.291"></a>
+<span class="sourceLineNo">292</span>      }<a name="line.292"></a>
+<span class="sourceLineNo">293</span>      sb.append(", batchSize=").append(batchSize);<a name="line.293"></a>
+<span class="sourceLineNo">294</span>      sb.append(", numHashFiles=").append(numHashFiles);<a name="line.294"></a>
+<span class="sourceLineNo">295</span>      if (!isTableStartRow(startRow)) {<a name="line.295"></a>
+<span class="sourceLineNo">296</span>        sb.append(", startRowHex=").append(Bytes.toHex(startRow));<a name="line.296"></a>
+<span class="sourceLineNo">297</span>      }<a name="line.297"></a>
+<span class="sourceLineNo">298</span>      if (!isTableEndRow(stopRow)) {<a name="line.298"></a>
+<span class="sourceLineNo">299</span>        sb.append(", stopRowHex=").append(Bytes.toHex(stopRow));<a name="line.299"></a>
+<span class="sourceLineNo">300</span>      }<a name="line.300"></a>
+<span class="sourceLineNo">301</span>      if (scanBatch &gt;= 0) {<a name="line.301"></a>
+<span class="sourceLineNo">302</span>        sb.append(", scanBatch=").append(scanBatch);<a name="line.302"></a>
+<span class="sourceLineNo">303</span>      }<a name="line.303"></a>
+<span class="sourceLineNo">304</span>      if (versions &gt;= 0) {<a name="line.304"></a>
+<span class="sourceLineNo">305</span>        sb.append(", versions=").append(versions);<a name="line.305"></a>
+<span class="sourceLineNo">306</span>      }<a name="line.306"></a>
+<span class="sourceLineNo">307</span>      if (startTime != 0) {<a name="line.307"></a>
+<span class="sourceLineNo">308</span>        sb.append("startTime=").append(startTime);<a name="line.308"></a>
+<span class="sourceLineNo">309</span>      }<a name="line.309"></a>
+<span class="sourceLineNo">310</span>      if (endTime != 0) {<a name="line.310"></a>
+<span class="sourceLineNo">311</span>        sb.append("endTime=").append(endTime);<a name="line.311"></a>
+<span class="sourceLineNo">312</span>      }<a name="line.312"></a>
+<span class="sourceLineNo">313</span>      return sb.toString();<a name="line.313"></a>
+<span class="sourceLineNo">314</span>    }<a name="line.314"></a>
+<span class="sourceLineNo">315</span><a name="line.315"></a>
+<span class="sourceLineNo">316</span>    static String getDataFileName(int hashFileIndex) {<a name="line.316"></a>
+<span class="sourceLineNo">317</span>      return String.format(HashTable.OUTPUT_DATA_FILE_PREFIX + "%05d", hashFileIndex);<a name="line.317"></a>
+<span class="sourceLineNo">318</span>    }<a name="line.318"></a>
+<span class="sourceLineNo">319</span><a name="line.319"></a>
+<span class="sourceLineNo">320</span>    /**<a name="line.320"></a>
+<span class="sourceLineNo">321</span>     * Open a TableHash.Reader starting at the first hash at or after the given key.<a name="line.321"></a>
+<span class="sourceLineNo">322</span>     * @throws IOException<a name="line.322"></a>
+<span class="sourceLineNo">323</span>     */<a name="line.323"></a>
+<span class="sourceLineNo">324</span>    public Reader newReader(Configuration conf, ImmutableBytesWritable startKey)<a name="line.324"></a>
+<span class="sourceLineNo">325</span>        throws IOException {<a name="line.325"></a>
+<span class="sourceLineNo">326</span>      return new Reader(conf, startKey);<a name="line.326"></a>
+<span class="sourceLineNo">327</span>    }<a name="line.327"></a>
+<span class="sourceLineNo">328</span><a name="line.328"></a>
+<span class="sourceLineNo">329</span>    public class Reader implements java.io.Closeable {<a name="line.329"></a>
+<span class="sourceLineNo">330</span>      private final Configuration conf;<a name="line.330"></a>
+<span class="sourceLineNo">331</span><a name="line.331"></a>
+<span class="sourceLineNo">332</span>      private int hashFileIndex;<a name="line.332"></a>
+<span class="sourceLineNo">333</span>      private MapFile.Reader mapFileReader;<a name="line.333"></a>
+<span class="sourceLineNo">334</span><a name="line.334"></a>
+<span class="sourceLineNo">335</span>      private boolean cachedNext;<a name="line.335"></a>
+<span class="sourceLineNo">336</span>      private ImmutableBytesWritable key;<a name="line.336"></a>
+<span class="sourceLineNo">337</span>      private ImmutableBytesWritable hash;<a name="line.337"></a>
+<span class="sourceLineNo">338</span><a name="line.338"></a>
+<span class="sourceLineNo">339</span>      Reader(Configuration conf, ImmutableBytesWritable startKey) throws IOException {<a name="line.339"></a>
+<span class="sourceLineNo">340</span>        this.conf = conf;<a name="line.340"></a>
+<span class="sourceLineNo">341</span>        int partitionIndex = Collections.binarySearch(partitions, startKey);<a name="line.341"></a>
+<span class="sourceLineNo">342</span>        if (partitionIndex &gt;= 0) {<a name="line.342"></a>
+<span class="sourceLineNo">343</span>          // if the key is equal to a partition, then go the file after that partition<a name="line.343"></a>
+<span class="sourceLineNo">344</span>          hashFileIndex = partitionIndex+1;<a name="line.344"></a>
+<span class="sourceLineNo">345</span>        } else {<a name="line.345"></a>
+<span class="sourceLineNo">346</span>          // if the key is between partitions, then go to the file between those partitions<a name="line.346"></a>
+<span class="sourceLineNo">347</span>          hashFileIndex = -1-partitionIndex;<a name="line.347"></a>
+<span class="sourceLineNo">348</span>        }<a name="line.348"></a>
+<span class="sourceLineNo">349</span>        openHashFile();<a name="line.349"></a>
+<span class="sourceLineNo">350</span><a name="line.350"></a>
+<span class="sourceLineNo">351</span>        // MapFile's don't make it easy to seek() so that the subsequent next() returns<a name="line.351"></a>
+<span class="sourceLineNo">352</span>        // the desired key/value pair.  So we cache it for the first call of next().<a name="line.352"></a>
+<span class="sourceLineNo">353</span>        hash = new ImmutableBytesWritable();<a name="line.353"></a>
+<span class="sourceLineNo">354</span>        key = (ImmutableBytesWritable) mapFileReader.getClosest(startKey, hash);<a name="line.354"></a>
+<span class="sourceLineNo">355</span>        if (key == null) {<a name="line.355"></a>
+<span class="sourceLineNo">356</span>          cachedNext = false;<a name="line.356"></a>
+<span class="sourceLineNo">357</span>          hash = null;<a name="line.357"></a>
+<span class="sourceLineNo">358</span>        } else {<a name="line.358"></a>
+<span class="sourceLineNo">359</span>          cachedNext = true;<a name="line.359"></a>
+<span class="sourceLineNo">360</span>        }<a name="line.360"></a>
+<span class="sourceLineNo">361</span>      }<a name="line.361"></a>
+<span class="sourceLineNo">362</span><a name="line.362"></a>
+<span class="sourceLineNo">363</span>      /**<a name="line.363"></a>
+<span class="sourceLineNo">364</span>       * Read the next key/hash pair.<a name="line.364"></a>
+<span class="sourceLineNo">365</span>       * Returns true if such a pair exists and false when at the end of the data.<a name="line.365"></a>
+<span class="sourceLineNo">366</span>       */<a name="line.366"></a>
+<span class="sourceLineNo">367</span>      public boolean next() throws IOException {<a name="line.367"></a>
+<span class="sourceLineNo">368</span>        if (cachedNext) {<a name="line.368"></a>
+<span class="sourceLineNo">369</span>          cachedNext = false;<a name="line.369"></a>
+<span class="sourceLineNo">370</span>          return true;<a name="line.370"></a>
+<span class="sourceLineNo">371</span>        }<a name="line.371"></a>
+<span class="sourceLineNo">372</span>        key = new ImmutableBytesWritable();<a name="line.372"></a>
+<span class="sourceLineNo">373</span>        hash = new ImmutableBytesWritable();<a name="line.373"></a>
+<span class="sourceLineNo">374</span>        while (true) {<a name="line.374"></a>
+<span class="sourceLineNo">375</span>          boolean hasNext = mapFileReader.next(key, hash);<a name="line.375"></a>
+<span class="sourceLineNo">376</span>          if (hasNext) {<a name="line.376"></a>
+<span class="sourceLineNo">377</span>            return true;<a name="line.377"></a>
+<span class="sourceLineNo">378</span>          }<a name="line.378"></a>
+<span class="sourceLineNo">379</span>          hashFileIndex++;<a name="line.379"></a>
+<span class="sourceLineNo">380</span>          if (hashFileIndex &lt; TableHash.this.numHashFiles) {<a name="line.380"></a>
+<span class="sourceLineNo">381</span>            mapFileReader.close();<a name="line.381"></a>
+<span class="sourceLineNo">382</span>            openHashFile();<a name="line.382"></a>
+<span class="sourceLineNo">383</span>          } else {<a name="line.383"></a>
+<span class="sourceLineNo">384</span>            key = null;<a name="line.384"></a>
+<span class="sourceLineNo">385</span>            hash = null;<a name="line.385"></a>
+<span class="sourceLineNo">386</span>            return false;<a name="line.386"></a>
+<span class="sourceLineNo">387</span>          }<a name="line.387"></a>
+<span class="sourceLineNo">388</span>        }<a name="line.388"></a>
+<span class="sourceLineNo">389</span>      }<a name="line.389"></a>
+<span class="sourceLineNo">390</span><a name="line.390"></a>
+<span class="sourceLineNo">391</span>      /**<a name="line.391"></a>
+<span class="sourceLineNo">392</span>       * Get the current key<a name="line.392"></a>
+<span class="sourceLineNo">393</span>       * @return the current key or null if there is no current key<a name="line.393"></a>
+<span class="sourceLineNo">394</span>       */<a name="line.394"></a>
+<span class="sourceLineNo">395</span>      public ImmutableBytesWritable getCurrentKey() {<a name="line.395"></a>
+<span class="sourceLineNo">396</span>        return key;<a name="line.396"></a>
+<span class="sourceLineNo">397</span>      }<a name="line.397"></a>
+<span class="sourceLineNo">398</span><a name="line.398"></a>
+<span class="sourceLineNo">399</span>      /**<a name="line.399"></a>
+<span class="sourceLineNo">400</span>       * Get the current hash<a name="line.400"></a>
+<span class="sourceLineNo">401</span>       * @return the current hash or null if there is no current hash<a name="line.401"></a>
+<span class="sourceLineNo">402</span>       */<a name="line.402"></a>
+<span class="sourceLineNo">403</span>      public ImmutableBytesWritable getCurrentHash() {<a name="line.403"></a>
+<span class="sourceLineNo">404</span>        return hash;<a name="line.404"></a>
+<span class="sourceLineNo">405</span>      }<a name="line.405"></a>
+<span class="sourceLineNo">406</span><a name="line.406"></a>
+<span class="sourceLineNo">407</span>      private void openHashFile() throws IOException {<a name="line.407"></a>
+<span class="sourceLineNo">408</span>        if (mapFileReader != null) {<a name="line.408"></a>
+<span class="sourceLineNo">409</span>          mapFileReader.close();<a name="line.409"></a>
+<span class="sourceLineNo">410</span>        }<a name="line.410"></a>
+<span class="sourceLineNo">411</span>        Path dataDir = new Path(TableHash.this.hashDir, HASH_DATA_DIR);<a name="line.411"></a>
+<span class="sourceLineNo">412</span>        Path dataFile = new Path(dataDir, getDataFileName(hashFileIndex));<a name="line.412"></a>
+<span class="sourceLineNo">413</span>        mapFileReader = new MapFile.Reader(dataFile, conf);<a name="line.413"></a>
+<span class="sourceLineNo">414</span>      }<a name="line.414"></a>
+<span class="sourceLineNo">415</span><a name="line.415"></a>
+<span class="sourceLineNo">416</span>      @Override<a name="line.416"></a>
+<span class="sourceLineNo">417</span>      public void close() throws IOException {<a name="line.417"></a>
+<span class="sourceLineNo">418</span>        mapFileReader.close();<a name="line.418"></a>
+<span class="sourceLineNo">419</span>      }<a name="line.419"></a>
+<span class="sourceLineNo">420</span>    }<a name="line.420"></a>
+<span class="sourceLineNo">421</span>  }<a name="line.421"></a>
+<span class="sourceLineNo">422</span><a name="line.422"></a>
+<span class="sourceLineNo">423</span>  static boolean isTableStartRow(byte[] row) {<a name="line.423"></a>
+<span class="sourceLineNo">424</span>    return Bytes.equals(HConstants.EMPTY_START_ROW, row);<a name="line.424"></a>
+<span class="sourceLineNo">425</span>  }<a name="line.425"></a>
+<span class="sourceLineNo">426</span><a name="line.426"></a>
+<span class="sourceLineNo">427</span>  static boolean isTableEndRow(byte[] row) {<a name="line.427"></a>
+<span class="sourceLineNo">428</span>    return Bytes.equals(HConstants.EMPTY_END_ROW, row);<a name="line.428"></a>
+<span class="sourceLineNo">429</span>  }<a name="line.429"></a>
+<span class="sourceLineNo">430</span><a name="line.430"></a>
+<span class="sourceLineNo">431</span>  public Job createSubmittableJob(String[] args) throws IOException {<a name="line.431"></a>
+<span class="sourceLineNo">432</span>    Path partitionsPath = new Path(destPath, PARTITIONS_FILE_NAME);<a name="line.432"></a>
+<span class="sourceLineNo">433</span>    generatePartitions(partitionsPath);<a name="line.433"></a>
+<span class="sourceLineNo">434</span><a name="line.434"></a>
+<span class="sourceLineNo">435</span>    Job job = Job.getInstance(getConf(),<a name="line.435"></a>
+<span class="sourceLineNo">436</span>          getConf().get("mapreduce.job.name", "hashTable_" + tableHash.tableName));<a name="line.436"></a>
+<span class="sourceLineNo">437</span>    Configuration jobConf = job.getConfiguration();<a name="line.437"></a>
+<span class="sourceLineNo">438</span>    jobConf.setLong(HASH_BATCH_SIZE_CONF_KEY, tableHash.batchSize);<a name="line.438"></a>
+<span class="sourceLineNo">439</span>    jobConf.setBoolean(IGNORE_TIMESTAMPS, tableHash.ignoreTimestamps);<a name="line.439"></a>
+<span class="sourceLineNo">440</span>    job.setJarByClass(HashTable.class);<a name="line.440"></a>
 <span class="sourceLineNo">441</span><a name="line.441"></a>
-<span class="sourceLineNo">442</span>    // use a TotalOrderPartitioner and reducers to group region output into hash files<a name="line.442"></a>
-<span class="sourceLineNo">443</span>    job.setPartitionerClass(TotalOrderPartitioner.class);<a name="line.443"></a>
-<span class="sourceLineNo">444</span>    TotalOrderPartitioner.setPartitionFile(jobConf, partitionsPath);<a name="line.444"></a>
-<span class="sourceLineNo">445</span>    job.setReducerClass(Reducer.class);  // identity reducer<a name="line.445"></a>
-<span class="sourceLineNo">446</span>    job.setNumReduceTasks(tableHash.numHashFiles);<a name="line.446"></a>
-<span class="sourceLineNo">447</span>    job.setOutputKeyClass(ImmutableBytesWritable.class);<a name="line.447"></a>
-<span class="sourceLineNo">448</span>    job.setOutputValueClass(ImmutableBytesWritable.class);<a name="line.448"></a>
-<span class="sourceLineNo">449</span>    job.setOutputFormatClass(MapFileOutputFormat.class);<a name="line.449"></a>
-<span class="sourceLineNo">450</span>    FileOutputFormat.setOutputPath(job, new Path(destPath, HASH_DATA_DIR));<a name="line.450"></a>
-<span class="sourceLineNo">451</span><a name="line.451"></a>
-<span class="sourceLineNo">452</span>    return job;<a name="line.452"></a>
-<span class="sourceLineNo">453</span>  }<a name="line.453"></a>
+<span class="sourceLineNo">442</span>    TableMapReduceUtil.initTableMapperJob(tableHash.tableName, tableHash.initScan(),<a name="line.442"></a>
+<span class="sourceLineNo">443</span>        HashMapper.class, ImmutableBytesWritable.class, ImmutableBytesWritable.class, job);<a name="line.443"></a>
+<span class="sourceLineNo">444</span><a name="line.444"></a>
+<span class="sourceLineNo">445</span>    // use a TotalOrderPartitioner and reducers to group region output into hash files<a name="line.445"></a>
+<span class="sourceLineNo">446</span>    job.setPartitionerClass(TotalOrderPartitioner.class);<a name="line.446"></a>
+<span class="sourceLineNo">447</span>    TotalOrderPartitioner.setPartitionFile(jobConf, partitionsPath);<a name="line.447"></a>
+<span class="sourceLineNo">448</span>    job.setReducerClass(Reducer.class);  // identity reducer<a name="line.448"></a>
+<span class="sourceLineNo">449</span>    job.setNumReduceTasks(tableHash.numHashFiles);<a name="line.449"></a>
+<span class="sourceLineNo">450</span>    job.setOutputKeyClass(ImmutableBytesWritable.class);<a name="line.450"></a>
+<span class="sourceLineNo">451</span>    job.setOutputValueClass(ImmutableBytesWritable.class);<a name="line.451"></a>
+<span class="sourceLineNo">452</span>    job.setOutputFormatClass(MapFileOutputFormat.class);<a name="line.452"></a>
+<span class="sourceLineNo">453</span>    FileOutputFormat.setOutputPath(job, new Path(destPath, HASH_DATA_DIR));<a name="line.453"></a>
 <span class="sourceLineNo">454</span><a name="line.454"></a>
-<span class="sourceLineNo">455</span>  private void generatePartitions(Path partitionsPath) throws IOException {<a name="line.455"></a>
-<span class="sourceLineNo">456</span>    Connection connection = ConnectionFactory.createConnection(getConf());<a name="line.456"></a>
-<span class="sourceLineNo">457</span>    Pair&lt;byte[][], byte[][]&gt; regionKeys<a name="line.457"></a>
-<span class="sourceLineNo">458</span>      = connection.getRegionLocator(TableName.valueOf(tableHash.tableName)).getStartEndKeys();<a name="line.458"></a>
-<span class="sourceLineNo">459</span>    connection.close();<a name="line.459"></a>
-<span class="sourceLineNo">460</span><a name="line.460"></a>
-<span class="sourceLineNo">461</span>    tableHash.selectPartitions(regionKeys);<a name="line.461"></a>
-<span class="sourceLineNo">462</span>    LOG.info("Writing " + tableHash.partitions.size() + " partition keys to " + partitionsPath);<a name="line.462"></a>
+<span class="sourceLineNo">455</span>    return job;<a name="line.455"></a>
+<span class="sourceLineNo">456</span>  }<a name="line.456"></a>
+<span class="sourceLineNo">457</span><a name="line.457"></a>
+<span class="sourceLineNo">458</span>  private void generatePartitions(Path partitionsPath) throws IOException {<a name="line.458"></a>
+<span class="sourceLineNo">459</span>    Connection connection = ConnectionFactory.createConnection(getConf());<a name="line.459"></a>
+<span class="sourceLineNo">460</span>    Pair&lt;byte[][], byte[][]&gt; regionKeys<a name="line.460"></a>
+<span class="sourceLineNo">461</span>      = connection.getRegionLocator(TableName.valueOf(tableHash.tableName)).getStartEndKeys();<a name="line.461"></a>
+<span class="sourceLineNo">462</span>    connection.close();<a name="line.462"></a>
 <span class="sourceLineNo">463</span><a name="line.463"></a>
-<span class="sourceLineNo">464</span>    tableHash.writePartitionFile(getConf(), partitionsPath);<a name="line.464"></a>
-<span class="sourceLineNo">465</span>  }<a name="line.465"></a>
+<span class="sourceLineNo">464</span>    tableHash.selectPartitions(regionKeys);<a name="line.464"></a>
+<span class="sourceLineNo">465</span>    LOG.info("Writing " + tableHash.partitions.size() + " partition keys to " + partitionsPath);<a name="line.465"></a>
 <span class="sourceLineNo">466</span><a name="line.466"></a>
-<span class="sourceLineNo">467</span>  static class ResultHasher {<a name="line.467"></a>
-<span class="sourceLineNo">468</span>    private MessageDigest digest;<a name="line.468"></a>
+<span class="sourceLineNo">467</span>    tableHash.writePartitionFile(getConf(), partitionsPath);<a name="line.467"></a>
+<span class="sourceLineNo">468</span>  }<a name="line.468"></a>
 <span class="sourceLineNo">469</span><a name="line.469"></a>
-<span class="sourceLineNo">470</span>    private boolean batchStarted = false;<a name="line.470"></a>
-<span class="sourceLineNo">471</span>    private ImmutableBytesWritable batchStartKey;<a name="line.471"></a>
-<span class="sourceLineNo">472</span>    private ImmutableBytesWritable batchHash;<a name="line.472"></a>
-<span class="sourceLineNo">473</span>    private long batchSize = 0;<a name="line.473"></a>
-<span class="sourceLineNo">474</span><a name="line.474"></a>
-<span class="sourceLineNo">475</span><a name="line.475"></a>
-<span class="sourceLineNo">476</span>    public ResultHasher() {<a name="line.476"></a>
-<span class="sourceLineNo">477</span>      try {<a name="line.477"></a>
-<span class="sourceLineNo">478</span>        digest = MessageDigest.getInstance("MD5");<a name="line.478"></a>
-<span class="sourceLineNo">479</span>      } catch (NoSuchAlgorithmException e) {<a name="line.479"></a>
-<span class="sourceLineNo">480</span>        Throwables.propagate(e);<a name="line.480"></a>
-<span class="sourceLineNo">481</span>      }<a name="line.481"></a>
-<span class="sourceLineNo">482</span>    }<a name="line.482"></a>
-<span class="sourceLineNo">483</span><a name="line.483"></a>
-<span class="sourceLineNo">484</span>    public void startBatch(ImmutableBytesWritable row) {<a name="line.484"></a>
-<span class="sourceLineNo">485</span>      if (batchStarted) {<a name="line.485"></a>
-<span class="sourceLineNo">486</span>        throw new RuntimeException("Cannot start new batch without finishing existing one.");<a name="line.486"></a>
-<span class="sourceLineNo">487</span>      }<a name="line.487"></a>
-<span class="sourceLineNo">488</span>      batchStarted = true;<a name="line.488"></a>
-<span class="sourceLineNo">489</span>      batchSize = 0;<a name="line.489"></a>
-<span class="sourceLineNo">490</span>      batchStartKey = row;<a name="line.490"></a>
-<span class="sourceLineNo">491</span>      batchHash = null;<a name="line.491"></a>
-<span class="sourceLineNo">492</span>    }<a name="line.492"></a>
-<span class="sourceLineNo">493</span><a name="line.493"></a>
-<span class="sourceLineNo">494</span>    public void hashResult(Result result) {<a name="line.494"></a>
-<span class="sourceLineNo">495</span>      if (!batchStarted) {<a name="line.495"></a>
-<span class="sourceLineNo">496</span>        throw new RuntimeException("Cannot add to batch that has not been started.");<a name="line.496"></a>
-<span class="sourceLineNo">497</span>      }<a name="line.497"></a>
-<span class="sourceLineNo">498</span>      for (Cell cell : result.rawCells()) {<a name="line.498"></a>
-<span class="sourceLineNo">499</span>        int rowLength = cell.getRowLength();<a name="line.499"></a>
-<span class="sourceLineNo">500</span>        int familyLength = cell.getFamilyLength();<a name="line.500"></a>
-<span class="sourceLineNo">501</span>        int qualifierLength = cell.getQualifierLength();<a name="line.501"></a>
-<span class="sourceLineNo">502</span>        int valueLength = cell.getValueLength();<a name="line.502"></a>
-<span class="sourceLineNo">503</span>        digest.update(cell.getRowArray(), cell.getRowOffset(), rowLength);<a name="line.503"></a>
-<span class="sourceLineNo">504</span>        digest.update(cell.getFamilyArray(), cell.getFamilyOffset(), familyLength);<a name="line.504"></a>
-<span class="sourceLineNo">505</span>        digest.update(cell.getQualifierArray(), cell.getQualifierOffset(), qualifierLength);<a name="line.505"></a>
-<span class="sourceLineNo">506</span>        long ts = cell.getTimestamp();<a name="line.506"></a>
-<span class="sourceLineNo">507</span>        for (int i = 8; i &gt; 0; i--) {<a name="line.507"></a>
-<span class="sourceLineNo">508</span>          digest.update((byte) ts);<a name="line.508"></a>
-<span class="sourceLineNo">509</span>          ts &gt;&gt;&gt;= 8;<a name="line.509"></a>
-<span class="sourceLineNo">510</span>        }<a name="line.510"></a>
-<span class="sourceLineNo">511</span>        digest.update(cell.getValueArray(), cell.getValueOffset(), valueLength);<a name="line.511"></a>
-<span class="sourceLineNo">512</span><a name="line.512"></a>
-<span class="sourceLineNo">513</span>        batchSize += rowLength + familyLength + qualifierLength + 8 + valueLength;<a name="line.513"></a>
-<span class="sourceLineNo">514</span>      }<a name="line.514"></a>
-<span class="sourceLineNo">515</span>    }<a name="line.515"></a>
-<span class="sourceLineNo">516</span><a name="line.516"></a>
-<span class="sourceLineNo">517</span>    public void finishBatch() {<a name="line.517"></a>
-<span class="sourceLineNo">518</span>      if (!batchStarted) {<a name="line.518"></a>
-<span class="sourceLineNo">519</span>        throw new RuntimeException("Cannot finish batch that has not started.");<a name="line.519"></a>
-<span class="sourceLineNo">520</span>      }<a name="line.520"></a>
-<span class="sourceLineNo">521</span>      batchStarted = false;<a name="line.521"></a>
-<span class="sourceLineNo">522</span>      batchHash = new ImmutableBytesWritable(digest.digest());<a name="line.522"></a>
-<span class="sourceLineNo">523</span>    }<a name="line.523"></a>
-<span class="sourceLineNo">524</span><a name="line.524"></a>
-<span class="sourceLineNo">525</span>    public boolean isBatchStarted() {<a name="line.525"></a>
-<span class="sourceLineNo">526</span>      return batchStarted;<a name="line.526"></a>
-<span class="sourceLineNo">527</span>    }<a name="line.527"></a>
-<span class="sourceLineNo">528</span><a name="line.528"></a>
-<span class="sourceLineNo">529</span>    public ImmutableBytesWritable getBatchStartKey() {<a name="line.529"></a>
-<span class="sourceLineNo">530</span>      return batchStartKey;<a name="line.530"></a>
-<span class="sourceLineNo">531</span>    }<a name="line.531"></a>
-<span class="sourceLineNo">532</span><a name="line.532"></a>
-<span class="sourceLineNo">533</span>    public ImmutableBytesWritable getBatchHash() {<a name="line.533"></a>
-<span class="sourceLineNo">534</span>      return batchHash;<a name="line.534"></a>
-<span class="sourceLineNo">535</span>    }<a name="line.535"></a>
-<span class="sourceLineNo">536</span><a name="line.536"></a>
-<span class="sourceLineNo">537</span>    public long getBatchSize() {<a name="line.537"></a>
-<span class="sourceLineNo">538</span>      return batchSize;<a name="line.538"></a>
-<span class="sourceLineNo">539</span>    }<a name="line.539"></a>
-<span class="sourceLineNo">540</span>  }<a name="line.540"></a>
-<span class="sourceLineNo">541</span><a name="line.541"></a>
-<span class="sourceLineNo">542</span>  public static class HashMapper<a name="line.542"></a>
-<span class="sourceLineNo">543</span>    extends TableMapper&lt;ImmutableBytesWritable, ImmutableBytesWritable&gt; {<a name="line.543"></a>
-<span class="sourceLineNo">544</span><a name="line.544"></a>
-<span class="sourceLineNo">545</span>    private ResultHasher hasher;<a name="line.545"></a>
-<span class="sourceLineNo">546</span>    private long targetBatchSize;<a name="line.546"></a>
-<span class="sourceLineNo">547</span><a name="line.547"></a>
-<span class="sourceLineNo">548</span>    private ImmutableBytesWritable currentRow;<a name="line.548"></a>
-<span class="sourceLineNo">549</span><a name="line.549"></a>
-<span class="sourceLineNo">550</span>    @Override<a name="line.550"></a>
-<span class="sourceLineNo">551</span>    protected void setup(Context context) throws IOException, InterruptedException {<a name="line.551"></a>
-<span class="sourceLineNo">552</span>      targetBatchSize = context.getConfiguration()<a name="line.552"></a>
-<span class="sourceLineNo">553</span>          .getLong(HASH_BATCH_SIZE_CONF_KEY, DEFAULT_BATCH_SIZE);<a name="line.553"></a>
-<span class="sourceLineNo">554</span>      hasher = new ResultHasher();<a name="line.554"></a>
-<span class="sourceLineNo">555</span><a name="line.555"></a>
-<span class="sourceLineNo">556</span>      TableSplit split = (TableSplit) context.getInputSplit();<a name="line.556"></a>
-<span class="sourceLineNo">557</span>      hasher.startBatch(new ImmutableBytesWritable(split.getStartRow()));<a name="line.557"></a>
-<span class="sourceLineNo">558</span>    }<a name="line.558"></a>
-<span class="sourceLineNo">559</span><a name="line.559"></a>
-<span class="sourceLineNo">560</span>    @Override<a name="line.560"></a>
-<span class="sourceLineNo">561</span>    protected void map(ImmutableBytesWritable key, Result value, Context context)<a name="line.561"></a>
-<span class="sourceLineNo">562</span>        throws IOException, InterruptedException {<a name="line.562"></a>
-<span class="sourceLineNo">563</span><a name="line.563"></a>
-<span class="sourceLineNo">564</span>      if (currentRow == null || !currentRow.equals(key)) {<a name="line.564"></a>
-<span class="sourceLineNo">565</span>        currentRow = new ImmutableBytesWritable(key); // not immutable<a name="line.565"></a>
-<span class="sourceLineNo">566</span><a name="line.566"></a>
-<span class="sourceLineNo">567</span>        if (hasher.getBatchSize() &gt;= targetBatchSize) {<a name="line.567"></a>
-<span class="sourceLineNo">568</span>          hasher.finishBatch();<a name="line.568"></a>
-<span class="sourceLineNo">569</span>          context.write(hasher.getBatchStartKey(), hasher.getBatchHash());<a name="line.569"></a>
-<span class="sourceLineNo">570</span>          hasher.startBatch(currentRow);<a name="line.570"></a>
-<span class="sourceLineNo">571</span>        }<a name="line.571"></a>
-<span class="sourceLineNo">572</span>      }<a name="line.572"></a>
-<span class="sourceLineNo">573</span><a name="line.573"></a>
-<span class="sourceLineNo">574</span>      hasher.hashResult(value);<a name="line.574"></a>
-<span class="sourceLineNo">575</span>    }<a name="line.575"></a>
-<span class="sourceLineNo">576</span><a name="line.576"></a>
-<span class="sourceLineNo">577</span>    @Override<a name="line.577"></a>
-<span class="sourceLineNo">578</span>    protected void cleanup(Context context) throws IOException, InterruptedException {<a name="line.578"></a>
-<span class="sourceLineNo">579</span>      hasher.finishBatch();<a name="line.579"></a>
-<span class="sourceLineNo">580</span>      context.write(hasher.getBatchStartKey(), hasher.getBatchHash());<a name="line.580"></a>
-<span class="sourceLineNo">581</span>    }<a name="line.581"></a>
-<span class="sourceLineNo">582</span>  }<a name="line.582"></a>
-<span class="sourceLineNo">583</span><a name="line.583"></a>
-<span class="sourceLineNo">584</span>  private void writeTempManifestFile() throws IOException {<a name="line.584"></a>
-<span class="sourceLineNo">585</span>    Path tempManifestPath = new Path(destPath, TMP_MANIFEST_FILE_NAME);<a name="line.585"></a>
-<span class="sourceLineNo">586</span>    FileSystem fs = tempManifestPath.getFileSystem(getConf());<a name="line.586"></a>
-<span class="sourceLineNo">587</span>    tableHash.writePropertiesFile(fs, tempManifestPath);<a name="line.587"></a>
-<span class="sourceLineNo">588</span>  }<a name="line.588"></a>
-<span class="sourceLineNo">589</span><a name="line.589"></a>
-<span class="sourceLineNo">590</span>  private void completeManifest() throws IOException {<a name="line.590"></a>
-<span class="sourceLineNo">591</span>    Path tempManifestPath = new Path(destPath, TMP_MANIFEST_FILE_NAME);<a name="line.591"></a>
-<span class="sourceLineNo">592</span>    Path manifestPath = new Path(destPath, MANIFEST_FILE_NAME);<a name="line.592"></a>
-<span class="sourceLineNo">593</span>    FileSystem fs = tempManifestPath.getFileSystem(getConf());<a name="line.593"></a>
-<span class="sourceLineNo">594</span>    fs.rename(tempManifestPath, manifestPath);<a name="line.594"></a>
-<span class="sourceLineNo">595</span>  }<a name="line.595"></a>
-<span class="sourceLineNo">596</span><a name="line.596"></a>
-<span class="sourceLineNo">597</span>  private static final int NUM_ARGS = 2;<a name="line.597"></a>
-<span class="sourceLineNo">598</span>  private static void printUsage(final String errorMsg) {<a name="line.598"></a>
-<span class="sourceLineNo">599</span>    if (errorMsg != null &amp;&amp; errorMsg.length() &gt; 0) {<a name="line.599"></a>
-<span class="sourceLineNo">600</span>      System.err.println("ERROR: " + errorMsg);<a name="line.600"></a>
-<span class="sourceLineNo">601</span>      System.err.println();<a name="line.601"></a>
-<span class="sourceLineNo">602</span>    }<a name="line.602"></a>
-<span class="sourceLineNo">603</span>    System.err.println("Usage: HashTable [options] &lt;tablename&gt; &lt;outputpath&gt;");<a name="line.603"></a>
-<span class="sourceLineNo">604</span>    System.err.println();<a name="line.604"></a>
-<span class="sourceLineNo">605</span>    System.err.println("Options:");<a name="line.605"></a>
-<span class="sourceLineNo">606</span>    System.err.println(" batchsize     the target amount of bytes to hash in each batch");<a name="line.606"></a>
-<span class="sourceLineNo">607</span>    System.err.println("               rows are added to the batch until this size is reached");<a name="line.607"></a>
-<span class="sourceLineNo">608</span>    System.err.println("               (defaults to " + DEFAULT_BATCH_SIZE + " bytes)");<a name="line.608"></a>
-<span class="sourceLineNo">609</span>    System.err.println(" numhashfiles  the number of hash files to create");<a name="line.609"></a>
-<span class="sourceLineNo">610</span>    System.err.println("               if set to fewer than number of regions then");<a name="line.610"></a>
-<span class="sourceLineNo">611</span>    System.err.println("               the job will create this number of reducers");<a name="line.611"></a>
-<span class="sourceLineNo">612</span>    System.err.println("               (defaults to 1/100 of regions -- at least 1)");<a name="line.612"></a>
-<span class="sourceLineNo">613</span>    System.err.println(" startrow      the start row");<a name="line.613"></a>
-<span class="sourceLineNo">614</span>    System.err.println(" stoprow       the stop row");<a name="line.614"></a>
-<span class="sourceLineNo">615</span>    System.err.println(" starttime     beginning of the time range (unixtime in millis)");<a name="line.615"></a>
-<span class="sourceLineNo">616</span>    System.err.println("               without endtime means from starttime to forever");<a name="line.616"></a>
-<span class="sourceLineNo">617</span>    System.err.println(" endtime       end of the time range.  Ignored if no starttime specified.");<a name="line.617"></a>
-<span class="sourceLineNo">618</span>    System.err.println(" scanbatch     scanner batch size to support intra row scans");<a name="line.618"></a>
-<span class="sourceLineNo">619</span>    System.err.println(" versions      number of cell versions to include");<a name="line.619"></a>
-<span class="sourceLineNo">620</span>    System.err.println(" families      comma-separated list of families to include");<a name="line.620"></a>
-<span class="sourceLineNo">621</span>    System.err.println();<a name="line.621"></a>
-<span class="sourceLineNo">622</span>    System.err.println("Args:");<a name="line.622"></a>
-<span class="sourceLineNo">623</span>    System.err.println(" tablename     Name of the table to hash");<a name="line.623"></a>
-<span class="sourceLineNo">624</span>    System.err.println(" outputpath    Filesystem path to put the output data");<a name="line.624"></a>
-<span class="sourceLineNo">625</span>    System.err.println();<a name="line.625"></a>
-<span class="sourceLineNo">626</span>    System.err.println("Examples:");<a name="line.626"></a>
-<span class="sourceLineNo">627</span>    System.err.println(" To hash 'TestTable' in 32kB batches for a 1 hour window into 50 files:");<a name="line.627"></a>
-<span class="sourceLineNo">628</span>    System.err.println(" $ hbase " +<a name="line.628"></a>
-<span class="sourceLineNo">629</span>        "org.apache.hadoop.hbase.mapreduce.HashTable --batchsize=32000 --numhashfiles=50"<a name="line.629"></a>
-<span class="sourceLineNo">630</span>        + " --starttime=1265875194289 --endtime=1265878794289 --families=cf2,cf3"<a name="line.630"></a>
-<span class="sourceLineNo">631</span>        + " TestTable /hashes/testTable");<a name="line.631"></a>
-<span class="sourceLineNo">632</span>  }<a name="line.632"></a>
-<span class="sourceLineNo">633</span><a name="line.633"></a>
-<span class="sourceLineNo">634</span>  private boolean doCommandLine(final String[] args) {<a name="line.634"></a>
-<span class="sourceLineNo">635</span>    if (args.length &lt; NUM_ARGS) {<a name="line.635"></a>
-<span class="sourceLineNo">636</span>      printUsage(null);<a name="line.636"></a>
-<span class="sourceLineNo">637</span>      return false;<a name="line.637"></a>
-<span class="sourceLineNo">638</span>    }<a name="line.638"></a>
-<span class="sourceLineNo">639</span>    try {<a name="line.639"></a>
-<span class="sourceLineNo">640</span><a name="line.640"></a>
-<span class="sourceLineNo">641</span>      tableHash.tableName = args[args.length-2];<a name="line.641"></a>
-<span class="sourceLineNo">642</span>      destPath = new Path(args[args.length-1]);<a name="line.642"></a>
-<span class="sourceLineNo">643</span><a name="line.643"></a>
-<span class="sourceLineNo">644</span>      for (int i = 0; i &lt; args.length - NUM_ARGS; i++) {<a name="line.644"></a>
-<span class="sourceLineNo">645</span>        String cmd = args[i];<a name="line.645"></a>
-<span class="sourceLineNo">646</span>        if (cmd.equals("-h") || cmd.startsWith("--h")) {<a name="line.646"></a>
-<span class="sourceLineNo">647</span>          printUsage(null);<a name="line.647"></a>
-<span class="sourceLineNo">648</span>          return false;<a name="line.648"></a>
-<span class="sourceLineNo">649</span>        }<a name="line.649"></a>
-<span class="sourceLineNo">650</span><a name="line.650"></a>
-<span class="sourceLineNo">651</span>        final String batchSizeArgKey = "--batchsize=";<a name="line.651"></a>
-<span class="sourceLineNo">652</span>        if (cmd.startsWith(batchSizeArgKey)) {<a name="line.652"></a>
-<span class="sourceLineNo">653</span>          tableHash.batchSize = Long.parseLong(cmd.substring(batchSizeArgKey.length()));<a name="line.653"></a>
-<span class="sourceLineNo">654</span>          continue;<a name="line.654"></a>
-<span class="sourceLineNo">655</span>        }<a name="line.655"></a>
-<span class="sourceLineNo">656</span><a name="line.656"></a>
-<span class="sourceLineNo">657</span>        final String numHashFilesArgKey = "--numhashfiles=";<a name="line.657"></a>
-<span class="sourceLineNo">658</span>        if (cmd.startsWith(numHashFilesArgKey)) {<a name="line.658"></a>
-<span class="sourceLineNo">659</span>          tableHash.numHashFiles = Integer.parseInt(cmd.substring(numHashFilesArgKey.length()));<a name="line.659"></a>
-<span class="sourceLineNo">660</span>          continue;<a name="line.660"></a>
-<span class="sourceLineNo">661</span>        }<a name="line.661"></a>
-<span class="sourceLineNo">662</span><a name="line.662"></a>
-<span class="sourceLineNo">663</span>        final String startRowArgKey = "--startrow=";<a name="line.663"></a>
-<span class="sourceLineNo">664</span>        if (cmd.startsWith(startRowArgKey)) {<a name="line.664"></a>
-<span class="sourceLineNo">665</span>          tableHash.startRow = Bytes.fromHex(cmd.substring(startRowArgKey.length()));<a name="line.665"></a>
-<span class="sourceLineNo">666</span>          continue;<a name="line.666"></a>
-<span class="sourceLineNo">667</span>        }<a name="line.667"></a>
-<span class="sourceLineNo">668</span><a name="line.668"></a>
-<span class="sourceLineNo">669</span>        final String stopRowArgKey = "--stoprow=";<a name="line.669"></a>
-<span class="sourceLineNo">670</span>        if (cmd.startsWith(stopRowArgKey)) {<a name="line.670"></a>
-<span class="sourceLineNo">671</span>          tableHash.stopRow = Bytes.fromHex(cmd.substring(stopRowArgKey.length()));<a name="line.671"></a>
-<span class="sourceLineNo">672</span>          continue;<a name="line.672"></a>
-<span class="sourceLineNo">673</span>        }<a name="line.673"></a>
-<span class="sourceLineNo">674</span><a name="line.674"></a>
-<span class="sourceLineNo">675</span>        final String startTimeArgKey = "--starttime=";<a name="line.675"></a>
-<span class="sourceLineNo">676</span>        if (cmd.startsWith(startTimeArgKey)) {<a name="line.676"></a>
-<span class="sourceLineNo">677</span>          tableHash.startTime = Long.parseLong(cmd.substring(startTimeArgKey.length()));<a name="line.677"></a>
-<span class="sourceLineNo">678</span>          continue;<a name="line.678"></a>
-<span class="sourceLineNo">679</span>        }<a name="line.679"></a>
-<span class="sourceLineNo">680</span><a name="line.680"></a>
-<span class="sourceLineNo">681</span>        final String endTimeArgKey = "--endtime=";<a name="line.681"></a>
-<span class="sourceLineNo">682</span>        if (cmd.startsWith(endTimeArgKey)) {<a name="line.682"></a>
-<span class="sourceLineNo">683</span>          tableHash.endTime = Long.parseLong(cmd.substring(endTimeArgKey.length()));<a name="line.683"></a>
-<span class="sourceLineNo">684</span>          continue;<a name="line.684"></a>
-<span class="sourceLineNo">685</span>        }<a name="line.685"></a>
-<span class="sourceLineNo">686</span><a name="line.686"></a>
-<span class="sourceLineNo">687</span>        final String scanBatchArgKey = "--scanbatch=";<a name="line.687"></a>
-<span class="sourceLineNo">688</span>        if (cmd.startsWith(scanBatchArgKey)) {<a name="line.688"></a>
-<span class="sourceLineNo">689</span>          tableHash.scanBatch = Integer.parseInt(cmd.substring(scanBatchArgKey.length()));<a name="line.689"></a>
-<span class="sourceLineNo">690</span>          continue;<a name="line.690"></a>
-<span class="sourceLineNo">691</span>        }<a name="line.691"></a>
-<span class="sourceLineNo">692</span><a name="line.692"></a>
-<span class="sourceLineNo">693</span>        final String versionsArgKey = "--versions=";<a name="line.693"></a>
-<span class="sourceLineNo">694</span>        if (cmd.startsWith(versionsArgKey)) {<a name="line.694"></a>
-<span class="sourceLineNo">695</span>          tableHash.versions = Integer.parseInt(cmd.substring(versionsArgKey.length()));<a name="line.695"></a>
-<span class="sourceLineNo">696</span>          continue;<a name="line.696"></a>
-<span class="sourceLineNo">697</span>        }<a name="line.697"></a>
-<span class="sourceLineNo">698</span><a name="line.698"></a>
-<span class="sourceLineNo">699</span>        final String familiesArgKey = "--families=";<a name="line.699"></a>
-<span class="sourceLineNo">700</span>        if (cmd.startsWith(familiesArgKey)) {<a name="line.700"></a>
-<span class="sourceLineNo">701</span>          tableHash.families = cmd.substring(familiesArgKey.length());<a name="line.701"></a>
-<span class="sourceLineNo">702</span>          continue;<a name="line.702"></a>
-<span class="sourceLineNo">703</span>        }<a name="line.703"></a>
-<span class="sourceLineNo">704</span><a name="line.704"></a>
-<span class="sourceLineNo">705</span>        printUsage("Invalid argument '" + cmd + "'");<a name="line.705"></a>
-<span class="sourceLineNo">706</span>        return false;<a name="line.706"></a>
-<span class="sourceLineNo">707</span>      }<a name="line.707"></a>
-<span class="sourceLineNo">708</span>      if ((tableHash.startTime != 0 || tableHash.endTime != 0)<a name="line.708"></a>
-<span class="sourceLineNo">709</span>          &amp;&amp; (tableHash.startTime &gt;= tableHash.endTime)) {<a name="line.709"></a>
-<span class="sourceLineNo">710</span>        printUsage("Invalid time range filter: starttime="<a name="line.710"></a>
-<span class="sourceLineNo">711</span>            + tableHash.startTime + " &gt;=  endtime=" + tableHash.endTime);<a name="line.711"></a>
-<span class="sourceLineNo">712</span>        return false;<a name="line.712"></a>
-<span class="sourceLineNo">713</span>      }<a name="line.713"></a>
-<span class="sourceLineNo">714</span><a name="line.714"></a>
-<span class="sourceLineNo">715</span>    } catch (Exception e) {<a name="line.715"></a>
-<span class="sourceLineNo">716</span>      LOG.error("Failed to parse commandLine arguments", e);<a name="line.716"></a>
-<span class="sourceLineNo">717</span>      printUsage("Can't start because " + e.getMessage());<a name="line.717"></a>
-<span class="sourceLineNo">718</span>      return false;<a name="line.718"></a>
-<span class="sourceLineNo">719</span>    }<a name="line.719"></a>
-<span class="sourceLineNo">720</span>    return true;<a name="line.720"></a>
-<span class="sourceLineNo">721</span>  }<a name="line.721"></a>
+<span class="sourceLineNo">470</span>  static class ResultHasher {<a name="line.470"></a>
+<span class="sourceLineNo">471</span>    private MessageDigest digest;<a name="line.471"></a>
+<span class="sourceLineNo">472</span><a name="line.472"></a>
+<span class="sourceLineNo">473</span>    private boolean batchStarted = false;<a name="line.473"></a>
+<span class="sourceLineNo">474</span>    private ImmutableBytesWritable batchStartKey;<a name="line.474"></a>
+<span class="sourceLineNo">475</span>    private ImmutableBytesWritable batchHash;<a name="line.475"></a>
+<span class="sourceLineNo">476</span>    private long batchSize = 0;<a name="line.476"></a>
+<span class="sourceLineNo">477</span>    boolean ignoreTimestamps;<a name="line.477"></a>
+<span class="sourceLineNo">478</span><a name="line.478"></a>
+<span class="sourceLineNo">479</span><a name="line.479"></a>
+<span class="sourceLineNo">480</span>    public ResultHasher() {<a name="line.480"></a>
+<span class="sourceLineNo">481</span>      try {<a name="line.481"></a>
+<span class="sourceLineNo">482</span>        digest = MessageDigest.getInstance("MD5");<a name="line.482"></a>
+<span class="sourceLineNo">483</span>      } catch (NoSuchAlgorithmException e) {<a name="line.483"></a>
+<span class="sourceLineNo">484</span>        Throwables.propagate(e);<a name="line.484"></a>
+<span class="sourceLineNo">485</span>      }<a name="line.485"></a>
+<span class="sourceLineNo">486</span>    }<a name="line.486"></a>
+<span class="sourceLineNo">487</span><a name="line.487"></a>
+<span class="sourceLineNo">488</span>    public void startBatch(ImmutableBytesWritable row) {<a name="line.488"></a>
+<span class="sourceLineNo">489</span>      if (batchStarted) {<a name="line.489"></a>
+<span class="sourceLineNo">490</span>        throw new RuntimeException("Cannot start new batch without finishing existing one.");<a name="line.490"></a>
+<span class="sourceLineNo">491</span>      }<a name="line.491"></a>
+<span class="sourceLineNo">492</span>      batchStarted = true;<a name="line.492"></a>
+<span class="sourceLineNo">493</span>      batchSize = 0;<a name="line.493"></a>
+<span class="sourceLineNo">494</span>      batchStartKey = row;<a name="line.494"></a>
+<span class="sourceLineNo">495</span>      batchHash = null;<a name="line.495"></a>
+<span class="sourceLineNo">496</span>    }<a name="line.496"></a>
+<span class="sourceLineNo">497</span><a name="line.497"></a>
+<span class="sourceLineNo">498</span>    public void hashResult(Result result) {<a name="line.498"></a>
+<span class="sourceLineNo">499</span>      if (!batchStarted) {<a name="line.499"></a>
+<span class="sourceLineNo">500</span>        throw new RuntimeException("Cannot add to batch that has not been started.");<a name="line.500"></a>
+<span class="sourceLineNo">501</span>      }<a name="line.501"></a>
+<span class="sourceLineNo">502</span>      for (Cell cell : result.rawCells()) {<a name="line.502"></a>
+<span class="sourceLineNo">503</span>        int rowLength = cell.getRowLength();<a name="line.503"></a>
+<span class="sourceLineNo">504</span>        int familyLength = cell.getFamilyLength();<a name="line.504"></a>
+<span class="sourceLineNo">505</span>        int qualifierLength = cell.getQualifierLength();<a name="line.505"></a>
+<span class="sourceLineNo">506</span>        int valueLength = cell.getValueLength();<a name="line.506"></a>
+<span class="sourceLineNo">507</span>        digest.update(cell.getRowArray(), cell.getRowOffset(), rowLength);<a name="line.507"></a>
+<span class="sourceLineNo">508</span>        digest.update(cell.getFamilyArray(), cell.getFamilyOffset(), familyLength);<a name="line.508"></a>
+<span class="sourceLineNo">509</span>        digest.update(cell.getQualifierArray(), cell.getQualifierOffset(), qualifierLength);<a name="line.509"></a>
+<span class="sourceLineNo">510</span><a name="line.510"></a>
+<span class="sourceLineNo">511</span>        if (!ignoreTimestamps) {<a name="line.511"></a>
+<span class="sourceLineNo">512</span>          long ts = cell.getTimestamp();<a name="line.512"></a>
+<span class="sourceLineNo">513</span>          for (int i = 8; i &gt; 0; i--) {<a name="line.513"></a>
+<span class="sourceLineNo">514</span>            digest.update((byte) ts);<a name="line.514"></a>
+<span class="sourceLineNo">515</span>            ts &gt;&gt;&gt;= 8;<a name="line.515"></a>
+<span class="sourceLineNo">516</span>          }<a name="line.516"></a>
+<span class="sourceLineNo">517</span>        }<a name="line.517"></a>
+<span class="sourceLineNo">518</span>        digest.update(cell.getValueArray(), cell.getValueOffset(), valueLength);<a name="line.518"></a>
+<span class="sourceLineNo">519</span><a name="line.519"></a>
+<span class="sourceLineNo">520</span>        batchSize += rowLength + familyLength + qualifierLength + 8 + valueLength;<a name="line.520"></a>
+<span class="sourceLineNo">521</span>      }<a name="line.521"></a>
+<span class="sourceLineNo">522</span>    }<a name="line.522"></a>
+<span class="sourceLineNo">523</span><a name="line.523"></a>
+<span class="sourceLineNo">524</span>    public void finishBatch() {<a name="line.524"></a>
+<span class="sourceLineNo">525</span>      if (!batchStarted) {<a name="line.525"></a>
+<span class="sourceLineNo">526</span>        throw new RuntimeException("Cannot finish batch that has not started.");<a name="line.526"></a>
+<span class="sourceLineNo">527</span>      }<a name="line.527"></a>
+<span class="sourceLineNo">528</span>      batchStarted = false;<a name="line.528"></a>
+<span class="sourceLineNo">529</span>      batchHash = new ImmutableBytesWritable(digest.digest());<a name="line.529"></a>
+<span class="sourceLineNo">530</span>    }<a name="line.530"></a>
+<span class="sourceLineNo">531</span><a name="line.531"></a>
+<span class="sourceLineNo">532</span>    public boolean isBatchStarted() {<a name="line.532"></a>
+<span class="sourceLineNo">533</span>      return batchStarted;<a name="line.533"></a>
+<span class="sourceLineNo">534</span>    }<a name="line.534"></a>
+<span class="sourceLineNo">535</span><a name="line.535"></a>
+<span class="sourceLineNo">536</span>    public ImmutableBytesWritable getBatchStartKey() {<a name="line.536"></a>
+<span class="sourceLineNo">537</span>      return batchStartKey;<a name="line.537"></a>
+<span class="sourceLineNo">538</span>    }<a name="line.538"></a>
+<span class="sourceLineNo">539</span><a name="line.539"></a>
+<span class="sourceLineNo">540</span>    public ImmutableBytesWritable getBatchHash() {<a name="line.540"></a>
+<span class="sourceLineNo">541</span>      return batchHash;<a name="line.541"></a>
+<span class="sourceLineNo">542</span>    }<a name="line.542"></a>
+<span class="sourceLineNo">543</span><a name="line.543"></a>
+<span class="sourceLineNo">544</span>    public long getBatchSize() {<a name="line.544"></a>
+<span class="sourceLineNo">545</span>      return batchSize;<a name="line.545"></a>
+<span class="sourceLineNo">546</span>    }<a name="line.546"></a>
+<span class="sourceLineNo">547</span>  }<a name="line.547"></a>
+<span class="sourceLineNo">548</span><a name="line.548"></a>
+<span class="sourceLineNo">549</span>  public static class HashMapper<a name="line.549"></a>
+<span class="sourceLineNo">550</span>    extends TableMapper&lt;ImmutableBytesWritable, ImmutableBytesWritable&gt; {<a name="line.550"></a>
+<span class="sourceLineNo">551</span><a name="line.551"></a>
+<span class="sourceLineNo">552</span>    private ResultHasher hasher;<a name="line.552"></a>
+<span class="sourceLineNo">553</span>    private long targetBatchSize;<a name="line.553"></a>
+<span class="sourceLineNo">554</span><a name="line.554"></a>
+<span class="sourceLineNo">555</span>    private ImmutableBytesWritable currentRow;<a name="line.555"></a>
+<span class="sourceLineNo">556</span><a name="line.556"></a>
+<span class="sourceLineNo">557</span>    @Override<a name="line.557"></a>
+<span class="sourceLineNo">558</span>    protected void setup(Context context) throws IOException, InterruptedException {<a name="line.558"></a>
+<span class="sourceLineNo">559</span>      targetBatchSize = context.getConfiguration()<a name="line.559"></a>
+<span class="sourceLineNo">560</span>          .getLong(HASH_BATCH_SIZE_CONF_KEY, DEFAULT_BATCH_SIZE);<a name="line.560"></a>
+<span class="sourceLineNo">561</span>      hasher = new ResultHasher();<a name="line.561"></a>
+<span class="sourceLineNo">562</span>      hasher.ignoreTimestamps = context.getConfiguration().<a name="line.562"></a>
+<span class="sourceLineNo">563</span>        getBoolean(IGNORE_TIMESTAMPS, false);<a name="line.563"></a>
+<span class="sourceLineNo">564</span>      TableSplit split = (TableSplit) context.getInputSplit();<a name="line.564"></a>
+<span class="sourceLineNo">565</span>      hasher.startBatch(new ImmutableBytesWritable(split.getStartRow()));<a name="line.565"></a>
+<span class="sourceLineNo">566</span>    }<a name="line.566"></a>
+<span class="sourceLineNo">567</span><a name="line.567"></a>
+<span class="sourceLineNo">568</span>    @Override<a name="line.568"></a>
+<span class="sourceLineNo">569</span>    protected void map(ImmutableBytesWritable key, Result value, Context context)<a name="line.569"></a>
+<span class="sourceLineNo">570</span>        throws IOException, InterruptedException {<a name="line.570"></a>
+<span class="sourceLineNo">571</span><a name="line.571"></a>
+<span class="sourceLineNo">572</span>      if (currentRow == null || !currentRow.equals(key)) {<a name="line.572"></a>
+<span class="sourceLineNo">573</span>        currentRow = new ImmutableBytesWritable(key); // not immutable<a name="line.573"></a>
+<span class="sourceLineNo">574</span><a name="line.574"></a>
+<span class="sourceLineNo">575</span>        if (hasher.getBatchSize() &gt;= targetBatchSize) {<a name="line.575"></a>
+<span class="sourceLineNo">576</span>          hasher.finishBatch();<a name="line.576"></a>
+<span class="sourceLineNo">577</span>          context.write(hasher.getBatchStartKey(), hasher.getBatchHash());<a name="line.577"></a>
+<span class="sourceLineNo">578</span>          hasher.startBatch(currentRow);<a name="line.578"></a>
+<span class="sourceLineNo">579</span>        }<a name="line.579"></a>
+<span class="sourceLineNo">580</span>      }<a name="line.580"></a>
+<span class="sourceLineNo">581</span><a name="line.581"></a>
+<span class="sourceLineNo">582</span>      hasher.hashResult(value);<a name="line.582"></a>
+<span class="sourceLineNo">583</span>    }<a name="line.583"></a>
+<span class="sourceLineNo">584</span><a name="line.584"></a>
+<span class="sourceLineNo">585</span>    @Override<a name="line.585"></a>
+<span class="sourceLineNo">586</span>    protected void cleanup(Context context) throws IOException, InterruptedException {<a name="line.586"></a>
+<span class="sourceLineNo">587</span>      hasher.finishBatch();<a name="line.587"></a>
+<span class="sourceLineNo">588</span>      context.write(hasher.getBatchStartKey(), hasher.getBatchHash());<a name="line.588"></a>
+<span class="sourceLineNo">589</span>    }<a name="line.589"></a>
+<span class="sourceLineNo">590</span>  }<a name="line.590"></a>
+<span class="sourceLineNo">591</span><a name="line.591"></a>
+<span class="sourceLineNo">592</span>  private void writeTempManifestFile() throws IOException {<a name="line.592"></a>
+<span class="sourceLineNo">593</span>    Path tempManifestPath = new Path(destPath, TMP_MANIFEST_FILE_NAME);<a name="line.593"></a>
+<span class="sourceLineNo">594</span>    FileSystem fs = tempManifestPath.getFileSystem(getConf());<a name="line.594"></a>
+<span class="sourceLineNo">595</span>    tableHash.writePropertiesFile(fs, tempManifestPath);<a name="line.595"></a>
+<span class="sourceLineNo">596</span>  }<a name="line.596"></a>
+<span class="sourceLineNo">597</span><a name="line.597"></a>
+<span class="sourceLineNo">598</span>  private void completeManifest() throws IOException {<a name="line.598"></a>
+<span class="sourceLineNo">599</span>    Path tempManifestPath = new Path(destPath, TMP_MANIFEST_FILE_NAME);<a name="line.599"></a>
+<span class="sourceLineNo">600</span>    Path manifestPath = new Path(destPath, MANIFEST_FILE_NAME);<a name="line.600"></a>
+<span class="sourceLineNo">601</span>    FileSystem fs = tempManifestPath.getFileSystem(getConf());<a name="line.601"></a>
+<span class="sourceLineNo">602</span>    fs.rename(tempManifestPath, manifestPath);<a name="line.602"></a>
+<span class="sourceLineNo">603</span>  }<a name="line.603"></a>
+<span class="sourceLineNo">604</span><a name="line.604"></a>
+<span class="sourceLineNo">605</span>  private static final int NUM_ARGS = 2;<a name="line.605"></a>
+<span class="sourceLineNo">606</span>  private static void printUsage(final String errorMsg) {<a name="line.606"></a>
+<span class="sourceLineNo">607</span>    if (errorMsg != null &amp;&amp; errorMsg.length() &gt; 0) {<a name="line.607"></a>
+<span class="sourceLineNo">608</span>      System.err.println("ERROR: " + errorMsg);<a name="line.608"></a>
+<span class="sourceLineNo">609</span>      System.err.println();<a name="line.609"></a>
+<span class="sourceLineNo">610</span>    }<a name="line.610"></a>
+<span class="sourceLineNo">611</span>    System.err.println("Usage: HashTable [options] &lt;tablename&gt; &lt;outputpath&gt;");<a name="line.611"></a>
+<span class="sourceLineNo">612</span>    System.err.println();<a name="line.612"></a>
+<span class="sourceLineNo">613</span>    System.err.println("Options:");<a name="line.613"></a>
+<span class="sourceLineNo">614</span>    System.err.println(" batchsize         the target amount of bytes to hash in each batch");<a name="line.614"></a>
+<span class="sourceLineNo">615</span>    System.err.println("                   rows are added to the batch until this size is reached");<a name="line.615"></a>
+<span class="sourceLineNo">616</span>    System.err.println("                   (defaults to " + DEFAULT_BATCH_SIZE + " bytes)");<a name="line.616"></a>
+<span class="sourceLineNo">617</span>    System.err.println(" numhashfiles      the number of hash files to create");<a name="line.617"></a>
+<span class="sourceLineNo">618</span>    System.err.println("                   if set to fewer than number of regions then");<a name="line.618"></a>
+<span class="sourceLineNo">619</span>    System.err.println("                   the job will create this number of reducers");<a name="line.619"></a>
+<span class="sourceLineNo">620</span>    System.err.println("                   (defaults to 1/100 of regions -- at least 1)");<a name="line.620"></a>
+<span class="sourceLineNo">621</span>    System.err.println(" startrow          the start row");<a name="line.621"></a>
+<span class="sourceLineNo">622</span>    System.err.println(" stoprow           the stop row");<a name="line.622"></a>
+<span class="sourceLineNo">623</span>    System.err.println(" starttime         beginning of the time range (unixtime in millis)");<a name="line.623"></a>
+<span class="sourceLineNo">624</span>    System.err.println("                   without endtime means from starttime to forever");<a name="line.624"></a>
+<span class="sourceLineNo">625</span>    System.err.println(" endtime           end of the time range.");<a name="line.625"></a>
+<span class="sourceLineNo">626</span>    System.err.println("                   Ignored if no starttime specified.");<a name="line.626"></a>
+<span class="sourceLineNo">627</span>    System.err.println(" scanbatch         scanner batch size to support intra row scans");<a name="line.627"></a>
+<span class="sourceLineNo">628</span>    System.err.println(" versions          number of cell versions to include");<a name="line.628"></a>
+<span class="sourceLineNo">629</span>    System.err.println(" families          comma-separated list of families to include");<a name="line.629"></a>
+<span class="sourceLineNo">630</span>    System.err.println(" ignoreTimestamps  if true, ignores cell timestamps");<a name="line.630"></a>
+<span class="sourceLineNo">631</span>    System.err.println("                   when calculating hashes");<a name="line.631"></a>
+<span class="sourceLineNo">632</span>    System.err.println();<a name="line.632"></a>
+<span class="sourceLineNo">633</span>    System.err.println("Args:");<a name="line.633"></a>
+<span class="sourceLineNo">634</span>    System.err.println(" tablename     Name of the table to hash");<a name="line.634"></a>
+<span class="sourceLineNo">635</span>    System.err.println(" outputpath    Filesystem path to put the output data");<a name="line.635"></a>
+<span class="sourceLineNo">636</span>    System.err.println();<a name="line.636"></a>
+<span class="sourceLineNo">637</span>    System.err.println("Examples:");<a name="line.637"></a>
+<span class="sourceLineNo">638</span>    System.err.println(" To hash 'TestTable' in 32kB batches for a 1 hour window into 50 files:");<a name="line.638"></a>
+<span class="sourceLineNo">639</span>    System.err.println(" $ hbase " +<a name="line.639"></a>
+<span class="sourceLineNo">640</span>        "org.apache.hadoop.hbase.mapreduce.HashTable --batchsize=32000 --numhashfiles=50"<a name="line.640"></a>
+<span class="sourceLineNo">641</span>        + " --starttime=1265875194289 --endtime=1265878794289 --families=cf2,cf3"<a name="line.641"></a>
+<span class="sourceLineNo">642</span>        + " TestTable /hashes/testTable");<a name="line.642"></a>
+<span class="sourceLineNo">643</span>  }<a name="line.643"></a>
+<span class="sourceLineNo">644</span><a name="line.644"></a>
+<span class="sourceLineNo">645</span>  private boolean doCommandLine(final String[] args) {<a name="line.645"></a>
+<span class="sourceLineNo">646</span>    if (args.length &lt; NUM_ARGS) {<a name="line.646"></a>
+<span class="sourceLineNo">647</span>      printUsage(null);<a name="line.647"></a>
+<span class="sourceLineNo">648</span>      return false;<a name="line.648"></a>
+<span class="sourceLineNo">649</span>    }<a name="line.649"></a>
+<span class="sourceLineNo">650</span>    try {<a name="line.650"></a>
+<span class="sourceLineNo">651</span><a name="line.651"></a>
+<span class="sourceLineNo">652</span>      tableHash.tableName = args[args.length-2];<a name="line.652"></a>
+<span class="sourceLineNo">653</span>      destPath = new Path(args[args.length-1]);<a name="line.653"></a>
+<span class="sourceLineNo">654</span><a name="line.654"></a>
+<span class="sourceLineNo">655</span>      for (int i = 0; i &lt; args.length - NUM_ARGS; i++) {<a name="line.655"></a>
+<span class="sourceLineNo">656</span>        String cmd = args[i];<a name="line.656"></a>
+<span class="sourceLineNo">657</span>        if (cmd.equals("-h") || cmd.startsWith("--h")) {<a name="line.657"></a>
+<span class="sourceLineNo">658</span>          printUsage(null);<a name="line.658"></a>
+<span class="sourceLineNo">659</span>          return false;<a name="line.659"></a>
+<span class="sourceLineNo">660</span>        }<a name="line.660"></a>
+<span class="sourceLineNo">661</span><a name="line.661"></a>
+<span class="sourceLineNo">662</span>        final String batchSizeArgKey = "--batchsize=";<a name="line.662"></a>
+<span class="sourceLineNo">663</span>        if (cmd.startsWith(batchSizeArgKey)) {<a name="line.663"></a>
+<span class="sourceLineNo">664</span>          tableHash.batchSize = Long.parseLong(cmd.substring(batchSizeArgKey.length()));<a name="line.664"></a>
+<span class="sourceLineNo">665</span>          continue;<a name="line.665"></a>
+<span class="sourceLineNo">666</span>        }<a name="line.666"></a>
+<span class="sourceLineNo">667</span><a name="line.667"></a>
+<span class="sourceLineNo">668</span>        final String numHashFilesArgKey = "--numhashfiles=";<a name="line.668"></a>
+<span class="sourceLineNo">669</span>        if (cmd.startsWith(numHashFilesArgKey)) {<a name="line.669"></a>
+<span class="sourceLineNo">670</span>          tableHash.numHashFiles = Integer.parseInt(cmd.substring(numHashFilesArgKey.length()));<a name="line.670"></a>
+<span class="sourceLineNo">671</span>          continue;<a name="line.671"></a>
+<span class="sourceLineNo">672</span>        }<a name="line.672"></a>
+<span class="sourceLineNo">673</span><a name="line.673"></a>
+<span class="sourceLineNo">674</span>        final String startRowArgKey = "--startrow=";<a name="line.674"></a>
+<span class="sourceLineNo">675</span>        if (cmd.startsWith(startRowArgKey)) {<a name="line.675"></a>
+<span class="sourceLineNo">676</span>          tableHash.startRow = Bytes.fromHex(cmd.substring(startRowArgKey.length()));<a name="line.676"></a>
+<span class="sourceLineNo">677</span>          continue;<a name="line.677"></a>
+<span class="sourceLineNo">678</span>        }<a name="line.678"></a>
+<span class="sourceLineNo">679</span><a name="line.679"></a>
+<span class="sourceLineNo">680</span>        final String stopRowArgKey = "--stoprow=";<a name="line.680"></a>
+<span class="sourceLineNo">681</span>        if (cmd.startsWith(stopRowArgKey)) {<a name="line.681"></a>
+<span class="sourceLineNo">682</span>          tableHash.stopRow = Bytes.fromHex(cmd.substring(stopRowArgKey.length()));<a name="line.682"></a>
+<span class="sourceLineNo">683</span>          continue;<a name="line.683"></a>
+<span class="sourceLineNo">684</span>        }<a name="line.684"></a>
+<span class="sourceLineNo">685</span><a name="line.685"></a>
+<span class="sourceLineNo">686</span>        final String startTimeArgKey = "--starttime=";<a name="line.686"></a>
+<span class="sourceLineNo">687</span>        if (cmd.startsWith(startTimeArgKey)) {<a name="line.687"></a>
+<span class="sourceLineNo">688</span>          tableHash.startTime = Long.parseLong(cmd.substring(startTimeArgKey.length()));<a name="line.688"></a>
+<span class="sourceLineNo">689</span>          continue;<a name="line.689"></a>
+<span class="sourceLineNo">690</span>        }<a name="line.690"></a>
+<span class="sourceLineNo">691</span><a name="line.691"></a>
+<span class="sourceLineNo">692</span>        final String endTimeArgKey = "--endtime=";<a name="line.692"></a>
+<span class="sourceLineNo">693</span>        if (cmd.startsWith(endTimeArgKey)) {<a name="line.693"></a>
+<span class="sourceLineNo">694</span>          tableHash.endTime = Long.parseLong(cmd.substring(endTimeArgKey.length()));<a name="line.694"></a>
+<span class="sourceLineNo">695</span>          continue;<a name="line.695"></a>
+<span class="sourceLineNo">696</span>        }<a name="line.696"></a>
+<span class="sourceLineNo">697</span><a name="line.697"></a>
+<span class="sourceLineNo">698</span>        final String scanBatchArgKey = "--scanbatch=";<a name="line.698"></a>
+<span class="sourceLineNo">699</span>        if (cmd.startsWith(scanBatchArgKey)) {<a name="line.699"></a>
+<span class="sourceLineNo">700</span>          tableHash.scanBatch = Integer.parseInt(cmd.substring(scanBatchArgKey.length()));<a name="line.700"></a>
+<span class="sourceLineNo">701</span>          continue;<a name="line.701"></a>
+<span class="sourceLineNo">702</span>        }<a name="line.702"></a>
+<span class="sourceLineNo">703</span><a name="line.703"></a>
+<span class="sourceLineNo">704</span>        final String versionsArgKey = "--versions=";<a name="line.704"></a>
+<span class="sourceLineNo">705</span>        if (cmd.startsWith(versionsArgKey)) {<a name="line.705"></a>
+<span class="sourceLineNo">706</span>          tableHash.versions = Integer.parseInt(cmd.substring(versionsArgKey.length()));<a name="line.706"></a>
+<span class="sourceLineNo">707</span>          continue;<a name="line.707"></a>
+<span class="sourceLineNo">708</span>        }<a name="line.708"></a>
+<span class="sourceLineNo">709</span><a name="line.709"></a>
+<span class="sourceLineNo">710</span>        final String familiesArgKey = "--families=";<a name="line.710"></a>
+<span class="sourceLineNo">711</span>        if (cmd.startsWith(familiesArgKey)) {<a name="line.711"></a>
+<span class="sourceLineNo">712</span>          tableHash.families = cmd.substring(familiesArgKey.length());<a name="line.712"></a>
+<span class="sourceLineNo">713</span>          continue;<a name="line.713"></a>
+<span class="sourceLineNo">714</span>        }<a name="line.714"></a>
+<span class="sourceLineNo">715</span><a name="line.715"></a>
+<span class="sourceLineNo">716</span>        final String ignoreTimestampsKey = "--ignoreTimestamps=";<a name="line.716"></a>
+<span class="sourceLineNo">717</span>        if (cmd.startsWith(ignoreTimestampsKey)) {<a name="line.717"></a>
+<span class="sourceLineNo">718</span>          tableHash.ignoreTimestamps = Boolean.<a name="line.718"></a>
+<span class="sourceLineNo">719</span>            parseBoolean(cmd.substring(ignoreTimestampsKey.length()));<a name="line.719"></a>
+<span class="sourceLineNo">720</span>          continue;<a name="line.720"></a>
+<span class="sourceLineNo">721</span>        }<a name="line.721"></a>
 <span class="sourceLineNo">722</span><a name="line.722"></a>
-<span class="sourceLineNo">723</span>  /**<a name="line.723"></a>
-<span class="sourceLineNo">724</span>   * Main entry point.<a name="line.724"></a>
-<span class="sourceLineNo">725</span>   */<a name="line.725"></a>
-<span class="sourceLineNo">726</span>  public static void main(String[] args) throws Exception {<a name="line.726"></a>
-<span class="sourceLineNo">727</span>    int ret = ToolRunner.run(new HashTable(HBaseConfiguration.create()), args);<a name="line.727"></a>
-<span class="sourceLineNo">728</span>    System.exit(ret);<a name="line.728"></a>
-<span class="sourceLineNo">729</span>  }<a name="line.729"></a>
-<span class="sourceLineNo">730</span><a name="line.730"></a>
-<span class="sourceLineNo">731</span>  @Override<a name="line.731"></a>
-<span class="sourceLineNo">732</span>  public int run(String[] args) throws Exception {<a name="line.732"></a>
-<span class="sourceLineNo">733</span>    String[] otherArgs = new GenericOptionsParser(getConf(), args).getRemainingArgs();<a name="line.733"></a>
-<span class="sourceLineNo">734</span>    if (!doCommandLine(otherArgs)) {<a name="line.734"></a>
-<span class="sourceLineNo">735</span>      return 1;<a name="line.735"></a>
-<span class="sourceLineNo">736</span>    }<a name="line.736"></a>
-<span class="sourceLineNo">737</span><a name="line.737"></a>
-<span class="sourceLineNo">738</span>    Job job = createSubmittableJob(otherArgs);<a name="line.738"></a>
-<span class="sourceLineNo">739</span>    writeTempManifestFile();<a name="line.739"></a>
-<span class="sourceLineNo">740</span>    if (!job.waitForCompletion(true)) {<a name="line.740"></a>
-<span class="sourceLineNo">741</span>      LOG.info("Map-reduce job failed!");<a name="line.741"></a>
-<span class="sourceLineNo">742</span>      return 1;<a name="line.742"></a>
-<span class="sourceLineNo">743</span>    }<a name="line.743"></a>
-<span class="sourceLineNo">744</span>    completeManifest();<a name="line.744"></a>
-<span class="sourceLineNo">745</span>    return 0;<a name="line.745"></a>
-<span class="sourceLineNo">746</span>  }<a name="line.746"></a>
-<span class="sourceLineNo">747</span><a name="line.747"></a>
-<span class="sourceLineNo">748</span>}<a name="line.748"></a>
+<span class="sourceLineNo">723</span>        printUsage("Invalid argument '" + cmd + "'");<a name="line.723"></a>
+<span class="sourceLineNo">724</span>        return false;<a name="line.724"></a>
+<span class="sourceLineNo">725</span>      }<a name="line.725"></a>
+<span class="sourceLineNo">726</span>      if ((tableHash.startTime != 0 || tableHash.endTime != 0)<a name="line.726"></a>
+<span class="sourceLineNo">727</span>          &amp;&amp; (tableHash.startTime &gt;= tableHash.endTime)) {<a name="line.727"></a>
+<span class="sourceLineNo">728</span>        printUsage("Invalid time range filter: starttime="<a name="line.728"></a>
+<span class="sourceLineNo">729</span>            + tableHash.startTime + " &gt;=  endtime=" + tableHash.endTime);<a name="line.729"></a>
+<span class="sourceLineNo">730</span>        return false;<a name="line.730"></a>
+<span class="sourceLineNo">731</span>      }<a name="line.731"></a>
+<span class="sourceLineNo">732</span><a name="line.732"></a>
+<span class="sourceLineNo">733</span>    } catch (Exception e) {<a name="line.733"></a>
+<span class="sourceLineNo">734</span>      LOG.error("Failed to parse commandLine arguments", e);<a name="line.734"></a>
+<span class="sourceLineNo">735</span>      printUsage("Can't start because " + e.getMessage());<a name="line.735"></a>
+<span class="sourceLineNo">736</span>      return false;<a name="line.736"></a>
+<span class="sourceLineNo">737</span>    }<a name="line.737"></a>
+<span class="sourceLineNo">738</span>    return true;<a name="line.738"></a>
+<span class="sourceLineNo">739</span>  }<a name="line.739"></a>
+<span class="sourceLineNo">740</span><a name="line.740"></a>
+<span class="sourceLineNo">741</span>  /**<a name="line.741"></a>
+<span class="sourceLineNo">742</span>   * Main entry point.<a name="line.742"></a>
+<span class="sourceLineNo">743</span>   */<a name="line.743"></a>
+<span class="sourceLineNo">744</span>  public static void main(String[] args) throws Exception {<a name="line.744"></a>
+<span class="sourceLineNo">745</span>    int ret = ToolRunner.run(new HashTable(HBaseConfiguration.create()), args);<a name="line.745"></a>
+<span class="sourceLineNo">746</span>    System.exit(ret);<a name="line.746"></a>
+<span class="sourceLineNo">747</span>  }<a name="line.747"></a>
+<span class="sourceLineNo">748</span><a name="line.748"></a>
+<span class="sourceLineNo">749</span>  @Override<a name="line.749"></a>
+<span class="sourceLineNo">750</span>  public int run(String[] args) throws Exception {<a name="line.750"></a>
+<span class="sourceLineNo">751</span>    String[] otherArgs = new GenericOptionsParser(getConf(), args).getRemainingArgs();<a name="line.751"></a>
+<span class="sourceLineNo">752</span>    if (!doCommandLine(otherArgs)) {<a name="line.752"></a>
+<span class="sourceLineNo">753</span>      return 1;<a name="line.753"></a>
+<span class="sourceLineNo">754</span>    }<a name="line.754"></a>
+<span class="sourceLineNo">755</span><a name="line.755"></a>
+<span class="sourceLineNo">756</span>    Job job = createSubmittableJob(otherArgs);<a name="line.756"></a>
+<span class="sourceLineNo">757</span>    writeTempManifestFile();<a name="line.757"></a>
+<span class="sourceLineNo">758</span>    if (!job.waitForCompletion(true)) {<a name="line.758"></a>
+<span class="sourceLineNo">759</span>      LOG.info("Map-reduce job failed!");<a name="line.759"></a>
+<span class="sourceLineNo">760</span>      return 1;<a name="line.760"></a>
+<span class="sourceLineNo">761</span>    }<a name="line.761"></a>
+<span class="sourceLineNo">762</span>    completeManifest();<a name="line.762"></a>
+<span class="sourceLineNo">763</span>    return 0;<a name="line.763"></a>
+<span class="sourceLineNo">764</span>  }<a name="line.764"></a>
+<span class="sourceLineNo">765</span><a name="line.765"></a>
+<span class="sourceLineNo">766</span>}<a name="line.766"></a>
 
 
 
diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html
index 55fc1c9..32934c7 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.Reader.html
@@ -81,679 +81,697 @@
 <span class="sourceLineNo">073</span>  final static String MANIFEST_FILE_NAME = "manifest";<a name="line.73"></a>
 <span class="sourceLineNo">074</span>  final static String HASH_DATA_DIR = "hashes";<a name="line.74"></a>
 <span class="sourceLineNo">075</span>  final static String OUTPUT_DATA_FILE_PREFIX = "part-r-";<a name="line.75"></a>
-<span class="sourceLineNo">076</span>  private final static String TMP_MANIFEST_FILE_NAME = "manifest.tmp";<a name="line.76"></a>
-<span class="sourceLineNo">077</span><a name="line.77"></a>
-<span class="sourceLineNo">078</span>  TableHash tableHash = new TableHash();<a name="line.78"></a>
-<span class="sourceLineNo">079</span>  Path destPath;<a name="line.79"></a>
-<span class="sourceLineNo">080</span><a name="line.80"></a>
-<span class="sourceLineNo">081</span>  public HashTable(Configuration conf) {<a name="line.81"></a>
-<span class="sourceLineNo">082</span>    super(conf);<a name="line.82"></a>
-<span class="sourceLineNo">083</span>  }<a name="line.83"></a>
-<span class="sourceLineNo">084</span><a name="line.84"></a>
-<span class="sourceLineNo">085</span>  public static class TableHash {<a name="line.85"></a>
-<span class="sourceLineNo">086</span><a name="line.86"></a>
-<span class="sourceLineNo">087</span>    Path hashDir;<a name="line.87"></a>
-<span class="sourceLineNo">088</span><a name="line.88"></a>
-<span class="sourceLineNo">089</span>    String tableName;<a name="line.89"></a>
-<span class="sourceLineNo">090</span>    String families = null;<a name="line.90"></a>
-<span class="sourceLineNo">091</span>    long batchSize = DEFAULT_BATCH_SIZE;<a name="line.91"></a>
-<span class="sourceLineNo">092</span>    int numHashFiles = 0;<a name="line.92"></a>
-<span class="sourceLineNo">093</span>    byte[] startRow = HConstants.EMPTY_START_ROW;<a name="line.93"></a>
-<span class="sourceLineNo">094</span>    byte[] stopRow = HConstants.EMPTY_END_ROW;<a name="line.94"></a>
-<span class="sourceLineNo">095</span>    int scanBatch = 0;<a name="line.95"></a>
-<span class="sourceLineNo">096</span>    int versions = -1;<a name="line.96"></a>
-<span class="sourceLineNo">097</span>    long startTime = 0;<a name="line.97"></a>
-<span class="sourceLineNo">098</span>    long endTime = 0;<a name="line.98"></a>
-<span class="sourceLineNo">099</span><a name="line.99"></a>
-<span class="sourceLineNo">100</span>    List&lt;ImmutableBytesWritable&gt; partitions;<a name="line.100"></a>
+<span class="sourceLineNo">076</span>  final static String IGNORE_TIMESTAMPS = "ignoreTimestamps";<a name="line.76"></a>
+<span class="sourceLineNo">077</span>  private final static String TMP_MANIFEST_FILE_NAME = "manifest.tmp";<a name="line.77"></a>
+<span class="sourceLineNo">078</span><a name="line.78"></a>
+<span class="sourceLineNo">079</span>  TableHash tableHash = new TableHash();<a name="line.79"></a>
+<span class="sourceLineNo">080</span>  Path destPath;<a name="line.80"></a>
+<span class="sourceLineNo">081</span><a name="line.81"></a>
+<span class="sourceLineNo">082</span>  public HashTable(Configuration conf) {<a name="line.82"></a>
+<span class="sourceLineNo">083</span>    super(conf);<a name="line.83"></a>
+<span class="sourceLineNo">084</span>  }<a name="line.84"></a>
+<span class="sourceLineNo">085</span><a name="line.85"></a>
+<span class="sourceLineNo">086</span>  public static class TableHash {<a name="line.86"></a>
+<span class="sourceLineNo">087</span><a name="line.87"></a>
+<span class="sourceLineNo">088</span>    Path hashDir;<a name="line.88"></a>
+<span class="sourceLineNo">089</span><a name="line.89"></a>
+<span class="sourceLineNo">090</span>    String tableName;<a name="line.90"></a>
+<span class="sourceLineNo">091</span>    String families = null;<a name="line.91"></a>
+<span class="sourceLineNo">092</span>    long batchSize = DEFAULT_BATCH_SIZE;<a name="line.92"></a>
+<span class="sourceLineNo">093</span>    int numHashFiles = 0;<a name="line.93"></a>
+<span class="sourceLineNo">094</span>    byte[] startRow = HConstants.EMPTY_START_ROW;<a name="line.94"></a>
+<span class="sourceLineNo">095</span>    byte[] stopRow = HConstants.EMPTY_END_ROW;<a name="line.95"></a>
+<span class="sourceLineNo">096</span>    int scanBatch = 0;<a name="line.96"></a>
+<span class="sourceLineNo">097</span>    int versions = -1;<a name="line.97"></a>
+<span class="sourceLineNo">098</span>    long startTime = 0;<a name="line.98"></a>
+<span class="sourceLineNo">099</span>    long endTime = 0;<a name="line.99"></a>
+<span class="sourceLineNo">100</span>    boolean ignoreTimestamps;<a name="line.100"></a>
 <span class="sourceLineNo">101</span><a name="line.101"></a>
-<span class="sourceLineNo">102</span>    public static TableHash read(Configuration conf, Path hashDir) throws IOException {<a name="line.102"></a>
-<span class="sourceLineNo">103</span>      TableHash tableHash = new TableHash();<a name="line.103"></a>
-<span class="sourceLineNo">104</span>      FileSystem fs = hashDir.getFileSystem(conf);<a name="line.104"></a>
-<span class="sourceLineNo">105</span>      tableHash.hashDir = hashDir;<a name="line.105"></a>
-<span class="sourceLineNo">106</span>      tableHash.readPropertiesFile(fs, new Path(hashDir, MANIFEST_FILE_NAME));<a name="line.106"></a>
-<span class="sourceLineNo">107</span>      tableHash.readPartitionFile(fs, conf, new Path(hashDir, PARTITIONS_FILE_NAME));<a name="line.107"></a>
-<span class="sourceLineNo">108</span>      return tableHash;<a name="line.108"></a>
-<span class="sourceLineNo">109</span>    }<a name="line.109"></a>
-<span class="sourceLineNo">110</span><a name="line.110"></a>
-<span class="sourceLineNo">111</span>    void writePropertiesFile(FileSystem fs, Path path) throws IOException {<a name="line.111"></a>
-<span class="sourceLineNo">112</span>      Properties p = new Properties();<a name="line.112"></a>
-<span class="sourceLineNo">113</span>      p.setProperty("table", tableName);<a name="line.113"></a>
-<span class="sourceLineNo">114</span>      if (families != null) {<a name="line.114"></a>
-<span class="sourceLineNo">115</span>        p.setProperty("columnFamilies", families);<a name="line.115"></a>
-<span class="sourceLineNo">116</span>      }<a name="line.116"></a>
-<span class="sourceLineNo">117</span>      p.setProperty("targetBatchSize", Long.toString(batchSize));<a name="line.117"></a>
-<span class="sourceLineNo">118</span>      p.setProperty("numHashFiles", Integer.toString(numHashFiles));<a name="line.118"></a>
-<span class="sourceLineNo">119</span>      if (!isTableStartRow(startRow)) {<a name="line.119"></a>
-<span class="sourceLineNo">120</span>        p.setProperty("startRowHex", Bytes.toHex(startRow));<a name="line.120"></a>
-<span class="sourceLineNo">121</span>      }<a name="line.121"></a>
-<span class="sourceLineNo">122</span>      if (!isTableEndRow(stopRow)) {<a name="line.122"></a>
-<span class="sourceLineNo">123</span>        p.setProperty("stopRowHex", Bytes.toHex(stopRow));<a name="line.123"></a>
-<span class="sourceLineNo">124</span>      }<a name="line.124"></a>
-<span class="sourceLineNo">125</span>      if (scanBatch &gt; 0) {<a name="line.125"></a>
-<span class="sourceLineNo">126</span>        p.setProperty("scanBatch", Integer.toString(scanBatch));<a name="line.126"></a>
-<span class="sourceLineNo">127</span>      }<a name="line.127"></a>
-<span class="sourceLineNo">128</span>      if (versions &gt;= 0) {<a name="line.128"></a>
-<span class="sourceLineNo">129</span>        p.setProperty("versions", Integer.toString(versions));<a name="line.129"></a>
-<span class="sourceLineNo">130</span>      }<a name="line.130"></a>
-<span class="sourceLineNo">131</span>      if (startTime != 0) {<a name="line.131"></a>
-<span class="sourceLineNo">132</span>        p.setProperty("startTimestamp", Long.toString(startTime));<a name="line.132"></a>
-<span class="sourceLineNo">133</span>      }<a name="line.133"></a>
-<span class="sourceLineNo">134</span>      if (endTime != 0) {<a name="line.134"></a>
-<span class="sourceLineNo">135</span>        p.setProperty("endTimestamp", Long.toString(endTime));<a name="line.135"></a>
-<span class="sourceLineNo">136</span>      }<a name="line.136"></a>
-<span class="sourceLineNo">137</span><a name="line.137"></a>
-<span class="sourceLineNo">138</span>      try (OutputStreamWriter osw = new OutputStreamWriter(fs.create(path), Charsets.UTF_8)) {<a name="line.138"></a>
-<span class="sourceLineNo">139</span>        p.store(osw, null);<a name="line.139"></a>
-<span class="sourceLineNo">140</span>      }<a name="line.140"></a>
-<span class="sourceLineNo">141</span>    }<a name="line.141"></a>
-<span class="sourceLineNo">142</span><a name="line.142"></a>
-<span class="sourceLineNo">143</span>    void readPropertiesFile(FileSystem fs, Path path) throws IOException {<a name="line.143"></a>
-<span class="sourceLineNo">144</span>      Properties p = new Properties();<a name="line.144"></a>
-<span class="sourceLineNo">145</span>      try (FSDataInputStream in = fs.open(path)) {<a name="line.145"></a>
-<span class="sourceLineNo">146</span>        try (InputStreamReader isr = new InputStreamReader(in, Charsets.UTF_8)) {<a name="line.146"></a>
-<span class="sourceLineNo">147</span>          p.load(isr);<a name="line.147"></a>
-<span class="sourceLineNo">148</span>        }<a name="line.148"></a>
-<span class="sourceLineNo">149</span>      }<a name="line.149"></a>
-<span class="sourceLineNo">150</span>      tableName = p.getProperty("table");<a name="line.150"></a>
-<span class="sourceLineNo">151</span>      families = p.getProperty("columnFamilies");<a name="line.151"></a>
-<span class="sourceLineNo">152</span>      batchSize = Long.parseLong(p.getProperty("targetBatchSize"));<a name="line.152"></a>
-<span class="sourceLineNo">153</span>      numHashFiles = Integer.parseInt(p.getProperty("numHashFiles"));<a name="line.153"></a>
-<span class="sourceLineNo">154</span><a name="line.154"></a>
-<span class="sourceLineNo">155</span>      String startRowHex = p.getProperty("startRowHex");<a name="line.155"></a>
-<span class="sourceLineNo">156</span>      if (startRowHex != null) {<a name="line.156"></a>
-<span class="sourceLineNo">157</span>        startRow = Bytes.fromHex(startRowHex);<a name="line.157"></a>
-<span class="sourceLineNo">158</span>      }<a name="line.158"></a>
-<span class="sourceLineNo">159</span>      String stopRowHex = p.getProperty("stopRowHex");<a name="line.159"></a>
-<span class="sourceLineNo">160</span>      if (stopRowHex != null) {<a name="line.160"></a>
-<span class="sourceLineNo">161</span>        stopRow = Bytes.fromHex(stopRowHex);<a name="line.161"></a>
-<span class="sourceLineNo">162</span>      }<a name="line.162"></a>
-<span class="sourceLineNo">163</span><a name="line.163"></a>
-<span class="sourceLineNo">164</span>      String scanBatchString = p.getProperty("scanBatch");<a name="line.164"></a>
-<span class="sourceLineNo">165</span>      if (scanBatchString != null) {<a name="line.165"></a>
-<span class="sourceLineNo">166</span>        scanBatch = Integer.parseInt(scanBatchString);<a name="line.166"></a>
-<span class="sourceLineNo">167</span>      }<a name="line.167"></a>
-<span class="sourceLineNo">168</span><a name="line.168"></a>
-<span class="sourceLineNo">169</span>      String versionString = p.getProperty("versions");<a name="line.169"></a>
-<span class="sourceLineNo">170</span>      if (versionString != null) {<a name="line.170"></a>
-<span class="sourceLineNo">171</span>        versions = Integer.parseInt(versionString);<a name="line.171"></a>
-<span class="sourceLineNo">172</span>      }<a name="line.172"></a>
-<span class="sourceLineNo">173</span><a name="line.173"></a>
-<span class="sourceLineNo">174</span>      String startTimeString = p.getProperty("startTimestamp");<a name="line.174"></a>
-<span class="sourceLineNo">175</span>      if (startTimeString != null) {<a name="line.175"></a>
-<span class="sourceLineNo">176</span>        startTime = Long.parseLong(startTimeString);<a name="line.176"></a>
-<span class="sourceLineNo">177</span>      }<a name="line.177"></a>
-<span class="sourceLineNo">178</span><a name="line.178"></a>
-<span class="sourceLineNo">179</span>      String endTimeString = p.getProperty("endTimestamp");<a name="line.179"></a>
-<span class="sourceLineNo">180</span>      if (endTimeString != null) {<a name="line.180"></a>
-<span class="sourceLineNo">181</span>        endTime = Long.parseLong(endTimeString);<a name="line.181"></a>
-<span class="sourceLineNo">182</span>      }<a name="line.182"></a>
-<span class="sourceLineNo">183</span>    }<a name="line.183"></a>
-<span class="sourceLineNo">184</span><a name="line.184"></a>
-<span class="sourceLineNo">185</span>    Scan initScan() throws IOException {<a name="line.185"></a>
-<span class="sourceLineNo">186</span>      Scan scan = new Scan();<a name="line.186"></a>
-<span class="sourceLineNo">187</span>      scan.setCacheBlocks(false);<a name="line.187"></a>
-<span class="sourceLineNo">188</span>      if (startTime != 0 || endTime != 0) {<a name="line.188"></a>
-<span class="sourceLineNo">189</span>        scan.setTimeRange(startTime, endTime == 0 ? HConstants.LATEST_TIMESTAMP : endTime);<a name="line.189"></a>
-<span class="sourceLineNo">190</span>      }<a name="line.190"></a>
-<span class="sourceLineNo">191</span>      if (scanBatch &gt; 0) {<a name="line.191"></a>
-<span class="sourceLineNo">192</span>        scan.setBatch(scanBatch);<a name="line.192"></a>
-<span class="sourceLineNo">193</span>      }<a name="line.193"></a>
-<span class="sourceLineNo">194</span>      if (versions &gt;= 0) {<a name="line.194"></a>
-<span class="sourceLineNo">195</span>        scan.readVersions(versions);<a name="line.195"></a>
-<span class="sourceLineNo">196</span>      }<a name="line.196"></a>
-<span class="sourceLineNo">197</span>      if (!isTableStartRow(startRow)) {<a name="line.197"></a>
-<span class="sourceLineNo">198</span>        scan.withStartRow(startRow);<a name="line.198"></a>
-<span class="sourceLineNo">199</span>      }<a name="line.199"></a>
-<span class="sourceLineNo">200</span>      if (!isTableEndRow(stopRow)) {<a name="line.200"></a>
-<span class="sourceLineNo">201</span>        scan.withStopRow(stopRow);<a name="line.201"></a>
-<span class="sourceLineNo">202</span>      }<a name="line.202"></a>
-<span class="sourceLineNo">203</span>      if(families != null) {<a name="line.203"></a>
-<span class="sourceLineNo">204</span>        for(String fam : families.split(",")) {<a name="line.204"></a>
-<span class="sourceLineNo">205</span>          scan.addFamily(Bytes.toBytes(fam));<a name="line.205"></a>
-<span class="sourceLineNo">206</span>        }<a name="line.206"></a>
-<span class="sourceLineNo">207</span>      }<a name="line.207"></a>
-<span class="sourceLineNo">208</span>      return scan;<a name="line.208"></a>
-<span class="sourceLineNo">209</span>    }<a name="line.209"></a>
-<span class="sourceLineNo">210</span><a name="line.210"></a>
-<span class="sourceLineNo">211</span>    /**<a name="line.211"></a>
-<span class="sourceLineNo">212</span>     * Choose partitions between row ranges to hash to a single output file<a name="line.212"></a>
-<span class="sourceLineNo">213</span>     * Selects region boundaries that fall within the scan range, and groups them<a name="line.213"></a>
-<span class="sourceLineNo">214</span>     * into the desired number of partitions.<a name="line.214"></a>
-<span class="sourceLineNo">215</span>     */<a name="line.215"></a>
-<span class="sourceLineNo">216</span>    void selectPartitions(Pair&lt;byte[][], byte[][]&gt; regionStartEndKeys) {<a name="line.216"></a>
-<span class="sourceLineNo">217</span>      List&lt;byte[]&gt; startKeys = new ArrayList&lt;&gt;();<a name="line.217"></a>
-<span class="sourceLineNo">218</span>      for (int i = 0; i &lt; regionStartEndKeys.getFirst().length; i++) {<a name="line.218"></a>
-<span class="sourceLineNo">219</span>        byte[] regionStartKey = regionStartEndKeys.getFirst()[i];<a name="line.219"></a>
-<span class="sourceLineNo">220</span>        byte[] regionEndKey = regionStartEndKeys.getSecond()[i];<a name="line.220"></a>
-<span class="sourceLineNo">221</span><a name="line.221"></a>
-<span class="sourceLineNo">222</span>        // if scan begins after this region, or starts before this region, then drop this region<a name="line.222"></a>
-<span class="sourceLineNo">223</span>        // in other words:<a name="line.223"></a>
-<span class="sourceLineNo">224</span>        //   IF (scan begins before the end of this region<a name="line.224"></a>
-<span class="sourceLineNo">225</span>        //      AND scan ends before the start of this region)<a name="line.225"></a>
-<span class="sourceLineNo">226</span>        //   THEN include this region<a name="line.226"></a>
-<span class="sourceLineNo">227</span>        if ((isTableStartRow(startRow) || isTableEndRow(regionEndKey)<a name="line.227"></a>
-<span class="sourceLineNo">228</span>            || Bytes.compareTo(startRow, regionEndKey) &lt; 0)<a name="line.228"></a>
-<span class="sourceLineNo">229</span>          &amp;&amp; (isTableEndRow(stopRow) || isTableStartRow(regionStartKey)<a name="line.229"></a>
-<span class="sourceLineNo">230</span>            || Bytes.compareTo(stopRow, regionStartKey) &gt; 0)) {<a name="line.230"></a>
-<span class="sourceLineNo">231</span>          startKeys.add(regionStartKey);<a name="line.231"></a>
-<span class="sourceLineNo">232</span>        }<a name="line.232"></a>
-<span class="sourceLineNo">233</span>      }<a name="line.233"></a>
-<span class="sourceLineNo">234</span><a name="line.234"></a>
-<span class="sourceLineNo">235</span>      int numRegions = startKeys.size();<a name="line.235"></a>
-<span class="sourceLineNo">236</span>      if (numHashFiles == 0) {<a name="line.236"></a>
-<span class="sourceLineNo">237</span>        numHashFiles = numRegions / 100;<a name="line.237"></a>
-<span class="sourceLineNo">238</span>      }<a name="line.238"></a>
-<span class="sourceLineNo">239</span>      if (numHashFiles == 0) {<a name="line.239"></a>
-<span class="sourceLineNo">240</span>        numHashFiles = 1;<a name="line.240"></a>
-<span class="sourceLineNo">241</span>      }<a name="line.241"></a>
-<span class="sourceLineNo">242</span>      if (numHashFiles &gt; numRegions) {<a name="line.242"></a>
-<span class="sourceLineNo">243</span>        // can't partition within regions<a name="line.243"></a>
-<span class="sourceLineNo">244</span>        numHashFiles = numRegions;<a name="line.244"></a>
-<span class="sourceLineNo">245</span>      }<a name="line.245"></a>
-<span class="sourceLineNo">246</span><a name="line.246"></a>
-<span class="sourceLineNo">247</span>      // choose a subset of start keys to group regions into ranges<a name="line.247"></a>
-<span class="sourceLineNo">248</span>      partitions = new ArrayList&lt;&gt;(numHashFiles - 1);<a name="line.248"></a>
-<span class="sourceLineNo">249</span>      // skip the first start key as it is not a partition between ranges.<a name="line.249"></a>
-<span class="sourceLineNo">250</span>      for (long i = 1; i &lt; numHashFiles; i++) {<a name="line.250"></a>
-<span class="sourceLineNo">251</span>        int splitIndex = (int) (numRegions * i / numHashFiles);<a name="line.251"></a>
-<span class="sourceLineNo">252</span>        partitions.add(new ImmutableBytesWritable(startKeys.get(splitIndex)));<a name="line.252"></a>
-<span class="sourceLineNo">253</span>      }<a name="line.253"></a>
-<span class="sourceLineNo">254</span>    }<a name="line.254"></a>
-<span class="sourceLineNo">255</span><a name="line.255"></a>
-<span class="sourceLineNo">256</span>    void writePartitionFile(Configuration conf, Path path) throws IOException {<a name="line.256"></a>
-<span class="sourceLineNo">257</span>      FileSystem fs = path.getFileSystem(conf);<a name="line.257"></a>
-<span class="sourceLineNo">258</span>      @SuppressWarnings("deprecation")<a name="line.258"></a>
-<span class="sourceLineNo">259</span>      SequenceFile.Writer writer = SequenceFile.createWriter(<a name="line.259"></a>
-<span class="sourceLineNo">260</span>        fs, conf, path, ImmutableBytesWritable.class, NullWritable.class);<a name="line.260"></a>
-<span class="sourceLineNo">261</span><a name="line.261"></a>
-<span class="sourceLineNo">262</span>      for (int i = 0; i &lt; partitions.size(); i++) {<a name="line.262"></a>
-<span class="sourceLineNo">263</span>        writer.append(partitions.get(i), NullWritable.get());<a name="line.263"></a>
-<span class="sourceLineNo">264</span>      }<a name="line.264"></a>
-<span class="sourceLineNo">265</span>      writer.close();<a name="line.265"></a>
-<span class="sourceLineNo">266</span>    }<a name="line.266"></a>
-<span class="sourceLineNo">267</span><a name="line.267"></a>
-<span class="sourceLineNo">268</span>    private void readPartitionFile(FileSystem fs, Configuration conf, Path path)<a name="line.268"></a>
-<span class="sourceLineNo">269</span>         throws IOException {<a name="line.269"></a>
-<span class="sourceLineNo">270</span>      @SuppressWarnings("deprecation")<a name="line.270"></a>
-<span class="sourceLineNo">271</span>      SequenceFile.Reader reader = new SequenceFile.Reader(fs, path, conf);<a name="line.271"></a>
-<span class="sourceLineNo">272</span>      ImmutableBytesWritable key = new ImmutableBytesWritable();<a name="line.272"></a>
-<span class="sourceLineNo">273</span>      partitions = new ArrayList&lt;&gt;();<a name="line.273"></a>
-<span class="sourceLineNo">274</span>      while (reader.next(key)) {<a name="line.274"></a>
-<span class="sourceLineNo">275</span>        partitions.add(new ImmutableBytesWritable(key.copyBytes()));<a name="line.275"></a>
-<span class="sourceLineNo">276</span>      }<a name="line.276"></a>
-<span class="sourceLineNo">277</span>      reader.close();<a name="line.277"></a>
-<span class="sourceLineNo">278</span><a name="line.278"></a>
-<span class="sourceLineNo">279</span>      if (!Ordering.natural().isOrdered(partitions)) {<a name="line.279"></a>
-<span class="sourceLineNo">280</span>        throw new IOException("Partitions are not ordered!");<a name="line.280"></a>
-<span class="sourceLineNo">281</span>      }<a name="line.281"></a>
-<span class="sourceLineNo">282</span>    }<a name="line.282"></a>
-<span class="sourceLineNo">283</span><a name="line.283"></a>
-<span class="sourceLineNo">284</span>    @Override<a name="line.284"></a>
-<span class="sourceLineNo">285</span>    public String toString() {<a name="line.285"></a>
-<span class="sourceLineNo">286</span>      StringBuilder sb = new StringBuilder();<a name="line.286"></a>
-<span class="sourceLineNo">287</span>      sb.append("tableName=").append(tableName);<a name="line.287"></a>
-<span class="sourceLineNo">288</span>      if (families != null) {<a name="line.288"></a>
-<span class="sourceLineNo">289</span>        sb.append(", families=").append(families);<a name="line.289"></a>
-<span class="sourceLineNo">290</span>      }<a name="line.290"></a>
-<span class="sourceLineNo">291</span>      sb.append(", batchSize=").append(batchSize);<a name="line.291"></a>
-<span class="sourceLineNo">292</span>      sb.append(", numHashFiles=").append(numHashFiles);<a name="line.292"></a>
-<span class="sourceLineNo">293</span>      if (!isTableStartRow(startRow)) {<a name="line.293"></a>
-<span class="sourceLineNo">294</span>        sb.append(", startRowHex=").append(Bytes.toHex(startRow));<a name="line.294"></a>
-<span class="sourceLineNo">295</span>      }<a name="line.295"></a>
-<span class="sourceLineNo">296</span>      if (!isTableEndRow(stopRow)) {<a name="line.296"></a>
-<span class="sourceLineNo">297</span>        sb.append(", stopRowHex=").append(Bytes.toHex(stopRow));<a name="line.297"></a>
-<span class="sourceLineNo">298</span>      }<a name="line.298"></a>
-<span class="sourceLineNo">299</span>      if (scanBatch &gt;= 0) {<a name="line.299"></a>
-<span class="sourceLineNo">300</span>        sb.append(", scanBatch=").append(scanBatch);<a name="line.300"></a>
-<span class="sourceLineNo">301</span>      }<a name="line.301"></a>
-<span class="sourceLineNo">302</span>      if (versions &gt;= 0) {<a name="line.302"></a>
-<span class="sourceLineNo">303</span>        sb.append(", versions=").append(versions);<a name="line.303"></a>
-<span class="sourceLineNo">304</span>      }<a name="line.304"></a>
-<span class="sourceLineNo">305</span>      if (startTime != 0) {<a name="line.305"></a>
-<span class="sourceLineNo">306</span>        sb.append("startTime=").append(startTime);<a name="line.306"></a>
-<span class="sourceLineNo">307</span>      }<a name="line.307"></a>
-<span class="sourceLineNo">308</span>      if (endTime != 0) {<a name="line.308"></a>
-<span class="sourceLineNo">309</span>        sb.append("endTime=").append(endTime);<a name="line.309"></a>
-<span class="sourceLineNo">310</span>      }<a name="line.310"></a>
-<span class="sourceLineNo">311</span>      return sb.toString();<a name="line.311"></a>
-<span class="sourceLineNo">312</span>    }<a name="line.312"></a>
-<span class="sourceLineNo">313</span><a name="line.313"></a>
-<span class="sourceLineNo">314</span>    static String getDataFileName(int hashFileIndex) {<a name="line.314"></a>
-<span class="sourceLineNo">315</span>      return String.format(HashTable.OUTPUT_DATA_FILE_PREFIX + "%05d", hashFileIndex);<a name="line.315"></a>
-<span class="sourceLineNo">316</span>    }<a name="line.316"></a>
-<span class="sourceLineNo">317</span><a name="line.317"></a>
-<span class="sourceLineNo">318</span>    /**<a name="line.318"></a>
-<span class="sourceLineNo">319</span>     * Open a TableHash.Reader starting at the first hash at or after the given key.<a name="line.319"></a>
-<span class="sourceLineNo">320</span>     * @throws IOException<a name="line.320"></a>
-<span class="sourceLineNo">321</span>     */<a name="line.321"></a>
-<span class="sourceLineNo">322</span>    public Reader newReader(Configuration conf, ImmutableBytesWritable startKey)<a name="line.322"></a>
-<span class="sourceLineNo">323</span>        throws IOException {<a name="line.323"></a>
-<span class="sourceLineNo">324</span>      return new Reader(conf, startKey);<a name="line.324"></a>
-<span class="sourceLineNo">325</span>    }<a name="line.325"></a>
-<span class="sourceLineNo">326</span><a name="line.326"></a>
-<span class="sourceLineNo">327</span>    public class Reader implements java.io.Closeable {<a name="line.327"></a>
-<span class="sourceLineNo">328</span>      private final Configuration conf;<a name="line.328"></a>
-<span class="sourceLineNo">329</span><a name="line.329"></a>
-<span class="sourceLineNo">330</span>      private int hashFileIndex;<a name="line.330"></a>
-<span class="sourceLineNo">331</span>      private MapFile.Reader mapFileReader;<a name="line.331"></a>
-<span class="sourceLineNo">332</span><a name="line.332"></a>
-<span class="sourceLineNo">333</span>      private boolean cachedNext;<a name="line.333"></a>
-<span class="sourceLineNo">334</span>      private ImmutableBytesWritable key;<a name="line.334"></a>
-<span class="sourceLineNo">335</span>      private ImmutableBytesWritable hash;<a name="line.335"></a>
-<span class="sourceLineNo">336</span><a name="line.336"></a>
-<span class="sourceLineNo">337</span>      Reader(Configuration conf, ImmutableBytesWritable startKey) throws IOException {<a name="line.337"></a>
-<span class="sourceLineNo">338</span>        this.conf = conf;<a name="line.338"></a>
-<span class="sourceLineNo">339</span>        int partitionIndex = Collections.binarySearch(partitions, startKey);<a name="line.339"></a>
-<span class="sourceLineNo">340</span>        if (partitionIndex &gt;= 0) {<a name="line.340"></a>
-<span class="sourceLineNo">341</span>          // if the key is equal to a partition, then go the file after that partition<a name="line.341"></a>
-<span class="sourceLineNo">342</span>          hashFileIndex = partitionIndex+1;<a name="line.342"></a>
-<span class="sourceLineNo">343</span>        } else {<a name="line.343"></a>
-<span class="sourceLineNo">344</span>          // if the key is between partitions, then go to the file between those partitions<a name="line.344"></a>
-<span class="sourceLineNo">345</span>          hashFileIndex = -1-partitionIndex;<a name="line.345"></a>
-<span class="sourceLineNo">346</span>        }<a name="line.346"></a>
-<span class="sourceLineNo">347</span>        openHashFile();<a name="line.347"></a>
-<span class="sourceLineNo">348</span><a name="line.348"></a>
-<span class="sourceLineNo">349</span>        // MapFile's don't make it easy to seek() so that the subsequent next() returns<a name="line.349"></a>
-<span class="sourceLineNo">350</span>        // the desired key/value pair.  So we cache it for the first call of next().<a name="line.350"></a>
-<span class="sourceLineNo">351</span>        hash = new ImmutableBytesWritable();<a name="line.351"></a>
-<span class="sourceLineNo">352</span>        key = (ImmutableBytesWritable) mapFileReader.getClosest(startKey, hash);<a name="line.352"></a>
-<span class="sourceLineNo">353</span>        if (key == null) {<a name="line.353"></a>
-<span class="sourceLineNo">354</span>          cachedNext = false;<a name="line.354"></a>
-<span class="sourceLineNo">355</span>          hash = null;<a name="line.355"></a>
-<span class="sourceLineNo">356</span>        } else {<a name="line.356"></a>
-<span class="sourceLineNo">357</span>          cachedNext = true;<a name="line.357"></a>
-<span class="sourceLineNo">358</span>        }<a name="line.358"></a>
-<span class="sourceLineNo">359</span>      }<a name="line.359"></a>
-<span class="sourceLineNo">360</span><a name="line.360"></a>
-<span class="sourceLineNo">361</span>      /**<a name="line.361"></a>
-<span class="sourceLineNo">362</span>       * Read the next key/hash pair.<a name="line.362"></a>
-<span class="sourceLineNo">363</span>       * Returns true if such a pair exists and false when at the end of the data.<a name="line.363"></a>
-<span class="sourceLineNo">364</span>       */<a name="line.364"></a>
-<span class="sourceLineNo">365</span>      public boolean next() throws IOException {<a name="line.365"></a>
-<span class="sourceLineNo">366</span>        if (cachedNext) {<a name="line.366"></a>
-<span class="sourceLineNo">367</span>          cachedNext = false;<a name="line.367"></a>
-<span class="sourceLineNo">368</span>          return true;<a name="line.368"></a>
-<span class="sourceLineNo">369</span>        }<a name="line.369"></a>
-<span class="sourceLineNo">370</span>        key = new ImmutableBytesWritable();<a name="line.370"></a>
-<span class="sourceLineNo">371</span>        hash = new ImmutableBytesWritable();<a name="line.371"></a>
-<span class="sourceLineNo">372</span>        while (true) {<a name="line.372"></a>
-<span class="sourceLineNo">373</span>          boolean hasNext = mapFileReader.next(key, hash);<a name="line.373"></a>
-<span class="sourceLineNo">374</span>          if (hasNext) {<a name="line.374"></a>
-<span class="sourceLineNo">375</span>            return true;<a name="line.375"></a>
-<span class="sourceLineNo">376</span>          }<a name="line.376"></a>
-<span class="sourceLineNo">377</span>          hashFileIndex++;<a name="line.377"></a>
-<span class="sourceLineNo">378</span>          if (hashFileIndex &lt; TableHash.this.numHashFiles) {<a name="line.378"></a>
-<span class="sourceLineNo">379</span>            mapFileReader.close();<a name="line.379"></a>
-<span class="sourceLineNo">380</span>            openHashFile();<a name="line.380"></a>
-<span class="sourceLineNo">381</span>          } else {<a name="line.381"></a>
-<span class="sourceLineNo">382</span>            key = null;<a name="line.382"></a>
-<span class="sourceLineNo">383</span>            hash = null;<a name="line.383"></a>
-<span class="sourceLineNo">384</span>            return false;<a name="line.384"></a>
-<span class="sourceLineNo">385</span>          }<a name="line.385"></a>
-<span class="sourceLineNo">386</span>        }<a name="line.386"></a>
-<span class="sourceLineNo">387</span>      }<a name="line.387"></a>
-<span class="sourceLineNo">388</span><a name="line.388"></a>
-<span class="sourceLineNo">389</span>      /**<a name="line.389"></a>
-<span class="sourceLineNo">390</span>       * Get the current key<a name="line.390"></a>
-<span class="sourceLineNo">391</span>       * @return the current key or null if there is no current key<a name="line.391"></a>
-<span class="sourceLineNo">392</span>       */<a name="line.392"></a>
-<span class="sourceLineNo">393</span>      public ImmutableBytesWritable getCurrentKey() {<a name="line.393"></a>
-<span class="sourceLineNo">394</span>        return key;<a name="line.394"></a>
-<span class="sourceLineNo">395</span>      }<a name="line.395"></a>
-<span class="sourceLineNo">396</span><a name="line.396"></a>
-<span class="sourceLineNo">397</span>      /**<a name="line.397"></a>
-<span class="sourceLineNo">398</span>       * Get the current hash<a name="line.398"></a>
-<span class="sourceLineNo">399</span>       * @return the current hash or null if there is no current hash<a name="line.399"></a>
-<span class="sourceLineNo">400</span>       */<a name="line.400"></a>
-<span class="sourceLineNo">401</span>      public ImmutableBytesWritable getCurrentHash() {<a name="line.401"></a>
-<span class="sourceLineNo">402</span>        return hash;<a name="line.402"></a>
-<span class="sourceLineNo">403</span>      }<a name="line.403"></a>
-<span class="sourceLineNo">404</span><a name="line.404"></a>
-<span class="sourceLineNo">405</span>      private void openHashFile() throws IOException {<a name="line.405"></a>
-<span class="sourceLineNo">406</span>        if (mapFileReader != null) {<a name="line.406"></a>
-<span class="sourceLineNo">407</span>          mapFileReader.close();<a name="line.407"></a>
-<span class="sourceLineNo">408</span>        }<a name="line.408"></a>
-<span class="sourceLineNo">409</span>        Path dataDir = new Path(TableHash.this.hashDir, HASH_DATA_DIR);<a name="line.409"></a>
-<span class="sourceLineNo">410</span>        Path dataFile = new Path(dataDir, getDataFileName(hashFileIndex));<a name="line.410"></a>
-<span class="sourceLineNo">411</span>        mapFileReader = new MapFile.Reader(dataFile, conf);<a name="line.411"></a>
-<span class="sourceLineNo">412</span>      }<a name="line.412"></a>
-<span class="sourceLineNo">413</span><a name="line.413"></a>
-<span class="sourceLineNo">414</span>      @Override<a name="line.414"></a>
-<span class="sourceLineNo">415</span>      public void close() throws IOException {<a name="line.415"></a>
-<span class="sourceLineNo">416</span>        mapFileReader.close();<a name="line.416"></a>
-<span class="sourceLineNo">417</span>      }<a name="line.417"></a>
-<span class="sourceLineNo">418</span>    }<a name="line.418"></a>
-<span class="sourceLineNo">419</span>  }<a name="line.419"></a>
-<span class="sourceLineNo">420</span><a name="line.420"></a>
-<span class="sourceLineNo">421</span>  static boolean isTableStartRow(byte[] row) {<a name="line.421"></a>
-<span class="sourceLineNo">422</span>    return Bytes.equals(HConstants.EMPTY_START_ROW, row);<a name="line.422"></a>
-<span class="sourceLineNo">423</span>  }<a name="line.423"></a>
-<span class="sourceLineNo">424</span><a name="line.424"></a>
-<span class="sourceLineNo">425</span>  static boolean isTableEndRow(byte[] row) {<a name="line.425"></a>
-<span class="sourceLineNo">426</span>    return Bytes.equals(HConstants.EMPTY_END_ROW, row);<a name="line.426"></a>
-<span class="sourceLineNo">427</span>  }<a name="line.427"></a>
-<span class="sourceLineNo">428</span><a name="line.428"></a>
-<span class="sourceLineNo">429</span>  public Job createSubmittableJob(String[] args) throws IOException {<a name="line.429"></a>
-<span class="sourceLineNo">430</span>    Path partitionsPath = new Path(destPath, PARTITIONS_FILE_NAME);<a name="line.430"></a>
-<span class="sourceLineNo">431</span>    generatePartitions(partitionsPath);<a name="line.431"></a>
-<span class="sourceLineNo">432</span><a name="line.432"></a>
-<span class="sourceLineNo">433</span>    Job job = Job.getInstance(getConf(),<a name="line.433"></a>
-<span class="sourceLineNo">434</span>          getConf().get("mapreduce.job.name", "hashTable_" + tableHash.tableName));<a name="line.434"></a>
-<span class="sourceLineNo">435</span>    Configuration jobConf = job.getConfiguration();<a name="line.435"></a>
-<span class="sourceLineNo">436</span>    jobConf.setLong(HASH_BATCH_SIZE_CONF_KEY, tableHash.batchSize);<a name="line.436"></a>
-<span class="sourceLineNo">437</span>    job.setJarByClass(HashTable.class);<a name="line.437"></a>
-<span class="sourceLineNo">438</span><a name="line.438"></a>
-<span class="sourceLineNo">439</span>    TableMapReduceUtil.initTableMapperJob(tableHash.tableName, tableHash.initScan(),<a name="line.439"></a>
-<span class="sourceLineNo">440</span>        HashMapper.class, ImmutableBytesWritable.class, ImmutableBytesWritable.class, job);<a name="line.440"></a>
+<span class="sourceLineNo">102</span>    List&lt;ImmutableBytesWritable&gt; partitions;<a name="line.102"></a>
+<span class="sourceLineNo">103</span><a name="line.103"></a>
+<span class="sourceLineNo">104</span>    public static TableHash read(Configuration conf, Path hashDir) throws IOException {<a name="line.104"></a>
+<span class="sourceLineNo">105</span>      TableHash tableHash = new TableHash();<a name="line.105"></a>
+<span class="sourceLineNo">106</span>      FileSystem fs = hashDir.getFileSystem(conf);<a name="line.106"></a>
+<span class="sourceLineNo">107</span>      tableHash.hashDir = hashDir;<a name="line.107"></a>
+<span class="sourceLineNo">108</span>      tableHash.readPropertiesFile(fs, new Path(hashDir, MANIFEST_FILE_NAME));<a name="line.108"></a>
+<span class="sourceLineNo">109</span>      tableHash.readPartitionFile(fs, conf, new Path(hashDir, PARTITIONS_FILE_NAME));<a name="line.109"></a>
+<span class="sourceLineNo">110</span>      return tableHash;<a name="line.110"></a>
+<span class="sourceLineNo">111</span>    }<a name="line.111"></a>
+<span class="sourceLineNo">112</span><a name="line.112"></a>
+<span class="sourceLineNo">113</span>    void writePropertiesFile(FileSystem fs, Path path) throws IOException {<a name="line.113"></a>
+<span class="sourceLineNo">114</span>      Properties p = new Properties();<a name="line.114"></a>
+<span class="sourceLineNo">115</span>      p.setProperty("table", tableName);<a name="line.115"></a>
+<span class="sourceLineNo">116</span>      if (families != null) {<a name="line.116"></a>
+<span class="sourceLineNo">117</span>        p.setProperty("columnFamilies", families);<a name="line.117"></a>
+<span class="sourceLineNo">118</span>      }<a name="line.118"></a>
+<span class="sourceLineNo">119</span>      p.setProperty("targetBatchSize", Long.toString(batchSize));<a name="line.119"></a>
+<span class="sourceLineNo">120</span>      p.setProperty("numHashFiles", Integer.toString(numHashFiles));<a name="line.120"></a>
+<span class="sourceLineNo">121</span>      if (!isTableStartRow(startRow)) {<a name="line.121"></a>
+<span class="sourceLineNo">122</span>        p.setProperty("startRowHex", Bytes.toHex(startRow));<a name="line.122"></a>
+<span class="sourceLineNo">123</span>      }<a name="line.123"></a>
+<span class="sourceLineNo">124</span>      if (!isTableEndRow(stopRow)) {<a name="line.124"></a>
+<span class="sourceLineNo">125</span>        p.setProperty("stopRowHex", Bytes.toHex(stopRow));<a name="line.125"></a>
+<span class="sourceLineNo">126</span>      }<a name="line.126"></a>
+<span class="sourceLineNo">127</span>      if (scanBatch &gt; 0) {<a name="line.127"></a>
+<span class="sourceLineNo">128</span>        p.setProperty("scanBatch", Integer.toString(scanBatch));<a name="line.128"></a>
+<span class="sourceLineNo">129</span>      }<a name="line.129"></a>
+<span class="sourceLineNo">130</span>      if (versions &gt;= 0) {<a name="line.130"></a>
+<span class="sourceLineNo">131</span>        p.setProperty("versions", Integer.toString(versions));<a name="line.131"></a>
+<span class="sourceLineNo">132</span>      }<a name="line.132"></a>
+<span class="sourceLineNo">133</span>      if (startTime != 0) {<a name="line.133"></a>
+<span class="sourceLineNo">134</span>        p.setProperty("startTimestamp", Long.toString(startTime));<a name="line.134"></a>
+<span class="sourceLineNo">135</span>      }<a name="line.135"></a>
+<span class="sourceLineNo">136</span>      if (endTime != 0) {<a name="line.136"></a>
+<span class="sourceLineNo">137</span>        p.setProperty("endTimestamp", Long.toString(endTime));<a name="line.137"></a>
+<span class="sourceLineNo">138</span>      }<a name="line.138"></a>
+<span class="sourceLineNo">139</span><a name="line.139"></a>
+<span class="sourceLineNo">140</span>      try (OutputStreamWriter osw = new OutputStreamWriter(fs.create(path), Charsets.UTF_8)) {<a name="line.140"></a>
+<span class="sourceLineNo">141</span>        p.store(osw, null);<a name="line.141"></a>
+<span class="sourceLineNo">142</span>      }<a name="line.142"></a>
+<span class="sourceLineNo">143</span>    }<a name="line.143"></a>
+<span class="sourceLineNo">144</span><a name="line.144"></a>
+<span class="sourceLineNo">145</span>    void readPropertiesFile(FileSystem fs, Path path) throws IOException {<a name="line.145"></a>
+<span class="sourceLineNo">146</span>      Properties p = new Properties();<a name="line.146"></a>
+<span class="sourceLineNo">147</span>      try (FSDataInputStream in = fs.open(path)) {<a name="line.147"></a>
+<span class="sourceLineNo">148</span>        try (InputStreamReader isr = new InputStreamReader(in, Charsets.UTF_8)) {<a name="line.148"></a>
+<span class="sourceLineNo">149</span>          p.load(isr);<a name="line.149"></a>
+<span class="sourceLineNo">150</span>        }<a name="line.150"></a>
+<span class="sourceLineNo">151</span>      }<a name="line.151"></a>
+<span class="sourceLineNo">152</span>      tableName = p.getProperty("table");<a name="line.152"></a>
+<span class="sourceLineNo">153</span>      families = p.getProperty("columnFamilies");<a name="line.153"></a>
+<span class="sourceLineNo">154</span>      batchSize = Long.parseLong(p.getProperty("targetBatchSize"));<a name="line.154"></a>
+<span class="sourceLineNo">155</span>      numHashFiles = Integer.parseInt(p.getProperty("numHashFiles"));<a name="line.155"></a>
+<span class="sourceLineNo">156</span><a name="line.156"></a>
+<span class="sourceLineNo">157</span>      String startRowHex = p.getProperty("startRowHex");<a name="line.157"></a>
+<span class="sourceLineNo">158</span>      if (startRowHex != null) {<a name="line.158"></a>
+<span class="sourceLineNo">159</span>        startRow = Bytes.fromHex(startRowHex);<a name="line.159"></a>
+<span class="sourceLineNo">160</span>      }<a name="line.160"></a>
+<span class="sourceLineNo">161</span>      String stopRowHex = p.getProperty("stopRowHex");<a name="line.161"></a>
+<span class="sourceLineNo">162</span>      if (stopRowHex != null) {<a name="line.162"></a>
+<span class="sourceLineNo">163</span>        stopRow = Bytes.fromHex(stopRowHex);<a name="line.163"></a>
+<span class="sourceLineNo">164</span>      }<a name="line.164"></a>
+<span class="sourceLineNo">165</span><a name="line.165"></a>
+<span class="sourceLineNo">166</span>      String scanBatchString = p.getProperty("scanBatch");<a name="line.166"></a>
+<span class="sourceLineNo">167</span>      if (scanBatchString != null) {<a name="line.167"></a>
+<span class="sourceLineNo">168</span>        scanBatch = Integer.parseInt(scanBatchString);<a name="line.168"></a>
+<span class="sourceLineNo">169</span>      }<a name="line.169"></a>
+<span class="sourceLineNo">170</span><a name="line.170"></a>
+<span class="sourceLineNo">171</span>      String versionString = p.getProperty("versions");<a name="line.171"></a>
+<span class="sourceLineNo">172</span>      if (versionString != null) {<a name="line.172"></a>
+<span class="sourceLineNo">173</span>        versions = Integer.parseInt(versionString);<a name="line.173"></a>
+<span class="sourceLineNo">174</span>      }<a name="line.174"></a>
+<span class="sourceLineNo">175</span><a name="line.175"></a>
+<span class="sourceLineNo">176</span>      String startTimeString = p.getProperty("startTimestamp");<a name="line.176"></a>
+<span class="sourceLineNo">177</span>      if (startTimeString != null) {<a name="line.177"></a>
+<span class="sourceLineNo">178</span>        startTime = Long.parseLong(startTimeString);<a name="line.178"></a>
+<span class="sourceLineNo">179</span>      }<a name="line.179"></a>
+<span class="sourceLineNo">180</span><a name="line.180"></a>
+<span class="sourceLineNo">181</span>      String endTimeString = p.getProperty("endTimestamp");<a name="line.181"></a>
+<span class="sourceLineNo">182</span>      if (endTimeString != null) {<a name="line.182"></a>
+<span class="sourceLineNo">183</span>        endTime = Long.parseLong(endTimeString);<a name="line.183"></a>
+<span class="sourceLineNo">184</span>      }<a name="line.184"></a>
+<span class="sourceLineNo">185</span>    }<a name="line.185"></a>
+<span class="sourceLineNo">186</span><a name="line.186"></a>
+<span class="sourceLineNo">187</span>    Scan initScan() throws IOException {<a name="line.187"></a>
+<span class="sourceLineNo">188</span>      Scan scan = new Scan();<a name="line.188"></a>
+<span class="sourceLineNo">189</span>      scan.setCacheBlocks(false);<a name="line.189"></a>
+<span class="sourceLineNo">190</span>      if (startTime != 0 || endTime != 0) {<a name="line.190"></a>
+<span class="sourceLineNo">191</span>        scan.setTimeRange(startTime, endTime == 0 ? HConstants.LATEST_TIMESTAMP : endTime);<a name="line.191"></a>
+<span class="sourceLineNo">192</span>      }<a name="line.192"></a>
+<span class="sourceLineNo">193</span>      if (scanBatch &gt; 0) {<a name="line.193"></a>
+<span class="sourceLineNo">194</span>        scan.setBatch(scanBatch);<a name="line.194"></a>
+<span class="sourceLineNo">195</span>      }<a name="line.195"></a>
+<span class="sourceLineNo">196</span>      if (versions &gt;= 0) {<a name="line.196"></a>
+<span class="sourceLineNo">197</span>        scan.readVersions(versions);<a name="line.197"></a>
+<span class="sourceLineNo">198</span>      }<a name="line.198"></a>
+<span class="sourceLineNo">199</span>      if (!isTableStartRow(startRow)) {<a name="line.199"></a>
+<span class="sourceLineNo">200</span>        scan.withStartRow(startRow);<a name="line.200"></a>
+<span class="sourceLineNo">201</span>      }<a name="line.201"></a>
+<span class="sourceLineNo">202</span>      if (!isTableEndRow(stopRow)) {<a name="line.202"></a>
+<span class="sourceLineNo">203</span>        scan.withStopRow(stopRow);<a name="line.203"></a>
+<span class="sourceLineNo">204</span>      }<a name="line.204"></a>
+<span class="sourceLineNo">205</span>      if(families != null) {<a name="line.205"></a>
+<span class="sourceLineNo">206</span>        for(String fam : families.split(",")) {<a name="line.206"></a>
+<span class="sourceLineNo">207</span>          scan.addFamily(Bytes.toBytes(fam));<a name="line.207"></a>
+<span class="sourceLineNo">208</span>        }<a name="line.208"></a>
+<span class="sourceLineNo">209</span>      }<a name="line.209"></a>
+<span class="sourceLineNo">210</span>      return scan;<a name="line.210"></a>
+<span class="sourceLineNo">211</span>    }<a name="line.211"></a>
+<span class="sourceLineNo">212</span><a name="line.212"></a>
+<span class="sourceLineNo">213</span>    /**<a name="line.213"></a>
+<span class="sourceLineNo">214</span>     * Choose partitions between row ranges to hash to a single output file<a name="line.214"></a>
+<span class="sourceLineNo">215</span>     * Selects region boundaries that fall within the scan range, and groups them<a name="line.215"></a>
+<span class="sourceLineNo">216</span>     * into the desired number of partitions.<a name="line.216"></a>
+<span class="sourceLineNo">217</span>     */<a name="line.217"></a>
+<span class="sourceLineNo">218</span>    void selectPartitions(Pair&lt;byte[][], byte[][]&gt; regionStartEndKeys) {<a name="line.218"></a>
+<span class="sourceLineNo">219</span>      List&lt;byte[]&gt; startKeys = new ArrayList&lt;&gt;();<a name="line.219"></a>
+<span class="sourceLineNo">220</span>      for (int i = 0; i &lt; regionStartEndKeys.getFirst().length; i++) {<a name="line.220"></a>
+<span class="sourceLineNo">221</span>        byte[] regionStartKey = regionStartEndKeys.getFirst()[i];<a name="line.221"></a>
+<span class="sourceLineNo">222</span>        byte[] regionEndKey = regionStartEndKeys.getSecond()[i];<a name="line.222"></a>
+<span class="sourceLineNo">223</span><a name="line.223"></a>
+<span class="sourceLineNo">224</span>        // if scan begins after this region, or starts before this region, then drop this region<a name="line.224"></a>
+<span class="sourceLineNo">225</span>        // in other words:<a name="line.225"></a>
+<span class="sourceLineNo">226</span>        //   IF (scan begins before the end of this region<a name="line.226"></a>
+<span class="sourceLineNo">227</span>        //      AND scan ends before the start of this region)<a name="line.227"></a>
+<span class="sourceLineNo">228</span>        //   THEN include this region<a name="line.228"></a>
+<span class="sourceLineNo">229</span>        if ((isTableStartRow(startRow) || isTableEndRow(regionEndKey)<a name="line.229"></a>
+<span class="sourceLineNo">230</span>            || Bytes.compareTo(startRow, regionEndKey) &lt; 0)<a name="line.230"></a>
+<span class="sourceLineNo">231</span>          &amp;&amp; (isTableEndRow(stopRow) || isTableStartRow(regionStartKey)<a name="line.231"></a>
+<span class="sourceLineNo">232</span>            || Bytes.compareTo(stopRow, regionStartKey) &gt; 0)) {<a name="line.232"></a>
+<span class="sourceLineNo">233</span>          startKeys.add(regionStartKey);<a name="line.233"></a>
+<span class="sourceLineNo">234</span>        }<a name="line.234"></a>
+<span class="sourceLineNo">235</span>      }<a name="line.235"></a>
+<span class="sourceLineNo">236</span><a name="line.236"></a>
+<span class="sourceLineNo">237</span>      int numRegions = startKeys.size();<a name="line.237"></a>
+<span class="sourceLineNo">238</span>      if (numHashFiles == 0) {<a name="line.238"></a>
+<span class="sourceLineNo">239</span>        numHashFiles = numRegions / 100;<a name="line.239"></a>
+<span class="sourceLineNo">240</span>      }<a name="line.240"></a>
+<span class="sourceLineNo">241</span>      if (numHashFiles == 0) {<a name="line.241"></a>
+<span class="sourceLineNo">242</span>        numHashFiles = 1;<a name="line.242"></a>
+<span class="sourceLineNo">243</span>      }<a name="line.243"></a>
+<span class="sourceLineNo">244</span>      if (numHashFiles &gt; numRegions) {<a name="line.244"></a>
+<span class="sourceLineNo">245</span>        // can't partition within regions<a name="line.245"></a>
+<span class="sourceLineNo">246</span>        numHashFiles = numRegions;<a name="line.246"></a>
+<span class="sourceLineNo">247</span>      }<a name="line.247"></a>
+<span class="sourceLineNo">248</span><a name="line.248"></a>
+<span class="sourceLineNo">249</span>      // choose a subset of start keys to group regions into ranges<a name="line.249"></a>
+<span class="sourceLineNo">250</span>      partitions = new ArrayList&lt;&gt;(numHashFiles - 1);<a name="line.250"></a>
+<span class="sourceLineNo">251</span>      // skip the first start key as it is not a partition between ranges.<a name="line.251"></a>
+<span class="sourceLineNo">252</span>      for (long i = 1; i &lt; numHashFiles; i++) {<a name="line.252"></a>
+<span class="sourceLineNo">253</span>        int splitIndex = (int) (numRegions * i / numHashFiles);<a name="line.253"></a>
+<span class="sourceLineNo">254</span>        partitions.add(new ImmutableBytesWritable(startKeys.get(splitIndex)));<a name="line.254"></a>
+<span class="sourceLineNo">255</span>      }<a name="line.255"></a>
+<span class="sourceLineNo">256</span>    }<a name="line.256"></a>
+<span class="sourceLineNo">257</span><a name="line.257"></a>
+<span class="sourceLineNo">258</span>    void writePartitionFile(Configuration conf, Path path) throws IOException {<a name="line.258"></a>
+<span class="sourceLineNo">259</span>      FileSystem fs = path.getFileSystem(conf);<a name="line.259"></a>
+<span class="sourceLineNo">260</span>      @SuppressWarnings("deprecation")<a name="line.260"></a>
+<span class="sourceLineNo">261</span>      SequenceFile.Writer writer = SequenceFile.createWriter(<a name="line.261"></a>
+<span class="sourceLineNo">262</span>        fs, conf, path, ImmutableBytesWritable.class, NullWritable.class);<a name="line.262"></a>
+<span class="sourceLineNo">263</span><a name="line.263"></a>
+<span class="sourceLineNo">264</span>      for (int i = 0; i &lt; partitions.size(); i++) {<a name="line.264"></a>
+<span class="sourceLineNo">265</span>        writer.append(partitions.get(i), NullWritable.get());<a name="line.265"></a>
+<span class="sourceLineNo">266</span>      }<a name="line.266"></a>
+<span class="sourceLineNo">267</span>      writer.close();<a name="line.267"></a>
+<span class="sourceLineNo">268</span>    }<a name="line.268"></a>
+<span class="sourceLineNo">269</span><a name="line.269"></a>
+<span class="sourceLineNo">270</span>    private void readPartitionFile(FileSystem fs, Configuration conf, Path path)<a name="line.270"></a>
+<span class="sourceLineNo">271</span>         throws IOException {<a name="line.271"></a>
+<span class="sourceLineNo">272</span>      @SuppressWarnings("deprecation")<a name="line.272"></a>
+<span class="sourceLineNo">273</span>      SequenceFile.Reader reader = new SequenceFile.Reader(fs, path, conf);<a name="line.273"></a>
+<span class="sourceLineNo">274</span>      ImmutableBytesWritable key = new ImmutableBytesWritable();<a name="line.274"></a>
+<span class="sourceLineNo">275</span>      partitions = new ArrayList&lt;&gt;();<a name="line.275"></a>
+<span class="sourceLineNo">276</span>      while (reader.next(key)) {<a name="line.276"></a>
+<span class="sourceLineNo">277</span>        partitions.add(new ImmutableBytesWritable(key.copyBytes()));<a name="line.277"></a>
+<span class="sourceLineNo">278</span>      }<a name="line.278"></a>
+<span class="sourceLineNo">279</span>      reader.close();<a name="line.279"></a>
+<span class="sourceLineNo">280</span><a name="line.280"></a>
+<span class="sourceLineNo">281</span>      if (!Ordering.natural().isOrdered(partitions)) {<a name="line.281"></a>
+<span class="sourceLineNo">282</span>        throw new IOException("Partitions are not ordered!");<a name="line.282"></a>
+<span class="sourceLineNo">283</span>      }<a name="line.283"></a>
+<span class="sourceLineNo">284</span>    }<a name="line.284"></a>
+<span class="sourceLineNo">285</span><a name="line.285"></a>
+<span class="sourceLineNo">286</span>    @Override<a name="line.286"></a>
+<span class="sourceLineNo">287</span>    public String toString() {<a name="line.287"></a>
+<span class="sourceLineNo">288</span>      StringBuilder sb = new StringBuilder();<a name="line.288"></a>
+<span class="sourceLineNo">289</span>      sb.append("tableName=").append(tableName);<a name="line.289"></a>
+<span class="sourceLineNo">290</span>      if (families != null) {<a name="line.290"></a>
+<span class="sourceLineNo">291</span>        sb.append(", families=").append(families);<a name="line.291"></a>
+<span class="sourceLineNo">292</span>      }<a name="line.292"></a>
+<span class="sourceLineNo">293</span>      sb.append(", batchSize=").append(batchSize);<a name="line.293"></a>
+<span class="sourceLineNo">294</span>      sb.append(", numHashFiles=").append(numHashFiles);<a name="line.294"></a>
+<span class="sourceLineNo">295</span>      if (!isTableStartRow(startRow)) {<a name="line.295"></a>
+<span class="sourceLineNo">296</span>        sb.append(", startRowHex=").append(Bytes.toHex(startRow));<a name="line.296"></a>
+<span class="sourceLineNo">297</span>      }<a name="line.297"></a>
+<span class="sourceLineNo">298</span>      if (!isTableEndRow(stopRow)) {<a name="line.298"></a>
+<span class="sourceLineNo">299</span>        sb.append(", stopRowHex=").append(Bytes.toHex(stopRow));<a name="line.299"></a>
+<span class="sourceLineNo">300</span>      }<a name="line.300"></a>
+<span class="sourceLineNo">301</span>      if (scanBatch &gt;= 0) {<a name="line.301"></a>
+<span class="sourceLineNo">302</span>        sb.append(", scanBatch=").append(scanBatch);<a name="line.302"></a>
+<span class="sourceLineNo">303</span>      }<a name="line.303"></a>
+<span class="sourceLineNo">304</span>      if (versions &gt;= 0) {<a name="line.304"></a>
+<span class="sourceLineNo">305</span>        sb.append(", versions=").append(versions);<a name="line.305"></a>
+<span class="sourceLineNo">306</span>      }<a name="line.306"></a>
+<span class="sourceLineNo">307</span>      if (startTime != 0) {<a name="line.307"></a>
+<span class="sourceLineNo">308</span>        sb.append("startTime=").append(startTime);<a name="line.308"></a>
+<span class="sourceLineNo">309</span>      }<a name="line.309"></a>
+<span class="sourceLineNo">310</span>      if (endTime != 0) {<a name="line.310"></a>
+<span class="sourceLineNo">311</span>        sb.append("endTime=").append(endTime);<a name="line.311"></a>
+<span class="sourceLineNo">312</span>      }<a name="line.312"></a>
+<span class="sourceLineNo">313</span>      return sb.toString();<a name="line.313"></a>
+<span class="sourceLineNo">314</span>    }<a name="line.314"></a>
+<span class="sourceLineNo">315</span><a name="line.315"></a>
+<span class="sourceLineNo">316</span>    static String getDataFileName(int hashFileIndex) {<a name="line.316"></a>
+<span class="sourceLineNo">317</span>      return String.format(HashTable.OUTPUT_DATA_FILE_PREFIX + "%05d", hashFileIndex);<a name="line.317"></a>
+<span class="sourceLineNo">318</span>    }<a name="line.318"></a>
+<span class="sourceLineNo">319</span><a name="line.319"></a>
+<span class="sourceLineNo">320</span>    /**<a name="line.320"></a>
+<span class="sourceLineNo">321</span>     * Open a TableHash.Reader starting at the first hash at or after the given key.<a name="line.321"></a>
+<span class="sourceLineNo">322</span>     * @throws IOException<a name="line.322"></a>
+<span class="sourceLineNo">323</span>     */<a name="line.323"></a>
+<span class="sourceLineNo">324</span>    public Reader newReader(Configuration conf, ImmutableBytesWritable startKey)<a name="line.324"></a>
+<span class="sourceLineNo">325</span>        throws IOException {<a name="line.325"></a>
+<span class="sourceLineNo">326</span>      return new Reader(conf, startKey);<a name="line.326"></a>
+<span class="sourceLineNo">327</span>    }<a name="line.327"></a>
+<span class="sourceLineNo">328</span><a name="line.328"></a>
+<span class="sourceLineNo">329</span>    public class Reader implements java.io.Closeable {<a name="line.329"></a>
+<span class="sourceLineNo">330</span>      private final Configuration conf;<a name="line.330"></a>
+<span class="sourceLineNo">331</span><a name="line.331"></a>
+<span class="sourceLineNo">332</span>      private int hashFileIndex;<a name="line.332"></a>
+<span class="sourceLineNo">333</span>      private MapFile.Reader mapFileReader;<a name="line.333"></a>
+<span class="sourceLineNo">334</span><a name="line.334"></a>
+<span class="sourceLineNo">335</span>      private boolean cachedNext;<a name="line.335"></a>
+<span class="sourceLineNo">336</span>      private ImmutableBytesWritable key;<a name="line.336"></a>
+<span class="sourceLineNo">337</span>      private ImmutableBytesWritable hash;<a name="line.337"></a>
+<span class="sourceLineNo">338</span><a name="line.338"></a>
+<span class="sourceLineNo">339</span>      Reader(Configuration conf, ImmutableBytesWritable startKey) throws IOException {<a name="line.339"></a>
+<span class="sourceLineNo">340</span>        this.conf = conf;<a name="line.340"></a>
+<span class="sourceLineNo">341</span>        int partitionIndex = Collections.binarySearch(partitions, startKey);<a name="line.341"></a>
+<span class="sourceLineNo">342</span>        if (partitionIndex &gt;= 0) {<a name="line.342"></a>
+<span class="sourceLineNo">343</span>          // if the key is equal to a partition, then go the file after that partition<a name="line.343"></a>
+<span class="sourceLineNo">344</span>          hashFileIndex = partitionIndex+1;<a name="line.344"></a>
+<span class="sourceLineNo">345</span>        } else {<a name="line.345"></a>
+<span class="sourceLineNo">346</span>          // if the key is between partitions, then go to the file between those partitions<a name="line.346"></a>
+<span class="sourceLineNo">347</span>          hashFileIndex = -1-partitionIndex;<a name="line.347"></a>
+<span class="sourceLineNo">348</span>        }<a name="line.348"></a>
+<span class="sourceLineNo">349</span>        openHashFile();<a name="line.349"></a>
+<span class="sourceLineNo">350</span><a name="line.350"></a>
+<span class="sourceLineNo">351</span>        // MapFile's don't make it easy to seek() so that the subsequent next() returns<a name="line.351"></a>
+<span class="sourceLineNo">352</span>        // the desired key/value pair.  So we cache it for the first call of next().<a name="line.352"></a>
+<span class="sourceLineNo">353</span>        hash = new ImmutableBytesWritable();<a name="line.353"></a>
+<span class="sourceLineNo">354</span>        key = (ImmutableBytesWritable) mapFileReader.getClosest(startKey, hash);<a name="line.354"></a>
+<span class="sourceLineNo">355</span>        if (key == null) {<a name="line.355"></a>
+<span class="sourceLineNo">356</span>          cachedNext = false;<a name="line.356"></a>
+<span class="sourceLineNo">357</span>          hash = null;<a name="line.357"></a>
+<span class="sourceLineNo">358</span>        } else {<a name="line.358"></a>
+<span class="sourceLineNo">359</span>          cachedNext = true;<a name="line.359"></a>
+<span class="sourceLineNo">360</span>        }<a name="line.360"></a>
+<span class="sourceLineNo">361</span>      }<a name="line.361"></a>
+<span class="sourceLineNo">362</span><a name="line.362"></a>
+<span class="sourceLineNo">363</span>      /**<a name="line.363"></a>
+<span class="sourceLineNo">364</span>       * Read the next key/hash pair.<a name="line.364"></a>
+<span class="sourceLineNo">365</span>       * Returns true if such a pair exists and false when at the end of the data.<a name="line.365"></a>
+<span class="sourceLineNo">366</span>       */<a name="line.366"></a>
+<span class="sourceLineNo">367</span>      public boolean next() throws IOException {<a name="line.367"></a>
+<span class="sourceLineNo">368</span>        if (cachedNext) {<a name="line.368"></a>
+<span class="sourceLineNo">369</span>          cachedNext = false;<a name="line.369"></a>
+<span class="sourceLineNo">370</span>          return true;<a name="line.370"></a>
+<span class="sourceLineNo">371</span>        }<a name="line.371"></a>
+<span class="sourceLineNo">372</span>        key = new ImmutableBytesWritable();<a name="line.372"></a>
+<span class="sourceLineNo">373</span>        hash = new ImmutableBytesWritable();<a name="line.373"></a>
+<span class="sourceLineNo">374</span>        while (true) {<a name="line.374"></a>
+<span class="sourceLineNo">375</span>          boolean hasNext = mapFileReader.next(key, hash);<a name="line.375"></a>
+<span class="sourceLineNo">376</span>          if (hasNext) {<a name="line.376"></a>
+<span class="sourceLineNo">377</span>            return true;<a name="line.377"></a>
+<span class="sourceLineNo">378</span>          }<a name="line.378"></a>
+<span class="sourceLineNo">379</span>          hashFileIndex++;<a name="line.379"></a>
+<span class="sourceLineNo">380</span>          if (hashFileIndex &lt; TableHash.this.numHashFiles) {<a name="line.380"></a>
+<span class="sourceLineNo">381</span>            mapFileReader.close();<a name="line.381"></a>
+<span class="sourceLineNo">382</span>            openHashFile();<a name="line.382"></a>
+<span class="sourceLineNo">383</span>          } else {<a name="line.383"></a>
+<span class="sourceLineNo">384</span>            key = null;<a name="line.384"></a>
+<span class="sourceLineNo">385</span>            hash = null;<a name="line.385"></a>
+<span class="sourceLineNo">386</span>            return false;<a name="line.386"></a>
+<span class="sourceLineNo">387</span>          }<a name="line.387"></a>
+<span class="sourceLineNo">388</span>        }<a name="line.388"></a>
+<span class="sourceLineNo">389</span>      }<a name="line.389"></a>
+<span class="sourceLineNo">390</span><a name="line.390"></a>
+<span class="sourceLineNo">391</span>      /**<a name="line.391"></a>
+<span class="sourceLineNo">392</span>       * Get the current key<a name="line.392"></a>
+<span class="sourceLineNo">393</span>       * @return the current key or null if there is no current key<a name="line.393"></a>
+<span class="sourceLineNo">394</span>       */<a name="line.394"></a>
+<span class="sourceLineNo">395</span>      public ImmutableBytesWritable getCurrentKey() {<a name="line.395"></a>
+<span class="sourceLineNo">396</span>        return key;<a name="line.396"></a>
+<span class="sourceLineNo">397</span>      }<a name="line.397"></a>
+<span class="sourceLineNo">398</span><a name="line.398"></a>
+<span class="sourceLineNo">399</span>      /**<a name="line.399"></a>
+<span class="sourceLineNo">400</span>       * Get the current hash<a name="line.400"></a>
+<span class="sourceLineNo">401</span>       * @return the current hash or null if there is no current hash<a name="line.401"></a>
+<span class="sourceLineNo">402</span>       */<a name="line.402"></a>
+<span class="sourceLineNo">403</span>      public ImmutableBytesWritable getCurrentHash() {<a name="line.403"></a>
+<span class="sourceLineNo">404</span>        return hash;<a name="line.404"></a>
+<span class="sourceLineNo">405</span>      }<a name="line.405"></a>
+<span class="sourceLineNo">406</span><a name="line.406"></a>
+<span class="sourceLineNo">407</span>      private void openHashFile() throws IOException {<a name="line.407"></a>
+<span class="sourceLineNo">408</span>        if (mapFileReader != null) {<a name="line.408"></a>
+<span class="sourceLineNo">409</span>          mapFileReader.close();<a name="line.409"></a>
+<span class="sourceLineNo">410</span>        }<a name="line.410"></a>
+<span class="sourceLineNo">411</span>        Path dataDir = new Path(TableHash.this.hashDir, HASH_DATA_DIR);<a name="line.411"></a>
+<span class="sourceLineNo">412</span>        Path dataFile = new Path(dataDir, getDataFileName(hashFileIndex));<a name="line.412"></a>
+<span class="sourceLineNo">413</span>        mapFileReader = new MapFile.Reader(dataFile, conf);<a name="line.413"></a>
+<span class="sourceLineNo">414</span>      }<a name="line.414"></a>
+<span class="sourceLineNo">415</span><a name="line.415"></a>
+<span class="sourceLineNo">416</span>      @Override<a name="line.416"></a>
+<span class="sourceLineNo">417</span>      public void close() throws IOException {<a name="line.417"></a>
+<span class="sourceLineNo">418</span>        mapFileReader.close();<a name="line.418"></a>
+<span class="sourceLineNo">419</span>      }<a name="line.419"></a>
+<span class="sourceLineNo">420</span>    }<a name="line.420"></a>
+<span class="sourceLineNo">421</span>  }<a name="line.421"></a>
+<span class="sourceLineNo">422</span><a name="line.422"></a>
+<span class="sourceLineNo">423</span>  static boolean isTableStartRow(byte[] row) {<a name="line.423"></a>
+<span class="sourceLineNo">424</span>    return Bytes.equals(HConstants.EMPTY_START_ROW, row);<a name="line.424"></a>
+<span class="sourceLineNo">425</span>  }<a name="line.425"></a>
+<span class="sourceLineNo">426</span><a name="line.426"></a>
+<span class="sourceLineNo">427</span>  static boolean isTableEndRow(byte[] row) {<a name="line.427"></a>
+<span class="sourceLineNo">428</span>    return Bytes.equals(HConstants.EMPTY_END_ROW, row);<a name="line.428"></a>
+<span class="sourceLineNo">429</span>  }<a name="line.429"></a>
+<span class="sourceLineNo">430</span><a name="line.430"></a>
+<span class="sourceLineNo">431</span>  public Job createSubmittableJob(String[] args) throws IOException {<a name="line.431"></a>
+<span class="sourceLineNo">432</span>    Path partitionsPath = new Path(destPath, PARTITIONS_FILE_NAME);<a name="line.432"></a>
+<span class="sourceLineNo">433</span>    generatePartitions(partitionsPath);<a name="line.433"></a>
+<span class="sourceLineNo">434</span><a name="line.434"></a>
+<span class="sourceLineNo">435</span>    Job job = Job.getInstance(getConf(),<a name="line.435"></a>
+<span class="sourceLineNo">436</span>          getConf().get("mapreduce.job.name", "hashTable_" + tableHash.tableName));<a name="line.436"></a>
+<span class="sourceLineNo">437</span>    Configuration jobConf = job.getConfiguration();<a name="line.437"></a>
+<span class="sourceLineNo">438</span>    jobConf.setLong(HASH_BATCH_SIZE_CONF_KEY, tableHash.batchSize);<a name="line.438"></a>
+<span class="sourceLineNo">439</span>    jobConf.setBoolean(IGNORE_TIMESTAMPS, tableHash.ignoreTimestamps);<a name="line.439"></a>
+<span class="sourceLineNo">440</span>    job.setJarByClass(HashTable.class);<a name="line.440"></a>
 <span class="sourceLineNo">441</span><a name="line.441"></a>
-<span class="sourceLineNo">442</span>    // use a TotalOrderPartitioner and reducers to group region output into hash files<a name="line.442"></a>
-<span class="sourceLineNo">443</span>    job.setPartitionerClass(TotalOrderPartitioner.class);<a name="line.443"></a>
-<span class="sourceLineNo">444</span>    TotalOrderPartitioner.setPartitionFile(jobConf, partitionsPath);<a name="line.444"></a>
-<span class="sourceLineNo">445</span>    job.setReducerClass(Reducer.class);  // identity reducer<a name="line.445"></a>
-<span class="sourceLineNo">446</span>    job.setNumReduceTasks(tableHash.numHashFiles);<a name="line.446"></a>
-<span class="sourceLineNo">447</span>    job.setOutputKeyClass(ImmutableBytesWritable.class);<a name="line.447"></a>
-<span class="sourceLineNo">448</span>    job.setOutputValueClass(ImmutableBytesWritable.class);<a name="line.448"></a>
-<span class="sourceLineNo">449</span>    job.setOutputFormatClass(MapFileOutputFormat.class);<a name="line.449"></a>
-<span class="sourceLineNo">450</span>    FileOutputFormat.setOutputPath(job, new Path(destPath, HASH_DATA_DIR));<a name="line.450"></a>
-<span class="sourceLineNo">451</span><a name="line.451"></a>
-<span class="sourceLineNo">452</span>    return job;<a name="line.452"></a>
-<span class="sourceLineNo">453</span>  }<a name="line.453"></a>
+<span class="sourceLineNo">442</span>    TableMapReduceUtil.initTableMapperJob(tableHash.tableName, tableHash.initScan(),<a name="line.442"></a>
+<span class="sourceLineNo">443</span>        HashMapper.class, ImmutableBytesWritable.class, ImmutableBytesWritable.class, job);<a name="line.443"></a>
+<span class="sourceLineNo">444</span><a name="line.444"></a>
+<span class="sourceLineNo">445</span>    // use a TotalOrderPartitioner and reducers to group region output into hash files<a name="line.445"></a>
+<span class="sourceLineNo">446</span>    job.setPartitionerClass(TotalOrderPartitioner.class);<a name="line.446"></a>
+<span class="sourceLineNo">447</span>    TotalOrderPartitioner.setPartitionFile(jobConf, partitionsPath);<a name="line.447"></a>
+<span class="sourceLineNo">448</span>    job.setReducerClass(Reducer.class);  // identity reducer<a name="line.448"></a>
+<span class="sourceLineNo">449</span>    job.setNumReduceTasks(tableHash.numHashFiles);<a name="line.449"></a>
+<span class="sourceLineNo">450</span>    job.setOutputKeyClass(ImmutableBytesWritable.class);<a name="line.450"></a>
+<span class="sourceLineNo">451</span>    job.setOutputValueClass(ImmutableBytesWritable.class);<a name="line.451"></a>
+<span class="sourceLineNo">452</span>    job.setOutputFormatClass(MapFileOutputFormat.class);<a name="line.452"></a>
+<span class="sourceLineNo">453</span>    FileOutputFormat.setOutputPath(job, new Path(destPath, HASH_DATA_DIR));<a name="line.453"></a>
 <span class="sourceLineNo">454</span><a name="line.454"></a>
-<span class="sourceLineNo">455</span>  private void generatePartitions(Path partitionsPath) throws IOException {<a name="line.455"></a>
-<span class="sourceLineNo">456</span>    Connection connection = ConnectionFactory.createConnection(getConf());<a name="line.456"></a>
-<span class="sourceLineNo">457</span>    Pair&lt;byte[][], byte[][]&gt; regionKeys<a name="line.457"></a>
-<span class="sourceLineNo">458</span>      = connection.getRegionLocator(TableName.valueOf(tableHash.tableName)).getStartEndKeys();<a name="line.458"></a>
-<span class="sourceLineNo">459</span>    connection.close();<a name="line.459"></a>
-<span class="sourceLineNo">460</span><a name="line.460"></a>
-<span class="sourceLineNo">461</span>    tableHash.selectPartitions(regionKeys);<a name="line.461"></a>
-<span class="sourceLineNo">462</span>    LOG.info("Writing " + tableHash.partitions.size() + " partition keys to " + partitionsPath);<a name="line.462"></a>
+<span class="sourceLineNo">455</span>    return job;<a name="line.455"></a>
+<span class="sourceLineNo">456</span>  }<a name="line.456"></a>
+<span class="sourceLineNo">457</span><a name="line.457"></a>
+<span class="sourceLineNo">458</span>  private void generatePartitions(Path partitionsPath) throws IOException {<a name="line.458"></a>
+<span class="sourceLineNo">459</span>    Connection connection = ConnectionFactory.createConnection(getConf());<a name="line.459"></a>
+<span class="sourceLineNo">460</span>    Pair&lt;byte[][], byte[][]&gt; regionKeys<a name="line.460"></a>
+<span class="sourceLineNo">461</span>      = connection.getRegionLocator(TableName.valueOf(tableHash.tableName)).getStartEndKeys();<a name="line.461"></a>
+<span class="sourceLineNo">462</span>    connection.close();<a name="line.462"></a>
 <span class="sourceLineNo">463</span><a name="line.463"></a>
-<span class="sourceLineNo">464</span>    tableHash.writePartitionFile(getConf(), partitionsPath);<a name="line.464"></a>
-<span class="sourceLineNo">465</span>  }<a name="line.465"></a>
+<span class="sourceLineNo">464</span>    tableHash.selectPartitions(regionKeys);<a name="line.464"></a>
+<span class="sourceLineNo">465</span>    LOG.info("Writing " + tableHash.partitions.size() + " partition keys to " + partitionsPath);<a name="line.465"></a>
 <span class="sourceLineNo">466</span><a name="line.466"></a>
-<span class="sourceLineNo">467</span>  static class ResultHasher {<a name="line.467"></a>
-<span class="sourceLineNo">468</span>    private MessageDigest digest;<a name="line.468"></a>
+<span class="sourceLineNo">467</span>    tableHash.writePartitionFile(getConf(), partitionsPath);<a name="line.467"></a>
+<span class="sourceLineNo">468</span>  }<a name="line.468"></a>
 <span class="sourceLineNo">469</span><a name="line.469"></a>
-<span class="sourceLineNo">470</span>    private boolean batchStarted = false;<a name="line.470"></a>
-<span class="sourceLineNo">471</span>    private ImmutableBytesWritable batchStartKey;<a name="line.471"></a>
-<span class="sourceLineNo">472</span>    private ImmutableBytesWritable batchHash;<a name="line.472"></a>
-<span class="sourceLineNo">473</span>    private long batchSize = 0;<a name="line.473"></a>
-<span class="sourceLineNo">474</span><a name="line.474"></a>
-<span class="sourceLineNo">475</span><a name="line.475"></a>
-<span class="sourceLineNo">476</span>    public ResultHasher() {<a name="line.476"></a>
-<span class="sourceLineNo">477</span>      try {<a name="line.477"></a>
-<span class="sourceLineNo">478</span>        digest = MessageDigest.getInstance("MD5");<a name="line.478"></a>
-<span class="sourceLineNo">479</span>      } catch (NoSuchAlgorithmException e) {<a name="line.479"></a>
-<span class="sourceLineNo">480</span>        Throwables.propagate(e);<a name="line.480"></a>
-<span class="sourceLineNo">481</span>      }<a name="line.481"></a>
-<span class="sourceLineNo">482</span>    }<a name="line.482"></a>
-<span class="sourceLineNo">483</span><a name="line.483"></a>
-<span class="sourceLineNo">484</span>    public void startBatch(ImmutableBytesWritable row) {<a name="line.484"></a>
-<span class="sourceLineNo">485</span>      if (batchStarted) {<a name="line.485"></a>
-<span class="sourceLineNo">486</span>        throw new RuntimeException("Cannot start new batch without finishing existing one.");<a name="line.486"></a>
-<span class="sourceLineNo">487</span>      }<a name="line.487"></a>
-<span class="sourceLineNo">488</span>      batchStarted = true;<a name="line.488"></a>
-<span class="sourceLineNo">489</span>      batchSize = 0;<a name="line.489"></a>
-<span class="sourceLineNo">490</span>      batchStartKey = row;<a name="line.490"></a>
-<span class="sourceLineNo">491</span>      batchHash = null;<a name="line.491"></a>
-<span class="sourceLineNo">492</span>    }<a name="line.492"></a>
-<span class="sourceLineNo">493</span><a name="line.493"></a>
-<span class="sourceLineNo">494</span>    public void hashResult(Result result) {<a name="line.494"></a>
-<span class="sourceLineNo">495</span>      if (!batchStarted) {<a name="line.495"></a>
-<span class="sourceLineNo">496</span>        throw new RuntimeException("Cannot add to batch that has not been started.");<a name="line.496"></a>
-<span class="sourceLineNo">497</span>      }<a name="line.497"></a>
-<span class="sourceLineNo">498</span>      for (Cell cell : result.rawCells()) {<a name="line.498"></a>
-<span class="sourceLineNo">499</span>        int rowLength = cell.getRowLength();<a name="line.499"></a>
-<span class="sourceLineNo">500</span>        int familyLength = cell.getFamilyLength();<a name="line.500"></a>
-<span class="sourceLineNo">501</span>        int qualifierLength = cell.getQualifierLength();<a name="line.501"></a>
-<span class="sourceLineNo">502</span>        int valueLength = cell.getValueLength();<a name="line.502"></a>
-<span class="sourceLineNo">503</span>        digest.update(cell.getRowArray(), cell.getRowOffset(), rowLength);<a name="line.503"></a>
-<span class="sourceLineNo">504</span>        digest.update(cell.getFamilyArray(), cell.getFamilyOffset(), familyLength);<a name="line.504"></a>
-<span class="sourceLineNo">505</span>        digest.update(cell.getQualifierArray(), cell.getQualifierOffset(), qualifierLength);<a name="line.505"></a>
-<span class="sourceLineNo">506</span>        long ts = cell.getTimestamp();<a name="line.506"></a>
-<span class="sourceLineNo">507</span>        for (int i = 8; i &gt; 0; i--) {<a name="line.507"></a>
-<span class="sourceLineNo">508</span>          digest.update((byte) ts);<a name="line.508"></a>
-<span class="sourceLineNo">509</span>          ts &gt;&gt;&gt;= 8;<a name="line.509"></a>
-<span class="sourceLineNo">510</span>        }<a name="line.510"></a>
-<span class="sourceLineNo">511</span>        digest.update(cell.getValueArray(), cell.getValueOffset(), valueLength);<a name="line.511"></a>
-<span class="sourceLineNo">512</span><a name="line.512"></a>
-<span class="sourceLineNo">513</span>        batchSize += rowLength + familyLength + qualifierLength + 8 + valueLength;<a name="line.513"></a>
-<span class="sourceLineNo">514</span>      }<a name="line.514"></a>
-<span class="sourceLineNo">515</span>    }<a name="line.515"></a>
-<span class="sourceLineNo">516</span><a name="line.516"></a>
-<span class="sourceLineNo">517</span>    public void finishBatch() {<a name="line.517"></a>
-<span class="sourceLineNo">518</span>      if (!batchStarted) {<a name="line.518"></a>
-<span class="sourceLineNo">519</span>        throw new RuntimeException("Cannot finish batch that has not started.");<a name="line.519"></a>
-<span class="sourceLineNo">520</span>      }<a name="line.520"></a>
-<span class="sourceLineNo">521</span>      batchStarted = false;<a name="line.521"></a>
-<span class="sourceLineNo">522</span>      batchHash = new ImmutableBytesWritable(digest.digest());<a name="line.522"></a>
-<span class="sourceLineNo">523</span>    }<a name="line.523"></a>
-<span class="sourceLineNo">524</span><a name="line.524"></a>
-<span class="sourceLineNo">525</span>    public boolean isBatchStarted() {<a name="line.525"></a>
-<span class="sourceLineNo">526</span>      return batchStarted;<a name="line.526"></a>
-<span class="sourceLineNo">527</span>    }<a name="line.527"></a>
-<span class="sourceLineNo">528</span><a name="line.528"></a>
-<span class="sourceLineNo">529</span>    public ImmutableBytesWritable getBatchStartKey() {<a name="line.529"></a>
-<span class="sourceLineNo">530</span>      return batchStartKey;<a name="line.530"></a>
-<span class="sourceLineNo">531</span>    }<a name="line.531"></a>
-<span class="sourceLineNo">532</span><a name="line.532"></a>
-<span class="sourceLineNo">533</span>    public ImmutableBytesWritable getBatchHash() {<a name="line.533"></a>
-<span class="sourceLineNo">534</span>      return batchHash;<a name="line.534"></a>
-<span class="sourceLineNo">535</span>    }<a name="line.535"></a>
-<span class="sourceLineNo">536</span><a name="line.536"></a>
-<span class="sourceLineNo">537</span>    public long getBatchSize() {<a name="line.537"></a>
-<span class="sourceLineNo">538</span>      return batchSize;<a name="line.538"></a>
-<span class="sourceLineNo">539</span>    }<a name="line.539"></a>
-<span class="sourceLineNo">540</span>  }<a name="line.540"></a>
-<span class="sourceLineNo">541</span><a name="line.541"></a>
-<span class="sourceLineNo">542</span>  public static class HashMapper<a name="line.542"></a>
-<span class="sourceLineNo">543</span>    extends TableMapper&lt;ImmutableBytesWritable, ImmutableBytesWritable&gt; {<a name="line.543"></a>
-<span class="sourceLineNo">544</span><a name="line.544"></a>
-<span class="sourceLineNo">545</span>    private ResultHasher hasher;<a name="line.545"></a>
-<span class="sourceLineNo">546</span>    private long targetBatchSize;<a name="line.546"></a>
-<span class="sourceLineNo">547</span><a name="line.547"></a>
-<span class="sourceLineNo">548</span>    private ImmutableBytesWritable currentRow;<a name="line.548"></a>
-<span class="sourceLineNo">549</span><a name="line.549"></a>
-<span class="sourceLineNo">550</span>    @Override<a name="line.550"></a>
-<span class="sourceLineNo">551</span>    protected void setup(Context context) throws IOException, InterruptedException {<a name="line.551"></a>
-<span class="sourceLineNo">552</span>      targetBatchSize = context.getConfiguration()<a name="line.552"></a>
-<span class="sourceLineNo">553</span>          .getLong(HASH_BATCH_SIZE_CONF_KEY, DEFAULT_BATCH_SIZE);<a name="line.553"></a>
-<span class="sourceLineNo">554</span>      hasher = new ResultHasher();<a name="line.554"></a>
-<span class="sourceLineNo">555</span><a name="line.555"></a>
-<span class="sourceLineNo">556</span>      TableSplit split = (TableSplit) context.getInputSplit();<a name="line.556"></a>
-<span class="sourceLineNo">557</span>      hasher.startBatch(new ImmutableBytesWritable(split.getStartRow()));<a name="line.557"></a>
-<span class="sourceLineNo">558</span>    }<a name="line.558"></a>
-<span class="sourceLineNo">559</span><a name="line.559"></a>
-<span class="sourceLineNo">560</span>    @Override<a name="line.560"></a>
-<span class="sourceLineNo">561</span>    protected void map(ImmutableBytesWritable key, Result value, Context context)<a name="line.561"></a>
-<span class="sourceLineNo">562</span>        throws IOException, InterruptedException {<a name="line.562"></a>
-<span class="sourceLineNo">563</span><a name="line.563"></a>
-<span class="sourceLineNo">564</span>      if (currentRow == null || !currentRow.equals(key)) {<a name="line.564"></a>
-<span class="sourceLineNo">565</span>        currentRow = new ImmutableBytesWritable(key); // not immutable<a name="line.565"></a>
-<span class="sourceLineNo">566</span><a name="line.566"></a>
-<span class="sourceLineNo">567</span>        if (hasher.getBatchSize() &gt;= targetBatchSize) {<a name="line.567"></a>
-<span class="sourceLineNo">568</span>          hasher.finishBatch();<a name="line.568"></a>
-<span class="sourceLineNo">569</span>          context.write(hasher.getBatchStartKey(), hasher.getBatchHash());<a name="line.569"></a>
-<span class="sourceLineNo">570</span>          hasher.startBatch(currentRow);<a name="line.570"></a>
-<span class="sourceLineNo">571</span>        }<a name="line.571"></a>
-<span class="sourceLineNo">572</span>      }<a name="line.572"></a>
-<span class="sourceLineNo">573</span><a name="line.573"></a>
-<span class="sourceLineNo">574</span>      hasher.hashResult(value);<a name="line.574"></a>
-<span class="sourceLineNo">575</span>    }<a name="line.575"></a>
-<span class="sourceLineNo">576</span><a name="line.576"></a>
-<span class="sourceLineNo">577</span>    @Override<a name="line.577"></a>
-<span class="sourceLineNo">578</span>    protected void cleanup(Context context) throws IOException, InterruptedException {<a name="line.578"></a>
-<span class="sourceLineNo">579</span>      hasher.finishBatch();<a name="line.579"></a>
-<span class="sourceLineNo">580</span>      context.write(hasher.getBatchStartKey(), hasher.getBatchHash());<a name="line.580"></a>
-<span class="sourceLineNo">581</span>    }<a name="line.581"></a>
-<span class="sourceLineNo">582</span>  }<a name="line.582"></a>
-<span class="sourceLineNo">583</span><a name="line.583"></a>
-<span class="sourceLineNo">584</span>  private void writeTempManifestFile() throws IOException {<a name="line.584"></a>
-<span class="sourceLineNo">585</span>    Path tempManifestPath = new Path(destPath, TMP_MANIFEST_FILE_NAME);<a name="line.585"></a>
-<span class="sourceLineNo">586</span>    FileSystem fs = tempManifestPath.getFileSystem(getConf());<a name="line.586"></a>
-<span class="sourceLineNo">587</span>    tableHash.writePropertiesFile(fs, tempManifestPath);<a name="line.587"></a>
-<span class="sourceLineNo">588</span>  }<a name="line.588"></a>
-<span class="sourceLineNo">589</span><a name="line.589"></a>
-<span class="sourceLineNo">590</span>  private void completeManifest() throws IOException {<a name="line.590"></a>
-<span class="sourceLineNo">591</span>    Path tempManifestPath = new Path(destPath, TMP_MANIFEST_FILE_NAME);<a name="line.591"></a>
-<span class="sourceLineNo">592</span>    Path manifestPath = new Path(destPath, MANIFEST_FILE_NAME);<a name="line.592"></a>
-<span class="sourceLineNo">593</span>    FileSystem fs = tempManifestPath.getFileSystem(getConf());<a name="line.593"></a>
-<span class="sourceLineNo">594</span>    fs.rename(tempManifestPath, manifestPath);<a name="line.594"></a>
-<span class="sourceLineNo">595</span>  }<a name="line.595"></a>
-<span class="sourceLineNo">596</span><a name="line.596"></a>
-<span class="sourceLineNo">597</span>  private static final int NUM_ARGS = 2;<a name="line.597"></a>
-<span class="sourceLineNo">598</span>  private static void printUsage(final String errorMsg) {<a name="line.598"></a>
-<span class="sourceLineNo">599</span>    if (errorMsg != null &amp;&amp; errorMsg.length() &gt; 0) {<a name="line.599"></a>
-<span class="sourceLineNo">600</span>      System.err.println("ERROR: " + errorMsg);<a name="line.600"></a>
-<span class="sourceLineNo">601</span>      System.err.println();<a name="line.601"></a>
-<span class="sourceLineNo">602</span>    }<a name="line.602"></a>
-<span class="sourceLineNo">603</span>    System.err.println("Usage: HashTable [options] &lt;tablename&gt; &lt;outputpath&gt;");<a name="line.603"></a>
-<span class="sourceLineNo">604</span>    System.err.println();<a name="line.604"></a>
-<span class="sourceLineNo">605</span>    System.err.println("Options:");<a name="line.605"></a>
-<span class="sourceLineNo">606</span>    System.err.println(" batchsize     the target amount of bytes to hash in each batch");<a name="line.606"></a>
-<span class="sourceLineNo">607</span>    System.err.println("               rows are added to the batch until this size is reached");<a name="line.607"></a>
-<span class="sourceLineNo">608</span>    System.err.println("               (defaults to " + DEFAULT_BATCH_SIZE + " bytes)");<a name="line.608"></a>
-<span class="sourceLineNo">609</span>    System.err.println(" numhashfiles  the number of hash files to create");<a name="line.609"></a>
-<span class="sourceLineNo">610</span>    System.err.println("               if set to fewer than number of regions then");<a name="line.610"></a>
-<span class="sourceLineNo">611</span>    System.err.println("               the job will create this number of reducers");<a name="line.611"></a>
-<span class="sourceLineNo">612</span>    System.err.println("               (defaults to 1/100 of regions -- at least 1)");<a name="line.612"></a>
-<span class="sourceLineNo">613</span>    System.err.println(" startrow      the start row");<a name="line.613"></a>
-<span class="sourceLineNo">614</span>    System.err.println(" stoprow       the stop row");<a name="line.614"></a>
-<span class="sourceLineNo">615</span>    System.err.println(" starttime     beginning of the time range (unixtime in millis)");<a name="line.615"></a>
-<span class="sourceLineNo">616</span>    System.err.println("               without endtime means from starttime to forever");<a name="line.616"></a>
-<span class="sourceLineNo">617</span>    System.err.println(" endtime       end of the time range.  Ignored if no starttime specified.");<a name="line.617"></a>
-<span class="sourceLineNo">618</span>    System.err.println(" scanbatch     scanner batch size to support intra row scans");<a name="line.618"></a>
-<span class="sourceLineNo">619</span>    System.err.println(" versions      number of cell versions to include");<a name="line.619"></a>
-<span class="sourceLineNo">620</span>    System.err.println(" families      comma-separated list of families to include");<a name="line.620"></a>
-<span class="sourceLineNo">621</span>    System.err.println();<a name="line.621"></a>
-<span class="sourceLineNo">622</span>    System.err.println("Args:");<a name="line.622"></a>
-<span class="sourceLineNo">623</span>    System.err.println(" tablename     Name of the table to hash");<a name="line.623"></a>
-<span class="sourceLineNo">624</span>    System.err.println(" outputpath    Filesystem path to put the output data");<a name="line.624"></a>
-<span class="sourceLineNo">625</span>    System.err.println();<a name="line.625"></a>
-<span class="sourceLineNo">626</span>    System.err.println("Examples:");<a name="line.626"></a>
-<span class="sourceLineNo">627</span>    System.err.println(" To hash 'TestTable' in 32kB batches for a 1 hour window into 50 files:");<a name="line.627"></a>
-<span class="sourceLineNo">628</span>    System.err.println(" $ hbase " +<a name="line.628"></a>
-<span class="sourceLineNo">629</span>        "org.apache.hadoop.hbase.mapreduce.HashTable --batchsize=32000 --numhashfiles=50"<a name="line.629"></a>
-<span class="sourceLineNo">630</span>        + " --starttime=1265875194289 --endtime=1265878794289 --families=cf2,cf3"<a name="line.630"></a>
-<span class="sourceLineNo">631</span>        + " TestTable /hashes/testTable");<a name="line.631"></a>
-<span class="sourceLineNo">632</span>  }<a name="line.632"></a>
-<span class="sourceLineNo">633</span><a name="line.633"></a>
-<span class="sourceLineNo">634</span>  private boolean doCommandLine(final String[] args) {<a name="line.634"></a>
-<span class="sourceLineNo">635</span>    if (args.length &lt; NUM_ARGS) {<a name="line.635"></a>
-<span class="sourceLineNo">636</span>      printUsage(null);<a name="line.636"></a>
-<span class="sourceLineNo">637</span>      return false;<a name="line.637"></a>
-<span class="sourceLineNo">638</span>    }<a name="line.638"></a>
-<span class="sourceLineNo">639</span>    try {<a name="line.639"></a>
-<span class="sourceLineNo">640</span><a name="line.640"></a>
-<span class="sourceLineNo">641</span>      tableHash.tableName = args[args.length-2];<a name="line.641"></a>
-<span class="sourceLineNo">642</span>      destPath = new Path(args[args.length-1]);<a name="line.642"></a>
-<span class="sourceLineNo">643</span><a name="line.643"></a>
-<span class="sourceLineNo">644</span>      for (int i = 0; i &lt; args.length - NUM_ARGS; i++) {<a name="line.644"></a>
-<span class="sourceLineNo">645</span>        String cmd = args[i];<a name="line.645"></a>
-<span class="sourceLineNo">646</span>        if (cmd.equals("-h") || cmd.startsWith("--h")) {<a name="line.646"></a>
-<span class="sourceLineNo">647</span>          printUsage(null);<a name="line.647"></a>
-<span class="sourceLineNo">648</span>          return false;<a name="line.648"></a>
-<span class="sourceLineNo">649</span>        }<a name="line.649"></a>
-<span class="sourceLineNo">650</span><a name="line.650"></a>
-<span class="sourceLineNo">651</span>        final String batchSizeArgKey = "--batchsize=";<a name="line.651"></a>
-<span class="sourceLineNo">652</span>        if (cmd.startsWith(batchSizeArgKey)) {<a name="line.652"></a>
-<span class="sourceLineNo">653</span>          tableHash.batchSize = Long.parseLong(cmd.substring(batchSizeArgKey.length()));<a name="line.653"></a>
-<span class="sourceLineNo">654</span>          continue;<a name="line.654"></a>
-<span class="sourceLineNo">655</span>        }<a name="line.655"></a>
-<span class="sourceLineNo">656</span><a name="line.656"></a>
-<span class="sourceLineNo">657</span>        final String numHashFilesArgKey = "--numhashfiles=";<a name="line.657"></a>
-<span class="sourceLineNo">658</span>        if (cmd.startsWith(numHashFilesArgKey)) {<a name="line.658"></a>
-<span class="sourceLineNo">659</span>          tableHash.numHashFiles = Integer.parseInt(cmd.substring(numHashFilesArgKey.length()));<a name="line.659"></a>
-<span class="sourceLineNo">660</span>          continue;<a name="line.660"></a>
-<span class="sourceLineNo">661</span>        }<a name="line.661"></a>
-<span class="sourceLineNo">662</span><a name="line.662"></a>
-<span class="sourceLineNo">663</span>        final String startRowArgKey = "--startrow=";<a name="line.663"></a>
-<span class="sourceLineNo">664</span>        if (cmd.startsWith(startRowArgKey)) {<a name="line.664"></a>
-<span class="sourceLineNo">665</span>          tableHash.startRow = Bytes.fromHex(cmd.substring(startRowArgKey.length()));<a name="line.665"></a>
-<span class="sourceLineNo">666</span>          continue;<a name="line.666"></a>
-<span class="sourceLineNo">667</span>        }<a name="line.667"></a>
-<span class="sourceLineNo">668</span><a name="line.668"></a>
-<span class="sourceLineNo">669</span>        final String stopRowArgKey = "--stoprow=";<a name="line.669"></a>
-<span class="sourceLineNo">670</span>        if (cmd.startsWith(stopRowArgKey)) {<a name="line.670"></a>
-<span class="sourceLineNo">671</span>          tableHash.stopRow = Bytes.fromHex(cmd.substring(stopRowArgKey.length()));<a name="line.671"></a>
-<span class="sourceLineNo">672</span>          continue;<a name="line.672"></a>
-<span class="sourceLineNo">673</span>        }<a name="line.673"></a>
-<span class="sourceLineNo">674</span><a name="line.674"></a>
-<span class="sourceLineNo">675</span>        final String startTimeArgKey = "--starttime=";<a name="line.675"></a>
-<span class="sourceLineNo">676</span>        if (cmd.startsWith(startTimeArgKey)) {<a name="line.676"></a>
-<span class="sourceLineNo">677</span>          tableHash.startTime = Long.parseLong(cmd.substring(startTimeArgKey.length()));<a name="line.677"></a>
-<span class="sourceLineNo">678</span>          continue;<a name="line.678"></a>
-<span class="sourceLineNo">679</span>        }<a name="line.679"></a>
-<span class="sourceLineNo">680</span><a name="line.680"></a>
-<span class="sourceLineNo">681</span>        final String endTimeArgKey = "--endtime=";<a name="line.681"></a>
-<span class="sourceLineNo">682</span>        if (cmd.startsWith(endTimeArgKey)) {<a name="line.682"></a>
-<span class="sourceLineNo">683</span>          tableHash.endTime = Long.parseLong(cmd.substring(endTimeArgKey.length()));<a name="line.683"></a>
-<span class="sourceLineNo">684</span>          continue;<a name="line.684"></a>
-<span class="sourceLineNo">685</span>        }<a name="line.685"></a>
-<span class="sourceLineNo">686</span><a name="line.686"></a>
-<span class="sourceLineNo">687</span>        final String scanBatchArgKey = "--scanbatch=";<a name="line.687"></a>
-<span class="sourceLineNo">688</span>        if (cmd.startsWith(scanBatchArgKey)) {<a name="line.688"></a>
-<span class="sourceLineNo">689</span>          tableHash.scanBatch = Integer.parseInt(cmd.substring(scanBatchArgKey.length()));<a name="line.689"></a>
-<span class="sourceLineNo">690</span>          continue;<a name="line.690"></a>
-<span class="sourceLineNo">691</span>        }<a name="line.691"></a>
-<span class="sourceLineNo">692</span><a name="line.692"></a>
-<span class="sourceLineNo">693</span>        final String versionsArgKey = "--versions=";<a name="line.693"></a>
-<span class="sourceLineNo">694</span>        if (cmd.startsWith(versionsArgKey)) {<a name="line.694"></a>
-<span class="sourceLineNo">695</span>          tableHash.versions = Integer.parseInt(cmd.substring(versionsArgKey.length()));<a name="line.695"></a>
-<span class="sourceLineNo">696</span>          continue;<a name="line.696"></a>
-<span class="sourceLineNo">697</span>        }<a name="line.697"></a>
-<span class="sourceLineNo">698</span><a name="line.698"></a>
-<span class="sourceLineNo">699</span>        final String familiesArgKey = "--families=";<a name="line.699"></a>
-<span class="sourceLineNo">700</span>        if (cmd.startsWith(familiesArgKey)) {<a name="line.700"></a>
-<span class="sourceLineNo">701</span>          tableHash.families = cmd.substring(familiesArgKey.length());<a name="line.701"></a>
-<span class="sourceLineNo">702</span>          continue;<a name="line.702"></a>
-<span class="sourceLineNo">703</span>        }<a name="line.703"></a>
-<span class="sourceLineNo">704</span><a name="line.704"></a>
-<span class="sourceLineNo">705</span>        printUsage("Invalid argument '" + cmd + "'");<a name="line.705"></a>
-<span class="sourceLineNo">706</span>        return false;<a name="line.706"></a>
-<span class="sourceLineNo">707</span>      }<a name="line.707"></a>
-<span class="sourceLineNo">708</span>      if ((tableHash.startTime != 0 || tableHash.endTime != 0)<a name="line.708"></a>
-<span class="sourceLineNo">709</span>          &amp;&amp; (tableHash.startTime &gt;= tableHash.endTime)) {<a name="line.709"></a>
-<span class="sourceLineNo">710</span>        printUsage("Invalid time range filter: starttime="<a name="line.710"></a>
-<span class="sourceLineNo">711</span>            + tableHash.startTime + " &gt;=  endtime=" + tableHash.endTime);<a name="line.711"></a>
-<span class="sourceLineNo">712</span>        return false;<a name="line.712"></a>
-<span class="sourceLineNo">713</span>      }<a name="line.713"></a>
-<span class="sourceLineNo">714</span><a name="line.714"></a>
-<span class="sourceLineNo">715</span>    } catch (Exception e) {<a name="line.715"></a>
-<span class="sourceLineNo">716</span>      LOG.error("Failed to parse commandLine arguments", e);<a name="line.716"></a>
-<span class="sourceLineNo">717</span>      printUsage("Can't start because " + e.getMessage());<a name="line.717"></a>
-<span class="sourceLineNo">718</span>      return false;<a name="line.718"></a>
-<span class="sourceLineNo">719</span>    }<a name="line.719"></a>
-<span class="sourceLineNo">720</span>    return true;<a name="line.720"></a>
-<span class="sourceLineNo">721</span>  }<a name="line.721"></a>
+<span class="sourceLineNo">470</span>  static class ResultHasher {<a name="line.470"></a>
+<span class="sourceLineNo">471</span>    private MessageDigest digest;<a name="line.471"></a>
+<span class="sourceLineNo">472</span><a name="line.472"></a>
+<span class="sourceLineNo">473</span>    private boolean batchStarted = false;<a name="line.473"></a>
+<span class="sourceLineNo">474</span>    private ImmutableBytesWritable batchStartKey;<a name="line.474"></a>
+<span class="sourceLineNo">475</span>    private ImmutableBytesWritable batchHash;<a name="line.475"></a>
+<span class="sourceLineNo">476</span>    private long batchSize = 0;<a name="line.476"></a>
+<span class="sourceLineNo">477</span>    boolean ignoreTimestamps;<a name="line.477"></a>
+<span class="sourceLineNo">478</span><a name="line.478"></a>
+<span class="sourceLineNo">479</span><a name="line.479"></a>
+<span class="sourceLineNo">480</span>    public ResultHasher() {<a name="line.480"></a>
+<span class="sourceLineNo">481</span>      try {<a name="line.481"></a>
+<span class="sourceLineNo">482</span>        digest = MessageDigest.getInstance("MD5");<a name="line.482"></a>
+<span class="sourceLineNo">483</span>      } catch (NoSuchAlgorithmException e) {<a name="line.483"></a>
+<span class="sourceLineNo">484</span>        Throwables.propagate(e);<a name="line.484"></a>
+<span class="sourceLineNo">485</span>      }<a name="line.485"></a>
+<span class="sourceLineNo">486</span>    }<a name="line.486"></a>
+<span class="sourceLineNo">487</span><a name="line.487"></a>
+<span class="sourceLineNo">488</span>    public void startBatch(ImmutableBytesWritable row) {<a name="line.488"></a>
+<span class="sourceLineNo">489</span>      if (batchStarted) {<a name="line.489"></a>
+<span class="sourceLineNo">490</span>        throw new RuntimeException("Cannot start new batch without finishing existing one.");<a name="line.490"></a>
+<span class="sourceLineNo">491</span>      }<a name="line.491"></a>
+<span class="sourceLineNo">492</span>      batchStarted = true;<a name="line.492"></a>
+<span class="sourceLineNo">493</span>      batchSize = 0;<a name="line.493"></a>
+<span class="sourceLineNo">494</span>      batchStartKey = row;<a name="line.494"></a>
+<span class="sourceLineNo">495</span>      batchHash = null;<a name="line.495"></a>
+<span class="sourceLineNo">496</span>    }<a name="line.496"></a>
+<span class="sourceLineNo">497</span><a name="line.497"></a>
+<span class="sourceLineNo">498</span>    public void hashResult(Result result) {<a name="line.498"></a>
+<span class="sourceLineNo">499</span>      if (!batchStarted) {<a name="line.499"></a>
+<span class="sourceLineNo">500</span>        throw new RuntimeException("Cannot add to batch that has not been started.");<a name="line.500"></a>
+<span class="sourceLineNo">501</span>      }<a name="line.501"></a>
+<span class="sourceLineNo">502</span>      for (Cell cell : result.rawCells()) {<a name="line.502"></a>
+<span class="sourceLineNo">503</span>        int rowLength = cell.getRowLength();<a name="line.503"></a>
+<span class="sourceLineNo">504</span>        int familyLength = cell.getFamilyLength();<a name="line.504"></a>
+<span class="sourceLineNo">505</span>        int qualifierLength = cell.getQualifierLength();<a name="line.505"></a>
+<span class="sourceLineNo">506</span>        int valueLength = cell.getValueLength();<a name="line.506"></a>
+<span class="sourceLineNo">507</span>        digest.update(cell.getRowArray(), cell.getRowOffset(), rowLength);<a name="line.507"></a>
+<span class="sourceLineNo">508</span>        digest.update(cell.getFamilyArray(), cell.getFamilyOffset(), familyLength);<a name="line.508"></a>
+<span class="sourceLineNo">509</span>        digest.update(cell.getQualifierArray(), cell.getQualifierOffset(), qualifierLength);<a name="line.509"></a>
+<span class="sourceLineNo">510</span><a name="line.510"></a>
+<span class="sourceLineNo">511</span>        if (!ignoreTimestamps) {<a name="line.511"></a>
+<span class="sourceLineNo">512</span>          long ts = cell.getTimestamp();<a name="line.512"></a>
+<span class="sourceLineNo">513</span>          for (int i = 8; i &gt; 0; i--) {<a name="line.513"></a>
+<span class="sourceLineNo">514</span>            digest.update((byte) ts);<a name="line.514"></a>
+<span class="sourceLineNo">515</span>            ts &gt;&gt;&gt;= 8;<a name="line.515"></a>
+<span class="sourceLineNo">516</span>          }<a name="line.516"></a>
+<span class="sourceLineNo">517</span>        }<a name="line.517"></a>
+<span class="sourceLineNo">518</span>        digest.update(cell.getValueArray(), cell.getValueOffset(), valueLength);<a name="line.518"></a>
+<span class="sourceLineNo">519</span><a name="line.519"></a>
+<span class="sourceLineNo">520</span>        batchSize += rowLength + familyLength + qualifierLength + 8 + valueLength;<a name="line.520"></a>
+<span class="sourceLineNo">521</span>      }<a name="line.521"></a>
+<span class="sourceLineNo">522</span>    }<a name="line.522"></a>
+<span class="sourceLineNo">523</span><a name="line.523"></a>
+<span class="sourceLineNo">524</span>    public void finishBatch() {<a name="line.524"></a>
+<span class="sourceLineNo">525</span>      if (!batchStarted) {<a name="line.525"></a>
+<span class="sourceLineNo">526</span>        throw new RuntimeException("Cannot finish batch that has not started.");<a name="line.526"></a>
+<span class="sourceLineNo">527</span>      }<a name="line.527"></a>
+<span class="sourceLineNo">528</span>      batchStarted = false;<a name="line.528"></a>
+<span class="sourceLineNo">529</span>      batchHash = new ImmutableBytesWritable(digest.digest());<a name="line.529"></a>
+<span class="sourceLineNo">530</span>    }<a name="line.530"></a>
+<span class="sourceLineNo">531</span><a name="line.531"></a>
+<span class="sourceLineNo">532</span>    public boolean isBatchStarted() {<a name="line.532"></a>
+<span class="sourceLineNo">533</span>      return batchStarted;<a name="line.533"></a>
+<span class="sourceLineNo">534</span>    }<a name="line.534"></a>
+<span class="sourceLineNo">535</span><a name="line.535"></a>
+<span class="sourceLineNo">536</span>    public ImmutableBytesWritable getBatchStartKey() {<a name="line.536"></a>
+<span class="sourceLineNo">537</span>      return batchStartKey;<a name="line.537"></a>
+<span class="sourceLineNo">538</span>    }<a name="line.538"></a>
+<span class="sourceLineNo">539</span><a name="line.539"></a>
+<span class="sourceLineNo">540</span>    public ImmutableBytesWritable getBatchHash() {<a name="line.540"></a>
+<span class="sourceLineNo">541</span>      return batchHash;<a name="line.541"></a>
+<span class="sourceLineNo">542</span>    }<a name="line.542"></a>
+<span class="sourceLineNo">543</span><a name="line.543"></a>
+<span class="sourceLineNo">544</span>    public long getBatchSize() {<a name="line.544"></a>
+<span class="sourceLineNo">545</span>      return batchSize;<a name="line.545"></a>
+<span class="sourceLineNo">546</span>    }<a name="line.546"></a>
+<span class="sourceLineNo">547</span>  }<a name="line.547"></a>
+<span class="sourceLineNo">548</span><a name="line.548"></a>
+<span class="sourceLineNo">549</span>  public static class HashMapper<a name="line.549"></a>
+<span class="sourceLineNo">550</span>    extends TableMapper&lt;ImmutableBytesWritable, ImmutableBytesWritable&gt; {<a name="line.550"></a>
+<span class="sourceLineNo">551</span><a name="line.551"></a>
+<span class="sourceLineNo">552</span>    private ResultHasher hasher;<a name="line.552"></a>
+<span class="sourceLineNo">553</span>    private long targetBatchSize;<a name="line.553"></a>
+<span class="sourceLineNo">554</span><a name="line.554"></a>
+<span class="sourceLineNo">555</span>    private ImmutableBytesWritable currentRow;<a name="line.555"></a>
+<span class="sourceLineNo">556</span><a name="line.556"></a>
+<span class="sourceLineNo">557</span>    @Override<a name="line.557"></a>
+<span class="sourceLineNo">558</span>    protected void setup(Context context) throws IOException, InterruptedException {<a name="line.558"></a>
+<span class="sourceLineNo">559</span>      targetBatchSize = context.getConfiguration()<a name="line.559"></a>
+<span class="sourceLineNo">560</span>          .getLong(HASH_BATCH_SIZE_CONF_KEY, DEFAULT_BATCH_SIZE);<a name="line.560"></a>
+<span class="sourceLineNo">561</span>      hasher = new ResultHasher();<a name="line.561"></a>
+<span class="sourceLineNo">562</span>      hasher.ignoreTimestamps = context.getConfiguration().<a name="line.562"></a>
+<span class="sourceLineNo">563</span>        getBoolean(IGNORE_TIMESTAMPS, false);<a name="line.563"></a>
+<span class="sourceLineNo">564</span>      TableSplit split = (TableSplit) context.getInputSplit();<a name="line.564"></a>
+<span class="sourceLineNo">565</span>      hasher.startBatch(new ImmutableBytesWritable(split.getStartRow()));<a name="line.565"></a>
+<span class="sourceLineNo">566</span>    }<a name="line.566"></a>
+<span class="sourceLineNo">567</span><a name="line.567"></a>
+<span class="sourceLineNo">568</span>    @Override<a name="line.568"></a>
+<span class="sourceLineNo">569</span>    protected void map(ImmutableBytesWritable key, Result value, Context context)<a name="line.569"></a>
+<span class="sourceLineNo">570</span>        throws IOException, InterruptedException {<a name="line.570"></a>
+<span class="sourceLineNo">571</span><a name="line.571"></a>
+<span class="sourceLineNo">572</span>      if (currentRow == null || !currentRow.equals(key)) {<a name="line.572"></a>
+<span class="sourceLineNo">573</span>        currentRow = new ImmutableBytesWritable(key); // not immutable<a name="line.573"></a>
+<span class="sourceLineNo">574</span><a name="line.574"></a>
+<span class="sourceLineNo">575</span>        if (hasher.getBatchSize() &gt;= targetBatchSize) {<a name="line.575"></a>
+<span class="sourceLineNo">576</span>          hasher.finishBatch();<a name="line.576"></a>
+<span class="sourceLineNo">577</span>          context.write(hasher.getBatchStartKey(), hasher.getBatchHash());<a name="line.577"></a>
+<span class="sourceLineNo">578</span>          hasher.startBatch(currentRow);<a name="line.578"></a>
+<span class="sourceLineNo">579</span>        }<a name="line.579"></a>
+<span class="sourceLineNo">580</span>      }<a name="line.580"></a>
+<span class="sourceLineNo">581</span><a name="line.581"></a>
+<span class="sourceLineNo">582</span>      hasher.hashResult(value);<a name="line.582"></a>
+<span class="sourceLineNo">583</span>    }<a name="line.583"></a>
+<span class="sourceLineNo">584</span><a name="line.584"></a>
+<span class="sourceLineNo">585</span>    @Override<a name="line.585"></a>
+<span class="sourceLineNo">586</span>    protected void cleanup(Context context) throws IOException, InterruptedException {<a name="line.586"></a>
+<span class="sourceLineNo">587</span>      hasher.finishBatch();<a name="line.587"></a>
+<span class="sourceLineNo">588</span>      context.write(hasher.getBatchStartKey(), hasher.getBatchHash());<a name="line.588"></a>
+<span class="sourceLineNo">589</span>    }<a name="line.589"></a>
+<span class="sourceLineNo">590</span>  }<a name="line.590"></a>
+<span class="sourceLineNo">591</span><a name="line.591"></a>
+<span class="sourceLineNo">592</span>  private void writeTempManifestFile() throws IOException {<a name="line.592"></a>
+<span class="sourceLineNo">593</span>    Path tempManifestPath = new Path(destPath, TMP_MANIFEST_FILE_NAME);<a name="line.593"></a>
+<span class="sourceLineNo">594</span>    FileSystem fs = tempManifestPath.getFileSystem(getConf());<a name="line.594"></a>
+<span class="sourceLineNo">595</span>    tableHash.writePropertiesFile(fs, tempManifestPath);<a name="line.595"></a>
+<span class="sourceLineNo">596</span>  }<a name="line.596"></a>
+<span class="sourceLineNo">597</span><a name="line.597"></a>
+<span class="sourceLineNo">598</span>  private void completeManifest() throws IOException {<a name="line.598"></a>
+<span class="sourceLineNo">599</span>    Path tempManifestPath = new Path(destPath, TMP_MANIFEST_FILE_NAME);<a name="line.599"></a>
+<span class="sourceLineNo">600</span>    Path manifestPath = new Path(destPath, MANIFEST_FILE_NAME);<a name="line.600"></a>
+<span class="sourceLineNo">601</span>    FileSystem fs = tempManifestPath.getFileSystem(getConf());<a name="line.601"></a>
+<span class="sourceLineNo">602</span>    fs.rename(tempManifestPath, manifestPath);<a name="line.602"></a>
+<span class="sourceLineNo">603</span>  }<a name="line.603"></a>
+<span class="sourceLineNo">604</span><a name="line.604"></a>
+<span class="sourceLineNo">605</span>  private static final int NUM_ARGS = 2;<a name="line.605"></a>
+<span class="sourceLineNo">606</span>  private static void printUsage(final String errorMsg) {<a name="line.606"></a>
+<span class="sourceLineNo">607</span>    if (errorMsg != null &amp;&amp; errorMsg.length() &gt; 0) {<a name="line.607"></a>
+<span class="sourceLineNo">608</span>      System.err.println("ERROR: " + errorMsg);<a name="line.608"></a>
+<span class="sourceLineNo">609</span>      System.err.println();<a name="line.609"></a>
+<span class="sourceLineNo">610</span>    }<a name="line.610"></a>
+<span class="sourceLineNo">611</span>    System.err.println("Usage: HashTable [options] &lt;tablename&gt; &lt;outputpath&gt;");<a name="line.611"></a>
+<span class="sourceLineNo">612</span>    System.err.println();<a name="line.612"></a>
+<span class="sourceLineNo">613</span>    System.err.println("Options:");<a name="line.613"></a>
+<span class="sourceLineNo">614</span>    System.err.println(" batchsize         the target amount of bytes to hash in each batch");<a name="line.614"></a>
+<span class="sourceLineNo">615</span>    System.err.println("                   rows are added to the batch until this size is reached");<a name="line.615"></a>
+<span class="sourceLineNo">616</span>    System.err.println("                   (defaults to " + DEFAULT_BATCH_SIZE + " bytes)");<a name="line.616"></a>
+<span class="sourceLineNo">617</span>    System.err.println(" numhashfiles      the number of hash files to create");<a name="line.617"></a>
+<span class="sourceLineNo">618</span>    System.err.println("                   if set to fewer than number of regions then");<a name="line.618"></a>
+<span class="sourceLineNo">619</span>    System.err.println("                   the job will create this number of reducers");<a name="line.619"></a>
+<span class="sourceLineNo">620</span>    System.err.println("                   (defaults to 1/100 of regions -- at least 1)");<a name="line.620"></a>
+<span class="sourceLineNo">621</span>    System.err.println(" startrow          the start row");<a name="line.621"></a>
+<span class="sourceLineNo">622</span>    System.err.println(" stoprow           the stop row");<a name="line.622"></a>
+<span class="sourceLineNo">623</span>    System.err.println(" starttime         beginning of the time range (unixtime in millis)");<a name="line.623"></a>
+<span class="sourceLineNo">624</span>    System.err.println("                   without endtime means from starttime to forever");<a name="line.624"></a>
+<span class="sourceLineNo">625</span>    System.err.println(" endtime           end of the time range.");<a name="line.625"></a>
+<span class="sourceLineNo">626</span>    System.err.println("                   Ignored if no starttime specified.");<a name="line.626"></a>
+<span class="sourceLineNo">627</span>    System.err.println(" scanbatch         scanner batch size to support intra row scans");<a name="line.627"></a>
+<span class="sourceLineNo">628</span>    System.err.println(" versions          number of cell versions to include");<a name="line.628"></a>
+<span class="sourceLineNo">629</span>    System.err.println(" families          comma-separated list of families to include");<a name="line.629"></a>
+<span class="sourceLineNo">630</span>    System.err.println(" ignoreTimestamps  if true, ignores cell timestamps");<a name="line.630"></a>
+<span class="sourceLineNo">631</span>    System.err.println("                   when calculating hashes");<a name="line.631"></a>
+<span class="sourceLineNo">632</span>    System.err.println();<a name="line.632"></a>
+<span class="sourceLineNo">633</span>    System.err.println("Args:");<a name="line.633"></a>
+<span class="sourceLineNo">634</span>    System.err.println(" tablename     Name of the table to hash");<a name="line.634"></a>
+<span class="sourceLineNo">635</span>    System.err.println(" outputpath    Filesystem path to put the output data");<a name="line.635"></a>
+<span class="sourceLineNo">636</span>    System.err.println();<a name="line.636"></a>
+<span class="sourceLineNo">637</span>    System.err.println("Examples:");<a name="line.637"></a>
+<span class="sourceLineNo">638</span>    System.err.println(" To hash 'TestTable' in 32kB batches for a 1 hour window into 50 files:");<a name="line.638"></a>
+<span class="sourceLineNo">639</span>    System.err.println(" $ hbase " +<a name="line.639"></a>
+<span class="sourceLineNo">640</span>        "org.apache.hadoop.hbase.mapreduce.HashTable --batchsize=32000 --numhashfiles=50"<a name="line.640"></a>
+<span class="sourceLineNo">641</span>        + " --starttime=1265875194289 --endtime=1265878794289 --families=cf2,cf3"<a name="line.641"></a>
+<span class="sourceLineNo">642</span>        + " TestTable /hashes/testTable");<a name="line.642"></a>
+<span class="sourceLineNo">643</span>  }<a name="line.643"></a>
+<span class="sourceLineNo">644</span><a name="line.644"></a>
+<span class="sourceLineNo">645</span>  private boolean doCommandLine(final String[] args) {<a name="line.645"></a>
+<span class="sourceLineNo">646</span>    if (args.length &lt; NUM_ARGS) {<a name="line.646"></a>
+<span class="sourceLineNo">647</span>      printUsage(null);<a name="line.647"></a>
+<span class="sourceLineNo">648</span>      return false;<a name="line.648"></a>
+<span class="sourceLineNo">649</span>    }<a name="line.649"></a>
+<span class="sourceLineNo">650</span>    try {<a name="line.650"></a>
+<span class="sourceLineNo">651</span><a name="line.651"></a>
+<span class="sourceLineNo">652</span>      tableHash.tableName = args[args.length-2];<a name="line.652"></a>
+<span class="sourceLineNo">653</span>      destPath = new Path(args[args.length-1]);<a name="line.653"></a>
+<span class="sourceLineNo">654</span><a name="line.654"></a>
+<span class="sourceLineNo">655</span>      for (int i = 0; i &lt; args.length - NUM_ARGS; i++) {<a name="line.655"></a>
+<span class="sourceLineNo">656</span>        String cmd = args[i];<a name="line.656"></a>
+<span class="sourceLineNo">657</span>        if (cmd.equals("-h") || cmd.startsWith("--h")) {<a name="line.657"></a>
+<span class="sourceLineNo">658</span>          printUsage(null);<a name="line.658"></a>
+<span class="sourceLineNo">659</span>          return false;<a name="line.659"></a>
+<span class="sourceLineNo">660</span>        }<a name="line.660"></a>
+<span class="sourceLineNo">661</span><a name="line.661"></a>
+<span class="sourceLineNo">662</span>        final String batchSizeArgKey = "--batchsize=";<a name="line.662"></a>
+<span class="sourceLineNo">663</span>        if (cmd.startsWith(batchSizeArgKey)) {<a name="line.663"></a>
+<span class="sourceLineNo">664</span>          tableHash.batchSize = Long.parseLong(cmd.substring(batchSizeArgKey.length()));<a name="line.664"></a>
+<span class="sourceLineNo">665</span>          continue;<a name="line.665"></a>
+<span class="sourceLineNo">666</span>        }<a name="line.666"></a>
+<span class="sourceLineNo">667</span><a name="line.667"></a>
+<span class="sourceLineNo">668</span>        final String numHashFilesArgKey = "--numhashfiles=";<a name="line.668"></a>
+<span class="sourceLineNo">669</span>        if (cmd.startsWith(numHashFilesArgKey)) {<a name="line.669"></a>
+<span class="sourceLineNo">670</span>          tableHash.numHashFiles = Integer.parseInt(cmd.substring(numHashFilesArgKey.length()));<a name="line.670"></a>
+<span class="sourceLineNo">671</span>          continue;<a name="line.671"></a>
+<span class="sourceLineNo">672</span>        }<a name="line.672"></a>
+<span class="sourceLineNo">673</span><a name="line.673"></a>
+<span class="sourceLineNo">674</span>        final String startRowArgKey = "--startrow=";<a name="line.674"></a>
+<span class="sourceLineNo">675</span>        if (cmd.startsWith(startRowArgKey)) {<a name="line.675"></a>
+<span class="sourceLineNo">676</span>          tableHash.startRow = Bytes.fromHex(cmd.substring(startRowArgKey.length()));<a name="line.676"></a>
+<span class="sourceLineNo">677</span>          continue;<a name="line.677"></a>
+<span class="sourceLineNo">678</span>        }<a name="line.678"></a>
+<span class="sourceLineNo">679</span><a name="line.679"></a>
+<span class="sourceLineNo">680</span>        final String stopRowArgKey = "--stoprow=";<a name="line.680"></a>
+<span class="sourceLineNo">681</span>        if (cmd.startsWith(stopRowArgKey)) {<a name="line.681"></a>
+<span class="sourceLineNo">682</span>          tableHash.stopRow = Bytes.fromHex(cmd.substring(stopRowArgKey.length()));<a name="line.682"></a>
+<span class="sourceLineNo">683</span>          continue;<a name="line.683"></a>
+<span class="sourceLineNo">684</span>        }<a name="line.684"></a>
+<span class="sourceLineNo">685</span><a name="line.685"></a>
+<span class="sourceLineNo">686</span>        final String startTimeArgKey = "--starttime=";<a name="line.686"></a>
+<span class="sourceLineNo">687</span>        if (cmd.startsWith(startTimeArgKey)) {<a name="line.687"></a>
+<span class="sourceLineNo">688</span>          tableHash.startTime = Long.parseLong(cmd.substring(startTimeArgKey.length()));<a name="line.688"></a>
+<span class="sourceLineNo">689</span>          continue;<a name="line.689"></a>
+<span class="sourceLineNo">690</span>        }<a name="line.690"></a>
+<span class="sourceLineNo">691</span><a name="line.691"></a>
+<span class="sourceLineNo">692</span>        final String endTimeArgKey = "--endtime=";<a name="line.692"></a>
+<span class="sourceLineNo">693</span>        if (cmd.startsWith(endTimeArgKey)) {<a name="line.693"></a>
+<span class="sourceLineNo">694</span>          tableHash.endTime = Long.parseLong(cmd.substring(endTimeArgKey.length()));<a name="line.694"></a>
+<span class="sourceLineNo">695</span>          continue;<a name="line.695"></a>
+<span class="sourceLineNo">696</span>        }<a name="line.696"></a>
+<span class="sourceLineNo">697</span><a name="line.697"></a>
+<span class="sourceLineNo">698</span>        final String scanBatchArgKey = "--scanbatch=";<a name="line.698"></a>
+<span class="sourceLineNo">699</span>        if (cmd.startsWith(scanBatchArgKey)) {<a name="line.699"></a>
+<span class="sourceLineNo">700</span>          tableHash.scanBatch = Integer.parseInt(cmd.substring(scanBatchArgKey.length()));<a name="line.700"></a>
+<span class="sourceLineNo">701</span>          continue;<a name="line.701"></a>
+<span class="sourceLineNo">702</span>        }<a name="line.702"></a>
+<span class="sourceLineNo">703</span><a name="line.703"></a>
+<span class="sourceLineNo">704</span>        final String versionsArgKey = "--versions=";<a name="line.704"></a>
+<span class="sourceLineNo">705</span>        if (cmd.startsWith(versionsArgKey)) {<a name="line.705"></a>
+<span class="sourceLineNo">706</span>          tableHash.versions = Integer.parseInt(cmd.substring(versionsArgKey.length()));<a name="line.706"></a>
+<span class="sourceLineNo">707</span>          continue;<a name="line.707"></a>
+<span class="sourceLineNo">708</span>        }<a name="line.708"></a>
+<span class="sourceLineNo">709</span><a name="line.709"></a>
+<span class="sourceLineNo">710</span>        final String familiesArgKey = "--families=";<a name="line.710"></a>
+<span class="sourceLineNo">711</span>        if (cmd.startsWith(familiesArgKey)) {<a name="line.711"></a>
+<span class="sourceLineNo">712</span>          tableHash.families = cmd.substring(familiesArgKey.length());<a name="line.712"></a>
+<span class="sourceLineNo">713</span>          continue;<a name="line.713"></a>
+<span class="sourceLineNo">714</span>        }<a name="line.714"></a>
+<span class="sourceLineNo">715</span><a name="line.715"></a>
+<span class="sourceLineNo">716</span>        final String ignoreTimestampsKey = "--ignoreTimestamps=";<a name="line.716"></a>
+<span class="sourceLineNo">717</span>        if (cmd.startsWith(ignoreTimestampsKey)) {<a name="line.717"></a>
+<span class="sourceLineNo">718</span>          tableHash.ignoreTimestamps = Boolean.<a name="line.718"></a>
+<span class="sourceLineNo">719</span>            parseBoolean(cmd.substring(ignoreTimestampsKey.length()));<a name="line.719"></a>
+<span class="sourceLineNo">720</span>          continue;<a name="line.720"></a>
+<span class="sourceLineNo">721</span>        }<a name="line.721"></a>
 <span class="sourceLineNo">722</span><a name="line.722"></a>
-<span class="sourceLineNo">723</span>  /**<a name="line.723"></a>
-<span class="sourceLineNo">724</span>   * Main entry point.<a name="line.724"></a>
-<span class="sourceLineNo">725</span>   */<a name="line.725"></a>
-<span class="sourceLineNo">726</span>  public static void main(String[] args) throws Exception {<a name="line.726"></a>
-<span class="sourceLineNo">727</span>    int ret = ToolRunner.run(new HashTable(HBaseConfiguration.create()), args);<a name="line.727"></a>
-<span class="sourceLineNo">728</span>    System.exit(ret);<a name="line.728"></a>
-<span class="sourceLineNo">729</span>  }<a name="line.729"></a>
-<span class="sourceLineNo">730</span><a name="line.730"></a>
-<span class="sourceLineNo">731</span>  @Override<a name="line.731"></a>
-<span class="sourceLineNo">732</span>  public int run(String[] args) throws Exception {<a name="line.732"></a>
-<span class="sourceLineNo">733</span>    String[] otherArgs = new GenericOptionsParser(getConf(), args).getRemainingArgs();<a name="line.733"></a>
-<span class="sourceLineNo">734</span>    if (!doCommandLine(otherArgs)) {<a name="line.734"></a>
-<span class="sourceLineNo">735</span>      return 1;<a name="line.735"></a>
-<span class="sourceLineNo">736</span>    }<a name="line.736"></a>
-<span class="sourceLineNo">737</span><a name="line.737"></a>
-<span class="sourceLineNo">738</span>    Job job = createSubmittableJob(otherArgs);<a name="line.738"></a>
-<span class="sourceLineNo">739</span>    writeTempManifestFile();<a name="line.739"></a>
-<span class="sourceLineNo">740</span>    if (!job.waitForCompletion(true)) {<a name="line.740"></a>
-<span class="sourceLineNo">741</span>      LOG.info("Map-reduce job failed!");<a name="line.741"></a>
-<span class="sourceLineNo">742</span>      return 1;<a name="line.742"></a>
-<span class="sourceLineNo">743</span>    }<a name="line.743"></a>
-<span class="sourceLineNo">744</span>    completeManifest();<a name="line.744"></a>
-<span class="sourceLineNo">745</span>    return 0;<a name="line.745"></a>
-<span class="sourceLineNo">746</span>  }<a name="line.746"></a>
-<span class="sourceLineNo">747</span><a name="line.747"></a>
-<span class="sourceLineNo">748</span>}<a name="line.748"></a>
+<span class="sourceLineNo">723</span>        printUsage("Invalid argument '" + cmd + "'");<a name="line.723"></a>
+<span class="sourceLineNo">724</span>        return false;<a name="line.724"></a>
+<span class="sourceLineNo">725</span>      }<a name="line.725"></a>
+<span class="sourceLineNo">726</span>      if ((tableHash.startTime != 0 || tableHash.endTime != 0)<a name="line.726"></a>
+<span class="sourceLineNo">727</span>          &amp;&amp; (tableHash.startTime &gt;= tableHash.endTime)) {<a name="line.727"></a>
+<span class="sourceLineNo">728</span>        printUsage("Invalid time range filter: starttime="<a name="line.728"></a>
+<span class="sourceLineNo">729</span>            + tableHash.startTime + " &gt;=  endtime=" + tableHash.endTime);<a name="line.729"></a>
+<span class="sourceLineNo">730</span>        return false;<a name="line.730"></a>
+<span class="sourceLineNo">731</span>      }<a name="line.731"></a>
+<span class="sourceLineNo">732</span><a name="line.732"></a>
+<span class="sourceLineNo">733</span>    } catch (Exception e) {<a name="line.733"></a>
+<span class="sourceLineNo">734</span>      LOG.error("Failed to parse commandLine arguments", e);<a name="line.734"></a>
+<span class="sourceLineNo">735</span>      printUsage("Can't start because " + e.getMessage());<a name="line.735"></a>
+<span class="sourceLineNo">736</span>      return false;<a name="line.736"></a>
+<span class="sourceLineNo">737</span>    }<a name="line.737"></a>
+<span class="sourceLineNo">738</span>    return true;<a name="line.738"></a>
+<span class="sourceLineNo">739</span>  }<a name="line.739"></a>
+<span class="sourceLineNo">740</span><a name="line.740"></a>
+<span class="sourceLineNo">741</span>  /**<a name="line.741"></a>
+<span class="sourceLineNo">742</span>   * Main entry point.<a name="line.742"></a>
+<span class="sourceLineNo">743</span>   */<a name="line.743"></a>
+<span class="sourceLineNo">744</span>  public static void main(String[] args) throws Exception {<a name="line.744"></a>
+<span class="sourceLineNo">745</span>    int ret = ToolRunner.run(new HashTable(HBaseConfiguration.create()), args);<a name="line.745"></a>
+<span class="sourceLineNo">746</span>    System.exit(ret);<a name="line.746"></a>
+<span class="sourceLineNo">747</span>  }<a name="line.747"></a>
+<span class="sourceLineNo">748</span><a name="line.748"></a>
+<span class="sourceLineNo">749</span>  @Override<a name="line.749"></a>
+<span class="sourceLineNo">750</span>  public int run(String[] args) throws Exception {<a name="line.750"></a>
+<span class="sourceLineNo">751</span>    String[] otherArgs = new GenericOptionsParser(getConf(), args).getRemainingArgs();<a name="line.751"></a>
+<span class="sourceLineNo">752</span>    if (!doCommandLine(otherArgs)) {<a name="line.752"></a>
+<span class="sourceLineNo">753</span>      return 1;<a name="line.753"></a>
+<span class="sourceLineNo">754</span>    }<a name="line.754"></a>
+<span class="sourceLineNo">755</span><a name="line.755"></a>
+<span class="sourceLineNo">756</span>    Job job = createSubmittableJob(otherArgs);<a name="line.756"></a>
+<span class="sourceLineNo">757</span>    writeTempManifestFile();<a name="line.757"></a>
+<span class="sourceLineNo">758</span>    if (!job.waitForCompletion(true)) {<a name="line.758"></a>
+<span class="sourceLineNo">759</span>      LOG.info("Map-reduce job failed!");<a name="line.759"></a>
+<span class="sourceLineNo">760</span>      return 1;<a name="line.760"></a>
+<span class="sourceLineNo">761</span>    }<a name="line.761"></a>
+<span class="sourceLineNo">762</span>    completeManifest();<a name="line.762"></a>
+<span class="sourceLineNo">763</span>    return 0;<a name="line.763"></a>
+<span class="sourceLineNo">764</span>  }<a name="line.764"></a>
+<span class="sourceLineNo">765</span><a name="line.765"></a>
+<span class="sourceLineNo">766</span>}<a name="line.766"></a>
 
 
 
diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html
index 55fc1c9..32934c7 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HashTable.TableHash.html
@@ -81,679 +81,697 @@
 <span class="sourceLineNo">073</span>  final static String MANIFEST_FILE_NAME = "manifest";<a name="line.73"></a>
 <span class="sourceLineNo">074</span>  final static String HASH_DATA_DIR = "hashes";<a name="line.74"></a>
 <span class="sourceLineNo">075</span>  final static String OUTPUT_DATA_FILE_PREFIX = "part-r-";<a name="line.75"></a>
-<span class="sourceLineNo">076</span>  private final static String TMP_MANIFEST_FILE_NAME = "manifest.tmp";<a name="line.76"></a>
-<span class="sourceLineNo">077</span><a name="line.77"></a>
-<span class="sourceLineNo">078</span>  TableHash tableHash = new TableHash();<a name="line.78"></a>
-<span class="sourceLineNo">079</span>  Path destPath;<a name="line.79"></a>
-<span class="sourceLineNo">080</span><a name="line.80"></a>
-<span class="sourceLineNo">081</span>  public HashTable(Configuration conf) {<a name="line.81"></a>
-<span class="sourceLineNo">082</span>    super(conf);<a name="line.82"></a>
-<span class="sourceLineNo">083</span>  }<a name="line.83"></a>
-<span class="sourceLineNo">084</span><a name="line.84"></a>
-<span class="sourceLineNo">085</span>  public static class TableHash {<a name="line.85"></a>
-<span class="sourceLineNo">086</span><a name="line.86"></a>
-<span class="sourceLineNo">087</span>    Path hashDir;<a name="line.87"></a>
-<span class="sourceLineNo">088</span><a name="line.88"></a>
-<span class="sourceLineNo">089</span>    String tableName;<a name="line.89"></a>
-<span class="sourceLineNo">090</span>    String families = null;<a name="line.90"></a>
-<span class="sourceLineNo">091</span>    long batchSize = DEFAULT_BATCH_SIZE;<a name="line.91"></a>
-<span class="sourceLineNo">092</span>    int numHashFiles = 0;<a name="line.92"></a>
-<span class="sourceLineNo">093</span>    byte[] startRow = HConstants.EMPTY_START_ROW;<a name="line.93"></a>
-<span class="sourceLineNo">094</span>    byte[] stopRow = HConstants.EMPTY_END_ROW;<a name="line.94"></a>
-<span class="sourceLineNo">095</span>    int scanBatch = 0;<a name="line.95"></a>
-<span class="sourceLineNo">096</span>    int versions = -1;<a name="line.96"></a>
-<span class="sourceLineNo">097</span>    long startTime = 0;<a name="line.97"></a>
-<span class="sourceLineNo">098</span>    long endTime = 0;<a name="line.98"></a>
-<span class="sourceLineNo">099</span><a name="line.99"></a>
-<span class="sourceLineNo">100</span>    List&lt;ImmutableBytesWritable&gt; partitions;<a name="line.100"></a>
+<span class="sourceLineNo">076</span>  final static String IGNORE_TIMESTAMPS = "ignoreTimestamps";<a name="line.76"></a>
+<span class="sourceLineNo">077</span>  private final static String TMP_MANIFEST_FILE_NAME = "manifest.tmp";<a name="line.77"></a>
+<span class="sourceLineNo">078</span><a name="line.78"></a>
+<span class="sourceLineNo">079</span>  TableHash tableHash = new TableHash();<a name="line.79"></a>
+<span class="sourceLineNo">080</span>  Path destPath;<a name="line.80"></a>
+<span class="sourceLineNo">081</span><a name="line.81"></a>
+<span class="sourceLineNo">082</span>  public HashTable(Configuration conf) {<a name="line.82"></a>
+<span class="sourceLineNo">083</span>    super(conf);<a name="line.83"></a>
+<span class="sourceLineNo">084</span>  }<a name="line.84"></a>
+<span class="sourceLineNo">085</span><a name="line.85"></a>
+<span class="sourceLineNo">086</span>  public static class TableHash {<a name="line.86"></a>
+<span class="sourceLineNo">087</span><a name="line.87"></a>
+<span class="sourceLineNo">088</span>    Path hashDir;<a name="line.88"></a>
+<span class="sourceLineNo">089</span><a name="line.89"></a>
+<span class="sourceLineNo">090</span>    String tableName;<a name="line.90"></a>
+<span class="sourceLineNo">091</span>    String families = null;<a name="line.91"></a>
+<span class="sourceLineNo">092</span>    long batchSize = DEFAULT_BATCH_SIZE;<a name="line.92"></a>
+<span class="sourceLineNo">093</span>    int numHashFiles = 0;<a name="line.93"></a>
+<span class="sourceLineNo">094</span>    byte[] startRow = HConstants.EMPTY_START_ROW;<a name="line.94"></a>
+<span class="sourceLineNo">095</span>    byte[] stopRow = HConstants.EMPTY_END_ROW;<a name="line.95"></a>
+<span class="sourceLineNo">096</span>    int scanBatch = 0;<a name="line.96"></a>
+<span class="sourceLineNo">097</span>    int versions = -1;<a name="line.97"></a>
+<span class="sourceLineNo">098</span>    long startTime = 0;<a name="line.98"></a>
+<span class="sourceLineNo">099</span>    long endTime = 0;<a name="line.99"></a>
+<span class="sourceLineNo">100</span>    boolean ignoreTimestamps;<a name="line.100"></a>
 <span class="sourceLineNo">101</span><a name="line.101"></a>
-<span class="sourceLineNo">102</span>    public static TableHash read(Configuration conf, Path hashDir) throws IOException {<a name="line.102"></a>
-<span class="sourceLineNo">103</span>      TableHash tableHash = new TableHash();<a name="line.103"></a>
-<span class="sourceLineNo">104</span>      FileSystem fs = hashDir.getFileSystem(conf);<a name="line.104"></a>
-<span class="sourceLineNo">105</span>      tableHash.hashDir = hashDir;<a name="line.105"></a>
-<span class="sourceLineNo">106</span>      tableHash.readPropertiesFile(fs, new Path(hashDir, MANIFEST_FILE_NAME));<a name="line.106"></a>
-<span class="sourceLineNo">107</span>      tableHash.readPartitionFile(fs, conf, new Path(hashDir, PARTITIONS_FILE_NAME));<a name="line.107"></a>
-<span class="sourceLineNo">108</span>      return tableHash;<a name="line.108"></a>
-<span class="sourceLineNo">109</span>    }<a name="line.109"></a>
-<span class="sourceLineNo">110</span><a name="line.110"></a>
-<span class="sourceLineNo">111</span>    void writePropertiesFile(FileSystem fs, Path path) throws IOException {<a name="line.111"></a>
-<span class="sourceLineNo">112</span>      Properties p = new Properties();<a name="line.112"></a>
-<span class="sourceLineNo">113</span>      p.setProperty("table", tableName);<a name="line.113"></a>
-<span class="sourceLineNo">114</span>      if (families != null) {<a name="line.114"></a>
-<span class="sourceLineNo">115</span>        p.setProperty("columnFamilies", families);<a name="line.115"></a>
-<span class="sourceLineNo">116</span>      }<a name="line.116"></a>
-<span class="sourceLineNo">117</span>      p.setProperty("targetBatchSize", Long.toString(batchSize));<a name="line.117"></a>
-<span class="sourceLineNo">118</span>      p.setProperty("numHashFiles", Integer.toString(numHashFiles));<a name="line.118"></a>
-<span class="sourceLineNo">119</span>      if (!isTableStartRow(startRow)) {<a name="line.119"></a>
-<span class="sourceLineNo">120</span>        p.setProperty("startRowHex", Bytes.toHex(startRow));<a name="line.120"></a>
-<span class="sourceLineNo">121</span>      }<a name="line.121"></a>
-<span class="sourceLineNo">122</span>      if (!isTableEndRow(stopRow)) {<a name="line.122"></a>
-<span class="sourceLineNo">123</span>        p.setProperty("stopRowHex", Bytes.toHex(stopRow));<a name="line.123"></a>
-<span class="sourceLineNo">124</span>      }<a name="line.124"></a>
-<span class="sourceLineNo">125</span>      if (scanBatch &gt; 0) {<a name="line.125"></a>
-<span class="sourceLineNo">126</span>        p.setProperty("scanBatch", Integer.toString(scanBatch));<a name="line.126"></a>
-<span class="sourceLineNo">127</span>      }<a name="line.127"></a>
-<span class="sourceLineNo">128</span>      if (versions &gt;= 0) {<a name="line.128"></a>
-<span class="sourceLineNo">129</span>        p.setProperty("versions", Integer.toString(versions));<a name="line.129"></a>
-<span class="sourceLineNo">130</span>      }<a name="line.130"></a>
-<span class="sourceLineNo">131</span>      if (startTime != 0) {<a name="line.131"></a>
-<span class="sourceLineNo">132</span>        p.setProperty("startTimestamp", Long.toString(startTime));<a name="line.132"></a>
-<span class="sourceLineNo">133</span>      }<a name="line.133"></a>
-<span class="sourceLineNo">134</span>      if (endTime != 0) {<a name="line.134"></a>
-<span class="sourceLineNo">135</span>        p.setProperty("endTimestamp", Long.toString(endTime));<a name="line.135"></a>
-<span class="sourceLineNo">136</span>      }<a name="line.136"></a>
-<span class="sourceLineNo">137</span><a name="line.137"></a>
-<span class="sourceLineNo">138</span>      try (OutputStreamWriter osw = new OutputStreamWriter(fs.create(path), Charsets.UTF_8)) {<a name="line.138"></a>
-<span class="sourceLineNo">139</span>        p.store(osw, null);<a name="line.139"></a>
-<span class="sourceLineNo">140</span>      }<a name="line.140"></a>
-<span class="sourceLineNo">141</span>    }<a name="line.141"></a>
-<span class="sourceLineNo">142</span><a name="line.142"></a>
-<span class="sourceLineNo">143</span>    void readPropertiesFile(FileSystem fs, Path path) throws IOException {<a name="line.143"></a>
-<span class="sourceLineNo">144</span>      Properties p = new Properties();<a name="line.144"></a>
-<span class="sourceLineNo">145</span>      try (FSDataInputStream in = fs.open(path)) {<a name="line.145"></a>
-<span class="sourceLineNo">146</span>        try (InputStreamReader isr = new InputStreamReader(in, Charsets.UTF_8)) {<a name="line.146"></a>
-<span class="sourceLineNo">147</span>          p.load(isr);<a name="line.147"></a>
-<span class="sourceLineNo">148</span>        }<a name="line.148"></a>
-<span class="sourceLineNo">149</span>      }<a name="line.149"></a>
-<span class="sourceLineNo">150</span>      tableName = p.getProperty("table");<a name="line.150"></a>
-<span class="sourceLineNo">151</span>      families = p.getProperty("columnFamilies");<a name="line.151"></a>
-<span class="sourceLineNo">152</span>      batchSize = Long.parseLong(p.getProperty("targetBatchSize"));<a name="line.152"></a>
-<span class="sourceLineNo">153</span>      numHashFiles = Integer.parseInt(p.getProperty("numHashFiles"));<a name="line.153"></a>
-<span class="sourceLineNo">154</span><a name="line.154"></a>
-<span class="sourceLineNo">155</span>      String startRowHex = p.getProperty("startRowHex");<a name="line.155"></a>
-<span class="sourceLineNo">156</span>      if (startRowHex != null) {<a name="line.156"></a>
-<span class="sourceLineNo">157</span>        startRow = Bytes.fromHex(startRowHex);<a name="line.157"></a>
-<span class="sourceLineNo">158</span>      }<a name="line.158"></a>
-<span class="sourceLineNo">159</span>      String stopRowHex = p.getProperty("stopRowHex");<a name="line.159"></a>
-<span class="sourceLineNo">160</span>      if (stopRowHex != null) {<a name="line.160"></a>
-<span class="sourceLineNo">161</span>        stopRow = Bytes.fromHex(stopRowHex);<a name="line.161"></a>
-<span class="sourceLineNo">162</span>      }<a name="line.162"></a>
-<span class="sourceLineNo">163</span><a name="line.163"></a>
-<span class="sourceLineNo">164</span>      String scanBatchString = p.getProperty("scanBatch");<a name="line.164"></a>
-<span class="sourceLineNo">165</span>      if (scanBatchString != null) {<a name="line.165"></a>
-<span class="sourceLineNo">166</span>        scanBatch = Integer.parseInt(scanBatchString);<a name="line.166"></a>
-<span class="sourceLineNo">167</span>      }<a name="line.167"></a>
-<span class="sourceLineNo">168</span><a name="line.168"></a>
-<span class="sourceLineNo">169</span>      String versionString = p.getProperty("versions");<a name="line.169"></a>
-<span class="sourceLineNo">170</span>      if (versionString != null) {<a name="line.170"></a>
-<span class="sourceLineNo">171</span>        versions = Integer.parseInt(versionString);<a name="line.171"></a>
-<span class="sourceLineNo">172</span>      }<a name="line.172"></a>
-<span class="sourceLineNo">173</span><a name="line.173"></a>
-<span class="sourceLineNo">174</span>      String startTimeString = p.getProperty("startTimestamp");<a name="line.174"></a>
-<span class="sourceLineNo">175</span>      if (startTimeString != null) {<a name="line.175"></a>
-<span class="sourceLineNo">176</span>        startTime = Long.parseLong(startTimeString);<a name="line.176"></a>
-<span class="sourceLineNo">177</span>      }<a name="line.177"></a>
-<span class="sourceLineNo">178</span><a name="line.178"></a>
-<span class="sourceLineNo">179</span>      String endTimeString = p.getProperty("endTimestamp");<a name="line.179"></a>
-<span class="sourceLineNo">180</span>      if (endTimeString != null) {<a name="line.180"></a>
-<span class="sourceLineNo">181</span>        endTime = Long.parseLong(endTimeString);<a name="line.181"></a>
-<span class="sourceLineNo">182</span>      }<a name="line.182"></a>
-<span class="sourceLineNo">183</span>    }<a name="line.183"></a>
-<span class="sourceLineNo">184</span><a name="line.184"></a>
-<span class="sourceLineNo">185</span>    Scan initScan() throws IOException {<a name="line.185"></a>
-<span class="sourceLineNo">186</span>      Scan scan = new Scan();<a name="line.186"></a>
-<span class="sourceLineNo">187</span>      scan.setCacheBlocks(false);<a name="line.187"></a>
-<span class="sourceLineNo">188</span>      if (startTime != 0 || endTime != 0) {<a name="line.188"></a>
-<span class="sourceLineNo">189</span>        scan.setTimeRange(startTime, endTime == 0 ? HConstants.LATEST_TIMESTAMP : endTime);<a name="line.189"></a>
-<span class="sourceLineNo">190</span>      }<a name="line.190"></a>
-<span class="sourceLineNo">191</span>      if (scanBatch &gt; 0) {<a name="line.191"></a>
-<span class="sourceLineNo">192</span>        scan.setBatch(scanBatch);<a name="line.192"></a>
-<span class="sourceLineNo">193</span>      }<a name="line.193"></a>
-<span class="sourceLineNo">194</span>      if (versions &gt;= 0) {<a name="line.194"></a>
-<span class="sourceLineNo">195</span>        scan.readVersions(versions);<a name="line.195"></a>
-<span class="sourceLineNo">196</span>      }<a name="line.196"></a>
-<span class="sourceLineNo">197</span>      if (!isTableStartRow(startRow)) {<a name="line.197"></a>
-<span class="sourceLineNo">198</span>        scan.withStartRow(startRow);<a name="line.198"></a>
-<span class="sourceLineNo">199</span>      }<a name="line.199"></a>
-<span class="sourceLineNo">200</span>      if (!isTableEndRow(stopRow)) {<a name="line.200"></a>
-<span class="sourceLineNo">201</span>        scan.withStopRow(stopRow);<a name="line.201"></a>
-<span class="sourceLineNo">202</span>      }<a name="line.202"></a>
-<span class="sourceLineNo">203</span>      if(families != null) {<a name="line.203"></a>
-<span class="sourceLineNo">204</span>        for(String fam : families.split(",")) {<a name="line.204"></a>
-<span class="sourceLineNo">205</span>          scan.addFamily(Bytes.toBytes(fam));<a name="line.205"></a>
-<span class="sourceLineNo">206</span>        }<a name="line.206"></a>
-<span class="sourceLineNo">207</span>      }<a name="line.207"></a>
-<span class="sourceLineNo">208</span>      return scan;<a name="line.208"></a>
-<span class="sourceLineNo">209</span>    }<a name="line.209"></a>
-<span class="sourceLineNo">210</span><a name="line.210"></a>
-<span class="sourceLineNo">211</span>    /**<a name="line.211"></a>
-<span class="sourceLineNo">212</span>     * Choose partitions between row ranges to hash to a single output file<a name="line.212"></a>
-<span class="sourceLineNo">213</span>     * Selects region boundaries that fall within the scan range, and groups them<a name="line.213"></a>
-<span class="sourceLineNo">214</span>     * into the desired number of partitions.<a name="line.214"></a>
-<span class="sourceLineNo">215</span>     */<a name="line.215"></a>
-<span class="sourceLineNo">216</span>    void selectPartitions(Pair&lt;byte[][], byte[][]&gt; regionStartEndKeys) {<a name="line.216"></a>
-<span class="sourceLineNo">217</span>      List&lt;byte[]&gt; startKeys = new ArrayList&lt;&gt;();<a name="line.217"></a>
-<span class="sourceLineNo">218</span>      for (int i = 0; i &lt; regionStartEndKeys.getFirst().length; i++) {<a name="line.218"></a>
-<span class="sourceLineNo">219</span>        byte[] regionStartKey = regionStartEndKeys.getFirst()[i];<a name="line.219"></a>
-<span class="sourceLineNo">220</span>        byte[] regionEndKey = regionStartEndKeys.getSecond()[i];<a name="line.220"></a>
-<span class="sourceLineNo">221</span><a name="line.221"></a>
-<span class="sourceLineNo">222</span>        // if scan begins after this region, or starts before this region, then drop this region<a name="line.222"></a>
-<span class="sourceLineNo">223</span>        // in other words:<a name="line.223"></a>
-<span class="sourceLineNo">224</span>        //   IF (scan begins before the end of this region<a name="line.224"></a>
-<span class="sourceLineNo">225</span>        //      AND scan ends before the start of this region)<a name="line.225"></a>
-<span class="sourceLineNo">226</span>        //   THEN include this region<a name="line.226"></a>
-<span class="sourceLineNo">227</span>        if ((isTableStartRow(startRow) || isTableEndRow(regionEndKey)<a name="line.227"></a>
-<span class="sourceLineNo">228</span>            || Bytes.compareTo(startRow, regionEndKey) &lt; 0)<a name="line.228"></a>
-<span class="sourceLineNo">229</span>          &amp;&amp; (isTableEndRow(stopRow) || isTableStartRow(regionStartKey)<a name="line.229"></a>
-<span class="sourceLineNo">230</span>            || Bytes.compareTo(stopRow, regionStartKey) &gt; 0)) {<a name="line.230"></a>
-<span class="sourceLineNo">231</span>          startKeys.add(regionStartKey);<a name="line.231"></a>
-<span class="sourceLineNo">232</span>        }<a name="line.232"></a>
-<span class="sourceLineNo">233</span>      }<a name="line.233"></a>
-<span class="sourceLineNo">234</span><a name="line.234"></a>
-<span class="sourceLineNo">235</span>      int numRegions = startKeys.size();<a name="line.235"></a>
-<span class="sourceLineNo">236</span>      if (numHashFiles == 0) {<a name="line.236"></a>
-<span class="sourceLineNo">237</span>        numHashFiles = numRegions / 100;<a name="line.237"></a>
-<span class="sourceLineNo">238</span>      }<a name="line.238"></a>
-<span class="sourceLineNo">239</span>      if (numHashFiles == 0) {<a name="line.239"></a>
-<span class="sourceLineNo">240</span>        numHashFiles = 1;<a name="line.240"></a>
-<span class="sourceLineNo">241</span>      }<a name="line.241"></a>
-<span class="sourceLineNo">242</span>      if (numHashFiles &gt; numRegions) {<a name="line.242"></a>
-<span class="sourceLineNo">243</span>        // can't partition within regions<a name="line.243"></a>
-<span class="sourceLineNo">244</span>        numHashFiles = numRegions;<a name="line.244"></a>
-<span class="sourceLineNo">245</span>      }<a name="line.245"></a>
-<span class="sourceLineNo">246</span><a name="line.246"></a>
-<span class="sourceLineNo">247</span>      // choose a subset of start keys to group regions into ranges<a name="line.247"></a>
-<span class="sourceLineNo">248</span>      partitions = new ArrayList&lt;&gt;(numHashFiles - 1);<a name="line.248"></a>
-<span class="sourceLineNo">249</span>      // skip the first start key as it is not a partition between ranges.<a name="line.249"></a>
-<span class="sourceLineNo">250</span>      for (long i = 1; i &lt; numHashFiles; i++) {<a name="line.250"></a>
-<span class="sourceLineNo">251</span>        int splitIndex = (int) (numRegions * i / numHashFiles);<a name="line.251"></a>
-<span class="sourceLineNo">252</span>        partitions.add(new ImmutableBytesWritable(startKeys.get(splitIndex)));<a name="line.252"></a>
-<span class="sourceLineNo">253</span>      }<a name="line.253"></a>
-<span class="sourceLineNo">254</span>    }<a name="line.254"></a>
-<span class="sourceLineNo">255</span><a name="line.255"></a>
-<span class="sourceLineNo">256</span>    void writePartitionFile(Configuration conf, Path path) throws IOException {<a name="line.256"></a>
-<span class="sourceLineNo">257</span>      FileSystem fs = path.getFileSystem(conf);<a name="line.257"></a>
-<span class="sourceLineNo">258</span>      @SuppressWarnings("deprecation")<a name="line.258"></a>
-<span class="sourceLineNo">259</span>      SequenceFile.Writer writer = SequenceFile.createWriter(<a name="line.259"></a>
-<span class="sourceLineNo">260</span>        fs, conf, path, ImmutableBytesWritable.class, NullWritable.class);<a name="line.260"></a>
-<span class="sourceLineNo">261</span><a name="line.261"></a>
-<span class="sourceLineNo">262</span>      for (int i = 0; i &lt; partitions.size(); i++) {<a name="line.262"></a>
-<span class="sourceLineNo">263</span>        writer.append(partitions.get(i), NullWritable.get());<a name="line.263"></a>
-<span class="sourceLineNo">264</span>      }<a name="line.264"></a>
-<span class="sourceLineNo">265</span>      writer.close();<a name="line.265"></a>
-<span class="sourceLineNo">266</span>    }<a name="line.266"></a>
-<span class="sourceLineNo">267</span><a name="line.267"></a>
-<span class="sourceLineNo">268</span>    private void readPartitionFile(FileSystem fs, Configuration conf, Path path)<a name="line.268"></a>
-<span class="sourceLineNo">269</span>         throws IOException {<a name="line.269"></a>
-<span class="sourceLineNo">270</span>      @SuppressWarnings("deprecation")<a name="line.270"></a>
-<span class="sourceLineNo">271</span>      SequenceFile.Reader reader = new SequenceFile.Reader(fs, path, conf);<a name="line.271"></a>
-<span class="sourceLineNo">272</span>      ImmutableBytesWritable key = new ImmutableBytesWritable();<a name="line.272"></a>
-<span class="sourceLineNo">273</span>      partitions = new ArrayList&lt;&gt;();<a name="line.273"></a>
-<span class="sourceLineNo">274</span>      while (reader.next(key)) {<a name="line.274"></a>
-<span class="sourceLineNo">275</span>        partitions.add(new ImmutableBytesWritable(key.copyBytes()));<a name="line.275"></a>
-<span class="sourceLineNo">276</span>      }<a name="line.276"></a>
-<span class="sourceLineNo">277</span>      reader.close();<a name="line.277"></a>
-<span class="sourceLineNo">278</span><a name="line.278"></a>
-<span class="sourceLineNo">279</span>      if (!Ordering.natural().isOrdered(partitions)) {<a name="line.279"></a>
-<span class="sourceLineNo">280</span>        throw new IOException("Partitions are not ordered!");<a name="line.280"></a>
-<span class="sourceLineNo">281</span>      }<a name="line.281"></a>
-<span class="sourceLineNo">282</span>    }<a name="line.282"></a>
-<span class="sourceLineNo">283</span><a name="line.283"></a>
-<span class="sourceLineNo">284</span>    @Override<a name="line.284"></a>
-<span class="sourceLineNo">285</span>    public String toString() {<a name="line.285"></a>
-<span class="sourceLineNo">286</span>      StringBuilder sb = new StringBuilder();<a name="line.286"></a>
-<span class="sourceLineNo">287</span>      sb.append("tableName=").append(tableName);<a name="line.287"></a>
-<span class="sourceLineNo">288</span>      if (families != null) {<a name="line.288"></a>
-<span class="sourceLineNo">289</span>        sb.append(", families=").append(families);<a name="line.289"></a>
-<span class="sourceLineNo">290</span>      }<a name="line.290"></a>
-<span class="sourceLineNo">291</span>      sb.append(", batchSize=").append(batchSize);<a name="line.291"></a>
-<span class="sourceLineNo">292</span>      sb.append(", numHashFiles=").append(numHashFiles);<a name="line.292"></a>
-<span class="sourceLineNo">293</span>      if (!isTableStartRow(startRow)) {<a name="line.293"></a>
-<span class="sourceLineNo">294</span>        sb.append(", startRowHex=").append(Bytes.toHex(startRow));<a name="line.294"></a>
-<span class="sourceLineNo">295</span>      }<a name="line.295"></a>
-<span class="sourceLineNo">296</span>      if (!isTableEndRow(stopRow)) {<a name="line.296"></a>
-<span class="sourceLineNo">297</span>        sb.append(", stopRowHex=").append(Bytes.toHex(stopRow));<a name="line.297"></a>
-<span class="sourceLineNo">298</span>      }<a name="line.298"></a>
-<span class="sourceLineNo">299</span>      if (scanBatch &gt;= 0) {<a name="line.299"></a>
-<span class="sourceLineNo">300</span>        sb.append(", scanBatch=").append(scanBatch);<a name="line.300"></a>
-<span class="sourceLineNo">301</span>      }<a name="line.301"></a>
-<span class="sourceLineNo">302</span>      if (versions &gt;= 0) {<a name="line.302"></a>
-<span class="sourceLineNo">303</span>        sb.append(", versions=").append(versions);<a name="line.303"></a>
-<span class="sourceLineNo">304</span>      }<a name="line.304"></a>
-<span class="sourceLineNo">305</span>      if (startTime != 0) {<a name="line.305"></a>
-<span class="sourceLineNo">306</span>        sb.append("startTime=").append(startTime);<a name="line.306"></a>
-<span class="sourceLineNo">307</span>      }<a name="line.307"></a>
-<span class="sourceLineNo">308</span>      if (endTime != 0) {<a name="line.308"></a>
-<span class="sourceLineNo">309</span>        sb.append("endTime=").append(endTime);<a name="line.309"></a>
-<span class="sourceLineNo">310</span>      }<a name="line.310"></a>
-<span class="sourceLineNo">311</span>      return sb.toString();<a name="line.311"></a>
-<span class="sourceLineNo">312</span>    }<a name="line.312"></a>
-<span class="sourceLineNo">313</span><a name="line.313"></a>
-<span class="sourceLineNo">314</span>    static String getDataFileName(int hashFileIndex) {<a name="line.314"></a>
-<span class="sourceLineNo">315</span>      return String.format(HashTable.OUTPUT_DATA_FILE_PREFIX + "%05d", hashFileIndex);<a name="line.315"></a>
-<span class="sourceLineNo">316</span>    }<a name="line.316"></a>
-<span class="sourceLineNo">317</span><a name="line.317"></a>
-<span class="sourceLineNo">318</span>    /**<a name="line.318"></a>
-<span class="sourceLineNo">319</span>     * Open a TableHash.Reader starting at the first hash at or after the given key.<a name="line.319"></a>
-<span class="sourceLineNo">320</span>     * @throws IOException<a name="line.320"></a>
-<span class="sourceLineNo">321</span>     */<a name="line.321"></a>
-<span class="sourceLineNo">322</span>    public Reader newReader(Configuration conf, ImmutableBytesWritable startKey)<a name="line.322"></a>
-<span class="sourceLineNo">323</span>        throws IOException {<a name="line.323"></a>
-<span class="sourceLineNo">324</span>      return new Reader(conf, startKey);<a name="line.324"></a>
-<span class="sourceLineNo">325</span>    }<a name="line.325"></a>
-<span class="sourceLineNo">326</span><a name="line.326"></a>
-<span class="sourceLineNo">327</span>    public class Reader implements java.io.Closeable {<a name="line.327"></a>
-<span class="sourceLineNo">328</span>      private final Configuration conf;<a name="line.328"></a>
-<span class="sourceLineNo">329</span><a name="line.329"></a>
-<span class="sourceLineNo">330</span>      private int hashFileIndex;<a name="line.330"></a>
-<span class="sourceLineNo">331</span>      private MapFile.Reader mapFileReader;<a name="line.331"></a>
-<span class="sourceLineNo">332</span><a name="line.332"></a>
-<span class="sourceLineNo">333</span>      private boolean cachedNext;<a name="line.333"></a>
-<span class="sourceLineNo">334</span>      private ImmutableBytesWritable key;<a name="line.334"></a>
-<span class="sourceLineNo">335</span>      private ImmutableBytesWritable hash;<a name="line.335"></a>
-<span class="sourceLineNo">336</span><a name="line.336"></a>
-<span class="sourceLineNo">337</span>      Reader(Configuration conf, ImmutableBytesWritable startKey) throws IOException {<a name="line.337"></a>
-<span class="sourceLineNo">338</span>        this.conf = conf;<a name="line.338"></a>
-<span class="sourceLineNo">339</span>        int partitionIndex = Collections.binarySearch(partitions, startKey);<a name="line.339"></a>
-<span class="sourceLineNo">340</span>        if (partitionIndex &gt;= 0) {<a name="line.340"></a>
-<span class="sourceLineNo">341</span>          // if the key is equal to a partition, then go the file after that partition<a name="line.341"></a>
-<span class="sourceLineNo">342</span>          hashFileIndex = partitionIndex+1;<a name="line.342"></a>
-<span class="sourceLineNo">343</span>        } else {<a name="line.343"></a>
-<span class="sourceLineNo">344</span>          // if the key is between partitions, then go to the file between those partitions<a name="line.344"></a>
-<span class="sourceLineNo">345</span>          hashFileIndex = -1-partitionIndex;<a name="line.345"></a>
-<span class="sourceLineNo">346</span>        }<a name="line.346"></a>
-<span class="sourceLineNo">347</span>        openHashFile();<a name="line.347"></a>
-<span class="sourceLineNo">348</span><a name="line.348"></a>
-<span class="sourceLineNo">349</span>        // MapFile's don't make it easy to seek() so that the subsequent next() returns<a name="line.349"></a>
-<span class="sourceLineNo">350</span>        // the desired key/value pair.  So we cache it for the first call of next().<a name="line.350"></a>
-<span class="sourceLineNo">351</span>        hash = new ImmutableBytesWritable();<a name="line.351"></a>
-<span class="sourceLineNo">352</span>        key = (ImmutableBytesWritable) mapFileReader.getClosest(startKey, hash);<a name="line.352"></a>
-<span class="sourceLineNo">353</span>        if (key == null) {<a name="line.353"></a>
-<span class="sourceLineNo">354</span>          cachedNext = false;<a name="line.354"></a>
-<span class="sourceLineNo">355</span>          hash = null;<a name="line.355"></a>
-<span class="sourceLineNo">356</span>        } else {<a name="line.356"></a>
-<span class="sourceLineNo">357</span>          cachedNext = true;<a name="line.357"></a>
-<span class="sourceLineNo">358</span>        }<a name="line.358"></a>
-<span class="sourceLineNo">359</span>      }<a name="line.359"></a>
-<span class="sourceLineNo">360</span><a name="line.360"></a>
-<span class="sourceLineNo">361</span>      /**<a name="line.361"></a>
-<span class="sourceLineNo">362</span>       * Read the next key/hash pair.<a name="line.362"></a>
-<span class="sourceLineNo">363</span>       * Returns true if such a pair exists and false when at the end of the data.<a name="line.363"></a>
-<span class="sourceLineNo">364</span>       */<a name="line.364"></a>
-<span class="sourceLineNo">365</span>      public boolean next() throws IOException {<a name="line.365"></a>
-<span class="sourceLineNo">366</span>        if (cachedNext) {<a name="line.366"></a>
-<span class="sourceLineNo">367</span>          cachedNext = false;<a name="line.367"></a>
-<span class="sourceLineNo">368</span>          return true;<a name="line.368"></a>
-<span class="sourceLineNo">369</span>        }<a name="line.369"></a>
-<span class="sourceLineNo">370</span>        key = new ImmutableBytesWritable();<a name="line.370"></a>
-<span class="sourceLineNo">371</span>        hash = new ImmutableBytesWritable();<a name="line.371"></a>
-<span class="sourceLineNo">372</span>        while (true) {<a name="line.372"></a>
-<span class="sourceLineNo">373</span>          boolean hasNext = mapFileReader.next(key, hash);<a name="line.373"></a>
-<span class="sourceLineNo">374</span>          if (hasNext) {<a name="line.374"></a>
-<span class="sourceLineNo">375</span>            return true;<a name="line.375"></a>
-<span class="sourceLineNo">376</span>          }<a name="line.376"></a>
-<span class="sourceLineNo">377</span>          hashFileIndex++;<a name="line.377"></a>
-<span class="sourceLineNo">378</span>          if (hashFileIndex &lt; TableHash.this.numHashFiles) {<a name="line.378"></a>
-<span class="sourceLineNo">379</span>            mapFileReader.close();<a name="line.379"></a>
-<span class="sourceLineNo">380</span>            openHashFile();<a name="line.380"></a>
-<span class="sourceLineNo">381</span>          } else {<a name="line.381"></a>
-<span class="sourceLineNo">382</span>            key = null;<a name="line.382"></a>
-<span class="sourceLineNo">383</span>            hash = null;<a name="line.383"></a>
-<span class="sourceLineNo">384</span>            return false;<a name="line.384"></a>
-<span class="sourceLineNo">385</span>          }<a name="line.385"></a>
-<span class="sourceLineNo">386</span>        }<a name="line.386"></a>
-<span class="sourceLineNo">387</span>      }<a name="line.387"></a>
-<span class="sourceLineNo">388</span><a name="line.388"></a>
-<span class="sourceLineNo">389</span>      /**<a name="line.389"></a>
-<span class="sourceLineNo">390</span>       * Get the current key<a name="line.390"></a>
-<span class="sourceLineNo">391</span>       * @return the current key or null if there is no current key<a name="line.391"></a>
-<span class="sourceLineNo">392</span>       */<a name="line.392"></a>
-<span class="sourceLineNo">393</span>      public ImmutableBytesWritable getCurrentKey() {<a name="line.393"></a>
-<span class="sourceLineNo">394</span>        return key;<a name="line.394"></a>
-<span class="sourceLineNo">395</span>      }<a name="line.395"></a>
-<span class="sourceLineNo">396</span><a name="line.396"></a>
-<span class="sourceLineNo">397</span>      /**<a name="line.397"></a>
-<span class="sourceLineNo">398</span>       * Get the current hash<a name="line.398"></a>
-<span class="sourceLineNo">399</span>       * @return the current hash or null if there is no current hash<a name="line.399"></a>
-<span class="sourceLineNo">400</span>       */<a name="line.400"></a>
-<span class="sourceLineNo">401</span>      public ImmutableBytesWritable getCurrentHash() {<a name="line.401"></a>
-<span class="sourceLineNo">402</span>        return hash;<a name="line.402"></a>
-<span class="sourceLineNo">403</span>      }<a name="line.403"></a>
-<span class="sourceLineNo">404</span><a name="line.404"></a>
-<span class="sourceLineNo">405</span>      private void openHashFile() throws IOException {<a name="line.405"></a>
-<span class="sourceLineNo">406</span>        if (mapFileReader != null) {<a name="line.406"></a>
-<span class="sourceLineNo">407</span>          mapFileReader.close();<a name="line.407"></a>
-<span class="sourceLineNo">408</span>        }<a name="line.408"></a>
-<span class="sourceLineNo">409</span>        Path dataDir = new Path(TableHash.this.hashDir, HASH_DATA_DIR);<a name="line.409"></a>
-<span class="sourceLineNo">410</span>        Path dataFile = new Path(dataDir, getDataFileName(hashFileIndex));<a name="line.410"></a>
-<span class="sourceLineNo">411</span>        mapFileReader = new MapFile.Reader(dataFile, conf);<a name="line.411"></a>
-<span class="sourceLineNo">412</span>      }<a name="line.412"></a>
-<span class="sourceLineNo">413</span><a name="line.413"></a>
-<span class="sourceLineNo">414</span>      @Override<a name="line.414"></a>
-<span class="sourceLineNo">415</span>      public void close() throws IOException {<a name="line.415"></a>
-<span class="sourceLineNo">416</span>        mapFileReader.close();<a name="line.416"></a>
-<span class="sourceLineNo">417</span>      }<a name="line.417"></a>
-<span class="sourceLineNo">418</span>    }<a name="line.418"></a>
-<span class="sourceLineNo">419</span>  }<a name="line.419"></a>
-<span class="sourceLineNo">420</span><a name="line.420"></a>
-<span class="sourceLineNo">421</span>  static boolean isTableStartRow(byte[] row) {<a name="line.421"></a>
-<span class="sourceLineNo">422</span>    return Bytes.equals(HConstants.EMPTY_START_ROW, row);<a name="line.422"></a>
-<span class="sourceLineNo">423</span>  }<a name="line.423"></a>
-<span class="sourceLineNo">424</span><a name="line.424"></a>
-<span class="sourceLineNo">425</span>  static boolean isTableEndRow(byte[] row) {<a name="line.425"></a>
-<span class="sourceLineNo">426</span>    return Bytes.equals(HConstants.EMPTY_END_ROW, row);<a name="line.426"></a>
-<span class="sourceLineNo">427</span>  }<a name="line.427"></a>
-<span class="sourceLineNo">428</span><a name="line.428"></a>
-<span class="sourceLineNo">429</span>  public Job createSubmittableJob(String[] args) throws IOException {<a name="line.429"></a>
-<span class="sourceLineNo">430</span>    Path partitionsPath = new Path(destPath, PARTITIONS_FILE_NAME);<a name="line.430"></a>
-<span class="sourceLineNo">431</span>    generatePartitions(partitionsPath);<a name="line.431"></a>
-<span class="sourceLineNo">432</span><a name="line.432"></a>
-<span class="sourceLineNo">433</span>    Job job = Job.getInstance(getConf(),<a name="line.433"></a>
-<span class="sourceLineNo">434</span>          getConf().get("mapreduce.job.name", "hashTable_" + tableHash.tableName));<a name="line.434"></a>
-<span class="sourceLineNo">435</span>    Configuration jobConf = job.getConfiguration();<a name="line.435"></a>
-<span class="sourceLineNo">436</span>    jobConf.setLong(HASH_BATCH_SIZE_CONF_KEY, tableHash.batchSize);<a name="line.436"></a>
-<span class="sourceLineNo">437</span>    job.setJarByClass(HashTable.class);<a name="line.437"></a>
-<span class="sourceLineNo">438</span><a name="line.438"></a>
-<span class="sourceLineNo">439</span>    TableMapReduceUtil.initTableMapperJob(tableHash.tableName, tableHash.initScan(),<a name="line.439"></a>
-<span class="sourceLineNo">440</span>        HashMapper.class, ImmutableBytesWritable.class, ImmutableBytesWritable.class, job);<a name="line.440"></a>
+<span class="sourceLineNo">102</span>    List&lt;ImmutableBytesWritable&gt; partitions;<a name="line.102"></a>
+<span class="sourceLineNo">103</span><a name="line.103"></a>
+<span class="sourceLineNo">104</span>    public static TableHash read(Configuration conf, Path hashDir) throws IOException {<a name="line.104"></a>
+<span class="sourceLineNo">105</span>      TableHash tableHash = new TableHash();<a name="line.105"></a>
+<span class="sourceLineNo">106</span>      FileSystem fs = hashDir.getFileSystem(conf);<a name="line.106"></a>
+<span class="sourceLineNo">107</span>      tableHash.hashDir = hashDir;<a name="line.107"></a>
+<span class="sourceLineNo">108</span>      tableHash.readPropertiesFile(fs, new Path(hashDir, MANIFEST_FILE_NAME));<a name="line.108"></a>
+<span class="sourceLineNo">109</span>      tableHash.readPartitionFile(fs, conf, new Path(hashDir, PARTITIONS_FILE_NAME));<a name="line.109"></a>
+<span class="sourceLineNo">110</span>      return tableHash;<a name="line.110"></a>
+<span class="sourceLineNo">111</span>    }<a name="line.111"></a>
+<span class="sourceLineNo">112</span><a name="line.112"></a>
+<span class="sourceLineNo">113</span>    void writePropertiesFile(FileSystem fs, Path path) throws IOException {<a name="line.113"></a>
+<span class="sourceLineNo">114</span>      Properties p = new Properties();<a name="line.114"></a>
+<span class="sourceLineNo">115</span>      p.setProperty("table", tableName);<a name="line.115"></a>
+<span class="sourceLineNo">116</span>      if (families != null) {<a name="line.116"></a>
+<span class="sourceLineNo">117</span>        p.setProperty("columnFamilies", families);<a name="line.117"></a>
+<span class="sourceLineNo">118</span>      }<a name="line.118"></a>
+<span class="sourceLineNo">119</span>      p.setProperty("targetBatchSize", Long.toString(batchSize));<a name="line.119"></a>
+<span class="sourceLineNo">120</span>      p.setProperty("numHashFiles", Integer.toString(numHashFiles));<a name="line.120"></a>
+<span class="sourceLineNo">121</span>      if (!isTableStartRow(startRow)) {<a name="line.121"></a>
+<span class="sourceLineNo">122</span>        p.setProperty("startRowHex", Bytes.toHex(startRow));<a name="line.122"></a>
+<span class="sourceLineNo">123</span>      }<a name="line.123"></a>
+<span class="sourceLineNo">124</span>      if (!isTableEndRow(stopRow)) {<a name="line.124"></a>
+<span class="sourceLineNo">125</span>        p.setProperty("stopRowHex", Bytes.toHex(stopRow));<a name="line.125"></a>
+<span class="sourceLineNo">126</span>      }<a name="line.126"></a>
+<span class="sourceLineNo">127</span>      if (scanBatch &gt; 0) {<a name="line.127"></a>
+<span class="sourceLineNo">128</span>        p.setProperty("scanBatch", Integer.toString(scanBatch));<a name="line.128"></a>
+<span class="sourceLineNo">129</span>      }<a name="line.129"></a>
+<span class="sourceLineNo">130</span>      if (versions &gt;= 0) {<a name="line.130"></a>
+<span class="sourceLineNo">131</span>        p.setProperty("versions", Integer.toString(versions));<a name="line.131"></a>
+<span class="sourceLineNo">132</span>      }<a name="line.132"></a>
+<span class="sourceLineNo">133</span>      if (startTime != 0) {<a name="line.133"></a>
+<span class="sourceLineNo">134</span>        p.setProperty("startTimestamp", Long.toString(startTime));<a name="line.134"></a>
+<span class="sourceLineNo">135</span>      }<a name="line.135"></a>
+<span class="sourceLineNo">136</span>      if (endTime != 0) {<a name="line.136"></a>
+<span class="sourceLineNo">137</span>        p.setProperty("endTimestamp", Long.toString(endTime));<a name="line.137"></a>
+<span class="sourceLineNo">138</span>      }<a name="line.138"></a>
+<span class="sourceLineNo">139</span><a name="line.139"></a>
+<span class="sourceLineNo">140</span>      try (OutputStreamWriter osw = new OutputStreamWriter(fs.create(path), Charsets.UTF_8)) {<a name="line.140"></a>
+<span class="sourceLineNo">141</span>        p.store(osw, null);<a name="line.141"></a>
+<span class="sourceLineNo">142</span>      }<a name="line.142"></a>
+<span class="sourceLineNo">143</span>    }<a name="line.143"></a>
+<span class="sourceLineNo">144</span><a name="line.144"></a>
+<span class="sourceLineNo">145</span>    void readPropertiesFile(FileSystem fs, Path path) throws IOException {<a name="line.145"></a>
+<span class="sourceLineNo">146</span>      Properties p = new Properties();<a name="line.146"></a>
+<span class="sourceLineNo">147</span>      try (FSDataInputStream in = fs.open(path)) {<a name="line.147"></a>
+<span class="sourceLineNo">148</span>        try (InputStreamReader isr = new InputStreamReader(in, Charsets.UTF_8)) {<a name="line.148"></a>
+<span class="sourceLineNo">149</span>          p.load(isr);<a name="line.149"></a>
+<span class="sourceLineNo">150</span>        }<a name="line.150"></a>
+<span class="sourceLineNo">151</span>      }<a name="line.151"></a>
+<span class="sourceLineNo">152</span>      tableName = p.getProperty("table");<a name="line.152"></a>
+<span class="sourceLineNo">153</span>      families = p.getProperty("columnFamilies");<a name="line.153"></a>
+<span class="sourceLineNo">154</span>      batchSize = Long.parseLong(p.getProperty("targetBatchSize"));<a name="line.154"></a>
+<span class="sourceLineNo">155</span>      numHashFiles = Integer.parseInt(p.getProperty("numHashFiles"));<a name="line.155"></a>
+<span class="sourceLineNo">156</span><a name="line.156"></a>
+<span class="sourceLineNo">157</span>      String startRowHex = p.getProperty("startRowHex");<a name="line.157"></a>
+<span class="sourceLineNo">158</span>      if (startRowHex != null) {<a name="line.158"></a>
+<span class="sourceLineNo">159</span>        startRow = Bytes.fromHex(startRowHex);<a name="line.159"></a>
+<span class="sourceLineNo">160</span>      }<a name="line.160"></a>
+<span class="sourceLineNo">161</span>      String stopRowHex = p.getProperty("stopRowHex");<a name="line.161"></a>
+<span class="sourceLineNo">162</span>      if (stopRowHex != null) {<a name="line.162"></a>
+<span class="sourceLineNo">163</span>        stopRow = Bytes.fromHex(stopRowHex);<a name="line.163"></a>
+<span class="sourceLineNo">164</span>      }<a name="line.164"></a>
+<span class="sourceLineNo">165</span><a name="line.165"></a>
+<span class="sourceLineNo">166</span>      String scanBatchString = p.getProperty("scanBatch");<a name="line.166"></a>
+<span class="sourceLineNo">167</span>      if (scanBatchString != null) {<a name="line.167"></a>
+<span class="sourceLineNo">168</span>        scanBatch = Integer.parseInt(scanBatchString);<a name="line.168"></a>
+<span class="sourceLineNo">169</span>      }<a name="line.169"></a>
+<span class="sourceLineNo">170</span><a name="line.170"></a>
+<span class="sourceLineNo">171</span>      String versionString = p.getProperty("versions");<a name="line.171"></a>
+<span class="sourceLineNo">172</span>      if (versionString != null) {<a name="line.172"></a>
+<span class="sourceLineNo">173</span>        versions = Integer.parseInt(versionString);<a name="line.173"></a>
+<span class="sourceLineNo">174</span>      }<a name="line.174"></a>
+<span class="sourceLineNo">175</span><a name="line.175"></a>
+<span class="sourceLineNo">176</span>      String startTimeString = p.getProperty("startTimestamp");<a name="line.176"></a>
+<span class="sourceLineNo">177</span>      if (startTimeString != null) {<a name="line.177"></a>
+<span class="sourceLineNo">178</span>        startTime = Long.parseLong(startTimeString);<a name="line.178"></a>
+<span class="sourceLineNo">179</span>      }<a name="line.179"></a>
+<span class="sourceLineNo">180</span><a name="line.180"></a>
+<span class="sourceLineNo">181</span>      String endTimeString = p.getProperty("endTimestamp");<a name="line.181"></a>
+<span class="sourceLineNo">182</span>      if (endTimeString != null) {<a name="line.182"></a>
+<span class="sourceLineNo">183</span>        endTime = Long.parseLong(endTimeString);<a name="line.183"></a>
+<span class="sourceLineNo">184</span>      }<a name="line.184"></a>
+<span class="sourceLineNo">185</span>    }<a name="line.185"></a>
+<span class="sourceLineNo">186</span><a name="line.186"></a>
+<span class="sourceLineNo">187</span>    Scan initScan() throws IOException {<a name="line.187"></a>
+<span class="sourceLineNo">188</span>      Scan scan = new Scan();<a name="line.188"></a>
+<span class="sourceLineNo">189</span>      scan.setCacheBlocks(false);<a name="line.189"></a>
+<span class="sourceLineNo">190</span>      if (startTime != 0 || endTime != 0) {<a name="line.190"></a>
+<span class="sourceLineNo">191</span>        scan.setTimeRange(startTime, endTime == 0 ? HConstants.LATEST_TIMESTAMP : endTime);<a name="line.191"></a>
+<span class="sourceLineNo">192</span>      }<a name="line.192"></a>
+<span class="sourceLineNo">193</span>      if (scanBatch &gt; 0) {<a name="line.193"></a>
+<span class="sourceLineNo">194</span>        scan.setBatch(scanBatch);<a name="line.194"></a>
+<span class="sourceLineNo">195</span>      }<a name="line.195"></a>
+<span class="sourceLineNo">196</span>      if (versions &gt;= 0) {<a name="line.196"></a>
+<span class="sourceLineNo">197</span>        scan.readVersions(versions);<a name="line.197"></a>
+<span class="sourceLineNo">198</span>      }<a name="line.198"></a>
+<span class="sourceLineNo">199</span>      if (!isTableStartRow(startRow)) {<a name="line.199"></a>
+<span class="sourceLineNo">200</span>        scan.withStartRow(startRow);<a name="line.200"></a>
+<span class="sourceLineNo">201</span>      }<a name="line.201"></a>
+<span class="sourceLineNo">202</span>      if (!isTableEndRow(stopRow)) {<a name="line.202"></a>
+<span class="sourceLineNo">203</span>        scan.withStopRow(stopRow);<a name="line.203"></a>
+<span class="sourceLineNo">204</span>      }<a name="line.204"></a>
+<span class="sourceLineNo">205</span>      if(families != null) {<a name="line.205"></a>
+<span class="sourceLineNo">206</span>        for(String fam : families.split(",")) {<a name="line.206"></a>
+<span class="sourceLineNo">207</span>          scan.addFamily(Bytes.toBytes(fam));<a name="line.207"></a>
+<span class="sourceLineNo">208</span>        }<a name="line.208"></a>
+<span class="sourceLineNo">209</span>      }<a name="line.209"></a>
+<span class="sourceLineNo">210</span>      return scan;<a name="line.210"></a>
+<span class="sourceLineNo">211</span>    }<a name="line.211"></a>
+<span class="sourceLineNo">212</span><a name="line.212"></a>
+<span class="sourceLineNo">213</span>    /**<a name="line.213"></a>
+<span class="sourceLineNo">214</span>     * Choose partitions between row ranges to hash to a single output file<a name="line.214"></a>
+<span class="sourceLineNo">215</span>     * Selects region boundaries that fall within the scan range, and groups them<a name="line.215"></a>
+<span class="sourceLineNo">216</span>     * into the desired number of partitions.<a name="line.216"></a>
+<span class="sourceLineNo">217</span>     */<a name="line.217"></a>
+<span class="sourceLineNo">218</span>    void selectPartitions(Pair&lt;byte[][], byte[][]&gt; regionStartEndKeys) {<a name="line.218"></a>
+<span class="sourceLineNo">219</span>      List&lt;byte[]&gt; startKeys = new ArrayList&lt;&gt;();<a name="line.219"></a>
+<span class="sourceLineNo">220</span>      for (int i = 0; i &lt; regionStartEndKeys.getFirst().length; i++) {<a name="line.220"></a>
+<span class="sourceLineNo">221</span>        byte[] regionStartKey = regionStartEndKeys.getFirst()[i];<a name="line.221"></a>
+<span class="sourceLineNo">222</span>        byte[] regionEndKey = regionStartEndKeys.getSecond()[i];<a name="line.222"></a>
+<span class="sourceLineNo">223</span><a name="line.223"></a>
+<span class="sourceLineNo">224</span>        // if scan begins after this region, or starts before this region, then drop this region<a name="line.224"></a>
+<span class="sourceLineNo">225</span>        // in other words:<a name="line.225"></a>
+<span class="sourceLineNo">226</span>        //   IF (scan begins before the end of this region<a name="line.226"></a>
+<span class="sourceLineNo">227</span>        //      AND scan ends before the start of this region)<a name="line.227"></a>
+<span class="sourceLineNo">228</span>        //   THEN include this region<a name="line.228"></a>
+<span class="sourceLineNo">229</span>        if ((isTableStartRow(startRow) || isTableEndRow(regionEndKey)<a name="line.229"></a>
+<span class="sourceLineNo">230</span>            || Bytes.compareTo(startRow, regionEndKey) &lt; 0)<a name="line.230"></a>
+<span class="sourceLineNo">231</span>          &amp;&amp; (isTableEndRow(stopRow) || isTableStartRow(regionStartKey)<a name="line.231"></a>
+<span class="sourceLineNo">232</span>            || Bytes.compareTo(stopRow, regionStartKey) &gt; 0)) {<a name="line.232"></a>
+<span class="sourceLineNo">233</span>          startKeys.add(regionStartKey);<a name="line.233"></a>
+<span class="sourceLineNo">234</span>        }<a name="line.234"></a>
+<span class="sourceLineNo">235</span>      }<a name="line.235"></a>
+<span class="sourceLineNo">236</span><a name="line.236"></a>
+<span class="sourceLineNo">237</span>      int numRegions = startKeys.size();<a name="line.237"></a>
+<span class="sourceLineNo">238</span>      if (numHashFiles == 0) {<a name="line.238"></a>
+<span class="sourceLineNo">239</span>        numHashFiles = numRegions / 100;<a name="line.239"></a>
+<span class="sourceLineNo">240</span>      }<a name="line.240"></a>
+<span class="sourceLineNo">241</span>      if (numHashFiles == 0) {<a name="line.241"></a>
+<span class="sourceLineNo">242</span>        numHashFiles = 1;<a name="line.242"></a>
+<span class="sourceLineNo">243</span>      }<a name="line.243"></a>
+<span class="sourceLineNo">244</span>      if (numHashFiles &gt; numRegions) {<a name="line.244"></a>
+<span class="sourceLineNo">245</span>        // can't partition within regions<a name="line.245"></a>
+<span class="sourceLineNo">246</span>        numHashFiles = numRegions;<a name="line.246"></a>
+<span class="sourceLineNo">247</span>      }<a name="line.247"></a>
+<span class="sourceLineNo">248</span><a name="line.248"></a>
+<span class="sourceLineNo">249</span>      // choose a subset of start keys to group regions into ranges<a name="line.249"></a>
+<span class="sourceLineNo">250</span>      partitions = new ArrayList&lt;&gt;(numHashFiles - 1);<a name="line.250"></a>
+<span class="sourceLineNo">251</span>      // skip the first start key as it is not a partition between ranges.<a name="line.251"></a>
+<span class="sourceLineNo">252</span>      for (long i = 1; i &lt; numHashFiles; i++) {<a name="line.252"></a>
+<span class="sourceLineNo">253</span>        int splitIndex = (int) (numRegions * i / numHashFiles);<a name="line.253"></a>
+<span class="sourceLineNo">254</span>        partitions.add(new ImmutableBytesWritable(startKeys.get(splitIndex)));<a name="line.254"></a>
+<span class="sourceLineNo">255</span>      }<a name="line.255"></a>
+<span class="sourceLineNo">256</span>    }<a name="line.256"></a>
+<span class="sourceLineNo">257</span><a name="line.257"></a>
+<span class="sourceLineNo">258</span>    void writePartitionFile(Configuration conf, Path path) throws IOException {<a name="line.258"></a>
+<span class="sourceLineNo">259</span>      FileSystem fs = path.getFileSystem(conf);<a name="line.259"></a>
+<span class="sourceLineNo">260</span>      @SuppressWarnings("deprecation")<a name="line.260"></a>
+<span class="sourceLineNo">261</span>      SequenceFile.Writer writer = SequenceFile.createWriter(<a name="line.261"></a>
+<span class="sourceLineNo">262</span>        fs, conf, path, ImmutableBytesWritable.class, NullWritable.class);<a name="line.262"></a>
+<span class="sourceLineNo">263</span><a name="line.263"></a>
+<span class="sourceLineNo">264</span>      for (int i = 0; i &lt; partitions.size(); i++) {<a name="line.264"></a>
+<span class="sourceLineNo">265</span>        writer.append(partitions.get(i), NullWritable.get());<a name="line.265"></a>
+<span class="sourceLineNo">266</span>      }<a name="line.266"></a>
+<span class="sourceLineNo">267</span>      writer.close();<a name="line.267"></a>
+<span class="sourceLineNo">268</span>    }<a name="line.268"></a>
+<span class="sourceLineNo">269</span><a name="line.269"></a>
+<span class="sourceLineNo">270</span>    private void readPartitionFile(FileSystem fs, Configuration conf, Path path)<a name="line.270"></a>
+<span class="sourceLineNo">271</span>         throws IOException {<a name="line.271"></a>
+<span class="sourceLineNo">272</span>      @SuppressWarnings("deprecation")<a name="line.272"></a>
+<span class="sourceLineNo">273</span>      SequenceFile.Reader reader = new SequenceFile.Reader(fs, path, conf);<a name="line.273"></a>
+<span class="sourceLineNo">274</span>      ImmutableBytesWritable key = new ImmutableBytesWritable();<a name="line.274"></a>
+<span class="sourceLineNo">275</span>      partitions = new ArrayList&lt;&gt;();<a name="line.275"></a>
+<span class="sourceLineNo">276</span>      while (reader.next(key)) {<a name="line.276"></a>
+<span class="sourceLineNo">277</span>        partitions.add(new ImmutableBytesWritable(key.copyBytes()));<a name="line.277"></a>
+<span class="sourceLineNo">278</span>      }<a name="line.278"></a>
+<span class="sourceLineNo">279</span>      reader.close();<a name="line.279"></a>
+<span class="sourceLineNo">280</span><a name="line.280"></a>
+<span class="sourceLineNo">281</span>      if (!Ordering.natural().isOrdered(partitions)) {<a name="line.281"></a>
+<span class="sourceLineNo">282</span>        throw new IOException("Partitions are not ordered!");<a name="line.282"></a>
+<span class="sourceLineNo">283</span>      }<a name="line.283"></a>
+<span class="sourceLineNo">284</span>    }<a name="line.284"></a>
+<span class="sourceLineNo">285</span><a name="line.285"></a>
+<span class="sourceLineNo">286</span>    @Override<a name="line.286"></a>
+<span class="sourceLineNo">287</span>    public String toString() {<a name="line.287"></a>
+<span class="sourceLineNo">288</span>      StringBuilder sb = new StringBuilder();<a name="line.288"></a>
+<span class="sourceLineNo">289</span>      sb.append("tableName=").append(tableName);<a name="line.289"></a>
+<span class="sourceLineNo">290</span>      if (families != null) {<a name="line.290"></a>
+<span class="sourceLineNo">291</span>        sb.append(", families=").append(families);<a name="line.291"></a>
+<span class="sourceLineNo">292</span>      }<a name="line.292"></a>
+<span class="sourceLineNo">293</span>      sb.append(", batchSize=").append(batchSize);<a name="line.293"></a>
+<span class="sourceLineNo">294</span>      sb.append(", numHashFiles=").append(numHashFiles);<a name="line.294"></a>
+<span class="sourceLineNo">295</span>      if (!isTableStartRow(startRow)) {<a name="line.295"></a>
+<span class="sourceLineNo">296</span>        sb.append(", startRowHex=").append(Bytes.toHex(startRow));<a name="line.296"></a>
+<span class="sourceLineNo">297</span>      }<a name="line.297"></a>
+<span class="sourceLineNo">298</span>      if (!isTableEndRow(stopRow)) {<a name="line.298"></a>
+<span class="sourceLineNo">299</span>        sb.append(", stopRowHex=").append(Bytes.toHex(stopRow));<a name="line.299"></a>
+<span class="sourceLineNo">300</span>      }<a name="line.300"></a>
+<span class="sourceLineNo">301</span>      if (scanBatch &gt;= 0) {<a name="line.301"></a>
+<span class="sourceLineNo">302</span>        sb.append(", scanBatch=").append(scanBatch);<a name="line.302"></a>
+<span class="sourceLineNo">303</span>      }<a name="line.303"></a>
+<span class="sourceLineNo">304</span>      if (versions &gt;= 0) {<a name="line.304"></a>
+<span class="sourceLineNo">305</span>        sb.append(", versions=").append(versions);<a name="line.305"></a>
+<span class="sourceLineNo">306</span>      }<a name="line.306"></a>
+<span class="sourceLineNo">307</span>      if (startTime != 0) {<a name="line.307"></a>
+<span class="sourceLineNo">308</span>        sb.append("startTime=").append(startTime);<a name="line.308"></a>
+<span class="sourceLineNo">309</span>      }<a name="line.309"></a>
+<span class="sourceLineNo">310</span>      if (endTime != 0) {<a name="line.310"></a>
+<span class="sourceLineNo">311</span>        sb.append("endTime=").append(endTime);<a name="line.311"></a>
+<span class="sourceLineNo">312</span>      }<a name="line.312"></a>
+<span class="sourceLineNo">313</span>      return sb.toString();<a name="line.313"></a>
+<span class="sourceLineNo">314</span>    }<a name="line.314"></a>
+<span class="sourceLineNo">315</span><a name="line.315"></a>
+<span class="sourceLineNo">316</span>    static String getDataFileName(int hashFileIndex) {<a name="line.316"></a>
+<span class="sourceLineNo">317</span>      return String.format(HashTable.OUTPUT_DATA_FILE_PREFIX + "%05d", hashFileIndex);<a name="line.317"></a>
+<span class="sourceLineNo">318</span>    }<a name="line.318"></a>
+<span class="sourceLineNo">319</span><a name="line.319"></a>
+<span class="sourceLineNo">320</span>    /**<a name="line.320"></a>
+<span class="sourceLineNo">321</span>     * Open a TableHash.Reader starting at the first hash at or after the given key.<a name="line.321"></a>
+<span class="sourceLineNo">322</span>     * @throws IOException<a name="line.322"></a>
+<span class="sourceLineNo">323</span>     */<a name="line.323"></a>
+<span class="sourceLineNo">324</span>    public Reader newReader(Configuration conf, ImmutableBytesWritable startKey)<a name="line.324"></a>
+<span class="sourceLineNo">325</span>        throws IOException {<a name="line.325"></a>
+<span class="sourceLineNo">326</span>      return new Reader(conf, startKey);<a name="line.326"></a>
+<span class="sourceLineNo">327</span>    }<a name="line.327"></a>
+<span class="sourceLineNo">328</span><a name="line.328"></a>
+<span class="sourceLineNo">329</span>    public class Reader implements java.io.Closeable {<a name="line.329"></a>
+<span class="sourceLineNo">330</span>      private final Configuration conf;<a name="line.330"></a>
+<span class="sourceLineNo">331</span><a name="line.331"></a>
+<span class="sourceLineNo">332</span>      private int hashFileIndex;<a name="line.332"></a>
+<span class="sourceLineNo">333</span>      private MapFile.Reader mapFileReader;<a name="line.333"></a>
+<span class="sourceLineNo">334</span><a name="line.334"></a>
+<span class="sourceLineNo">335</span>      private boolean cachedNext;<a name="line.335"></a>
+<span class="sourceLineNo">336</span>      private ImmutableBytesWritable key;<a name="line.336"></a>
+<span class="sourceLineNo">337</span>      private ImmutableBytesWritable hash;<a name="line.337"></a>
+<span class="sourceLineNo">338</span><a name="line.338"></a>
+<span class="sourceLineNo">339</span>      Reader(Configuration conf, ImmutableBytesWritable startKey) throws IOException {<a name="line.339"></a>
+<span class="sourceLineNo">340</span>        this.conf = conf;<a name="line.340"></a>
+<span class="sourceLineNo">341</span>        int partitionIndex = Collections.binarySearch(partitions, startKey);<a name="line.341"></a>
+<span class="sourceLineNo">342</span>        if (partitionIndex &gt;= 0) {<a name="line.342"></a>
+<span class="sourceLineNo">343</span>          // if the key is equal to a partition, then go the file after that partition<a name="line.343"></a>
+<span class="sourceLineNo">344</span>          hashFileIndex = partitionIndex+1;<a name="line.344"></a>
+<span class="sourceLineNo">345</span>        } else {<a name="line.345"></a>
+<span class="sourceLineNo">346</span>          // if the key is between partitions, then go to the file between those partitions<a name="line.346"></a>
+<span class="sourceLineNo">347</span>          hashFileIndex = -1-partitionIndex;<a name="line.347"></a>
+<span class="sourceLineNo">348</span>        }<a name="line.348"></a>
+<span class="sourceLineNo">349</span>        openHashFile();<a name="line.349"></a>
+<span class="sourceLineNo">350</span><a name="line.350"></a>
+<span class="sourceLineNo">351</span>        // MapFile's don't make it easy to seek() so that the subsequent next() returns<a name="line.351"></a>
+<span class="sourceLineNo">352</span>        // the desired key/value pair.  So we cache it for the first call of next().<a name="line.352"></a>
+<span class="sourceLineNo">353</span>        hash = new ImmutableBytesWritable();<a name="line.353"></a>
+<span class="sourceLineNo">354</span>        key = (ImmutableBytesWritable) mapFileReader.getClosest(startKey, hash);<a name="line.354"></a>
+<span class="sourceLineNo">355</span>        if (key == null) {<a name="line.355"></a>
+<span class="sourceLineNo">356</span>          cachedNext = false;<a name="line.356"></a>
+<span class="sourceLineNo">357</span>          hash = null;<a name="line.357"></a>
+<span class="sourceLineNo">358</span>        } else {<a name="line.358"></a>
+<span class="sourceLineNo">359</span>          cachedNext = true;<a name="line.359"></a>
+<span class="sourceLineNo">360</span>        }<a name="line.360"></a>
+<span class="sourceLineNo">361</span>      }<a name="line.361"></a>
+<span class="sourceLineNo">362</span><a name="line.362"></a>
+<span class="sourceLineNo">363</span>      /**<a name="line.363"></a>
+<span class="sourceLineNo">364</span>       * Read the next key/hash pair.<a name="line.364"></a>
+<span class="sourceLineNo">365</span>       * Returns true if such a pair exists and false when at the end of the data.<a name="line.365"></a>
+<span class="sourceLineNo">366</span>       */<a name="line.366"></a>
+<span class="sourceLineNo">367</span>      public boolean next() throws IOException {<a name="line.367"></a>
+<span class="sourceLineNo">368</span>        if (cachedNext) {<a name="line.368"></a>
+<span class="sourceLineNo">369</span>          cachedNext = false;<a name="line.369"></a>
+<span class="sourceLineNo">370</span>          return true;<a name="line.370"></a>
+<span class="sourceLineNo">371</span>        }<a name="line.371"></a>
+<span class="sourceLineNo">372</span>        key = new ImmutableBytesWritable();<a name="line.372"></a>
+<span class="sourceLineNo">373</span>        hash = new ImmutableBytesWritable();<a name="line.373"></a>
+<span class="sourceLineNo">374</span>        while (true) {<a name="line.374"></a>
+<span class="sourceLineNo">375</span>          boolean hasNext = mapFileReader.next(key, hash);<a name="line.375"></a>
+<span class="sourceLineNo">376</span>          if (hasNext) {<a name="line.376"></a>
+<span class="sourceLineNo">377</span>            return true;<a name="line.377"></a>
+<span class="sourceLineNo">378</span>          }<a name="line.378"></a>
+<span class="sourceLineNo">379</span>          hashFileIndex++;<a name="line.379"></a>
+<span class="sourceLineNo">380</span>          if (hashFileIndex &lt; TableHash.this.numHashFiles) {<a name="line.380"></a>
+<span class="sourceLineNo">381</span>            mapFileReader.close();<a name="line.381"></a>
+<span class="sourceLineNo">382</span>            openHashFile();<a name="line.382"></a>
+<span class="sourceLineNo">383</span>          } else {<a name="line.383"></a>
+<span class="sourceLineNo">384</span>            key = null;<a name="line.384"></a>
+<span class="sourceLineNo">385</span>            hash = null;<a name="line.385"></a>
+<span class="sourceLineNo">386</span>            return false;<a name="line.386"></a>
+<span class="sourceLineNo">387</span>          }<a name="line.387"></a>
+<span class="sourceLineNo">388</span>        }<a name="line.388"></a>
+<span class="sourceLineNo">389</span>      }<a name="line.389"></a>
+<span class="sourceLineNo">390</span><a name="line.390"></a>
+<span class="sourceLineNo">391</span>      /**<a name="line.391"></a>
+<span class="sourceLineNo">392</span>       * Get the current key<a name="line.392"></a>
+<span class="sourceLineNo">393</span>       * @return the current key or null if there is no current key<a name="line.393"></a>
+<span class="sourceLineNo">394</span>       */<a name="line.394"></a>
+<span class="sourceLineNo">395</span>      public ImmutableBytesWritable getCurrentKey() {<a name="line.395"></a>
+<span class="sourceLineNo">396</span>        return key;<a name="line.396"></a>
+<span class="sourceLineNo">397</span>      }<a name="line.397"></a>
+<span class="sourceLineNo">398</span><a name="line.398"></a>
+<span class="sourceLineNo">399</span>      /**<a name="line.399"></a>
+<span class="sourceLineNo">400</span>       * Get the current hash<a name="line.400"></a>
+<span class="sourceLineNo">401</span>       * @return the current hash or null if there is no current hash<a name="line.401"></a>
+<span class="sourceLineNo">402</span>       */<a name="line.402"></a>
+<span class="sourceLineNo">403</span>      public ImmutableBytesWritable getCurrentHash() {<a name="line.403"></a>
+<span class="sourceLineNo">404</span>        return hash;<a name="line.404"></a>
+<span class="sourceLineNo">405</span>      }<a name="line.405"></a>
+<span class="sourceLineNo">406</span><a name="line.406"></a>
+<span class="sourceLineNo">407</span>      private void openHashFile() throws IOException {<a name="line.407"></a>
+<span class="sourceLineNo">408</span>        if (mapFileReader != null) {<a name="line.408"></a>
+<span class="sourceLineNo">409</span>          mapFileReader.close();<a name="line.409"></a>
+<span class="sourceLineNo">410</span>        }<a name="line.410"></a>
+<span class="sourceLineNo">411</span>        Path dataDir = new Path(TableHash.this.hashDir, HASH_DATA_DIR);<a name="line.411"></a>
+<span class="sourceLineNo">412</span>        Path dataFile = new Path(dataDir, getDataFileName(hashFileIndex));<a name="line.412"></a>
+<span class="sourceLineNo">413</span>        mapFileReader = new MapFile.Reader(dataFile, conf);<a name="line.413"></a>
+<span class="sourceLineNo">414</span>      }<a name="line.414"></a>
+<span class="sourceLineNo">415</span><a name="line.415"></a>
+<span class="sourceLineNo">416</span>      @Override<a name="line.416"></a>
+<span class="sourceLineNo">417</span>      public void close() throws IOException {<a name="line.417"></a>
+<span class="sourceLineNo">418</span>        mapFileReader.close();<a name="line.418"></a>
+<span class="sourceLineNo">419</span>      }<a name="line.419"></a>
+<span class="sourceLineNo">420</span>    }<a name="line.420"></a>
+<span class="sourceLineNo">421</span>  }<a name="line.421"></a>
+<span class="sourceLineNo">422</span><a name="line.422"></a>
+<span class="sourceLineNo">423</span>  static boolean isTableStartRow(byte[] row) {<a name="line.423"></a>
+<span class="sourceLineNo">424</span>    return Bytes.equals(HConstants.EMPTY_START_ROW, row);<a name="line.424"></a>
+<span class="sourceLineNo">425</span>  }<a name="line.425"></a>
+<span class="sourceLineNo">426</span><a name="line.426"></a>
+<span class="sourceLineNo">427</span>  static boolean isTableEndRow(byte[] row) {<a name="line.427"></a>
+<span class="sourceLineNo">428</span>    return Bytes.equals(HConstants.EMPTY_END_ROW, row);<a name="line.428"></a>
+<span class="sourceLineNo">429</span>  }<a name="line.429"></a>
+<span class="sourceLineNo">430</span><a name="line.430"></a>
+<span class="sourceLineNo">431</span>  public Job createSubmittableJob(String[] args) throws IOException {<a name="line.431"></a>
+<span class="sourceLineNo">432</span>    Path partitionsPath = new Path(destPath, PARTITIONS_FILE_NAME);<a name="line.432"></a>
+<span class="sourceLineNo">433</span>    generatePartitions(partitionsPath);<a name="line.433"></a>
+<span class="sourceLineNo">434</span><a name="line.434"></a>
+<span class="sourceLineNo">435</span>    Job job = Job.getInstance(getConf(),<a name="line.435"></a>
+<span class="sourceLineNo">436</span>          getConf().get("mapreduce.job.name", "hashTable_" + tableHash.tableName));<a name="line.436"></a>
+<span class="sourceLineNo">437</span>    Configuration jobConf = job.getConfiguration();<a name="line.437"></a>
+<span class="sourceLineNo">438</span>    jobConf.setLong(HASH_BATCH_SIZE_CONF_KEY, tableHash.batchSize);<a name="line.438"></a>
+<span class="sourceLineNo">439</span>    jobConf.setBoolean(IGNORE_TIMESTAMPS, tableHash.ignoreTimestamps);<a name="line.439"></a>
+<span class="sourceLineNo">440</span>    job.setJarByClass(HashTable.class);<a name="line.440"></a>
 <span class="sourceLineNo">441</span><a name="line.441"></a>
-<span class="sourceLineNo">442</span>    // use a TotalOrderPartitioner and reducers to group region output into hash files<a name="line.442"></a>
-<span class="sourceLineNo">443</span>    job.setPartitionerClass(TotalOrderPartitioner.class);<a name="line.443"></a>
-<span class="sourceLineNo">444</span>    TotalOrderPartitioner.setPartitionFile(jobConf, partitionsPath);<a name="line.444"></a>
-<span class="sourceLineNo">445</span>    job.setReducerClass(Reducer.class);  // identity reducer<a name="line.445"></a>
-<span class="sourceLineNo">446</span>    job.setNumReduceTasks(tableHash.numHashFiles);<a name="line.446"></a>
-<span class="sourceLineNo">447</span>    job.setOutputKeyClass(ImmutableBytesWritable.class);<a name="line.447"></a>
-<span class="sourceLineNo">448</span>    job.setOutputValueClass(ImmutableBytesWritable.class);<a name="line.448"></a>
-<span class="sourceLineNo">449</span>    job.setOutputFormatClass(MapFileOutputFormat.class);<a name="line.449"></a>
-<span class="sourceLineNo">450</span>    FileOutputFormat.setOutputPath(job, new Path(destPath, HASH_DATA_DIR));<a name="line.450"></a>
-<span class="sourceLineNo">451</span><a name="line.451"></a>
-<span class="sourceLineNo">452</span>    return job;<a name="line.452"></a>
-<span class="sourceLineNo">453</span>  }<a name="line.453"></a>
+<span class="sourceLineNo">442</span>    TableMapReduceUtil.initTableMapperJob(tableHash.tableName, tableHash.initScan(),<a name="line.442"></a>
+<span class="sourceLineNo">443</span>        HashMapper.class, ImmutableBytesWritable.class, ImmutableBytesWritable.class, job);<a name="line.443"></a>
+<span class="sourceLineNo">444</span><a name="line.444"></a>
+<span class="sourceLineNo">445</span>    // use a TotalOrderPartitioner and reducers to group region output into hash files<a name="line.445"></a>
+<span class="sourceLineNo">446</span>    job.setPartitionerClass(TotalOrderPartitioner.class);<a name="line.446"></a>
+<span class="sourceLineNo">447</span>    TotalOrderPartitioner.setPartitionFile(jobConf, partitionsPath);<a name="line.447"></a>
+<span class="sourceLineNo">448</span>    job.setReducerClass(Reducer.class);  // identity reducer<a name="line.448"></a>
+<span class="sourceLineNo">449</span>    job.setNumReduceTasks(tableHash.numHashFiles);<a name="line.449"></a>
+<span class="sourceLineNo">450</span>    job.setOutputKeyClass(ImmutableBytesWritable.class);<a name="line.450"></a>
+<span class="sourceLineNo">451</span>    job.setOutputValueClass(ImmutableBytesWritable.class);<a name="line.451"></a>
+<span class="sourceLineNo">452</span>    job.setOutputFormatClass(MapFileOutputFormat.class);<a name="line.452"></a>
+<span class="sourceLineNo">453</span>    FileOutputFormat.setOutputPath(job, new Path(destPath, HASH_DATA_DIR));<a name="line.453"></a>
 <span class="sourceLineNo">454</span><a name="line.454"></a>
-<span class="sourceLineNo">455</span>  private void generatePartitions(Path partitionsPath) throws IOException {<a name="line.455"></a>
-<span class="sourceLineNo">456</span>    Connection connection = ConnectionFactory.createConnection(getConf());<a name="line.456"></a>
-<span class="sourceLineNo">457</span>    Pair&lt;byte[][], byte[][]&gt; regionKeys<a name="line.457"></a>
-<span class="sourceLineNo">458</span>      = connection.getRegionLocator(TableName.valueOf(tableHash.tableName)).getStartEndKeys();<a name="line.458"></a>
-<span class="sourceLineNo">459</span>    connection.close();<a name="line.459"></a>
-<span class="sourceLineNo">460</span><a name="line.460"></a>
-<span class="sourceLineNo">461</span>    tableHash.selectPartitions(regionKeys);<a name="line.461"></a>
-<span class="sourceLineNo">462</span>    LOG.info("Writing " + tableHash.partitions.size() + " partition keys to " + partitionsPath);<a name="line.462"></a>
+<span class="sourceLineNo">455</span>    return job;<a name="line.455"></a>
+<span class="sourceLineNo">456</span>  }<a name="line.456"></a>
+<span class="sourceLineNo">457</span><a name="line.457"></a>
+<span class="sourceLineNo">458</span>  private void generatePartitions(Path partitionsPath) throws IOException {<a name="line.458"></a>
+<span class="sourceLineNo">459</span>    Connection connection = ConnectionFactory.createConnection(getConf());<a name="line.459"></a>
+<span class="sourceLineNo">460</span>    Pair&lt;byte[][], byte[][]&gt; regionKeys<a name="line.460"></a>
+<span class="sourceLineNo">461</span>      = connection.getRegionLocator(TableName.valueOf(tableHash.tableName)).getStartEndKeys();<a name="line.461"></a>
+<span class="sourceLineNo">462</span>    connection.close();<a name="line.462"></a>
 <span class="sourceLineNo">463</span><a name="line.463"></a>
-<span class="sourceLineNo">464</span>    tableHash.writePartitionFile(getConf(), partitionsPath);<a name="line.464"></a>
-<span class="sourceLineNo">465</span>  }<a name="line.465"></a>
+<span class="sourceLineNo">464</span>    tableHash.selectPartitions(regionKeys);<a name="line.464"></a>
+<span class="sourceLineNo">465</span>    LOG.info("Writing " + tableHash.partitions.size() + " partition keys to " + partitionsPath);<a name="line.465"></a>
 <span class="sourceLineNo">466</span><a name="line.466"></a>
-<span class="sourceLineNo">467</span>  static class ResultHasher {<a name="line.467"></a>
-<span class="sourceLineNo">468</span>    private MessageDigest digest;<a name="line.468"></a>
+<span class="sourceLineNo">467</span>    tableHash.writePartitionFile(getConf(), partitionsPath);<a name="line.467"></a>
+<span class="sourceLineNo">468</span>  }<a name="line.468"></a>
 <span class="sourceLineNo">469</span><a name="line.469"></a>
-<span class="sourceLineNo">470</span>    private boolean batchStarted = false;<a name="line.470"></a>
-<span class="sourceLineNo">471</span>    private ImmutableBytesWritable batchStartKey;<a name="line.471"></a>
-<span class="sourceLineNo">472</span>    private ImmutableBytesWritable batchHash;<a name="line.472"></a>
-<span class="sourceLineNo">473</span>    private long batchSize = 0;<a name="line.473"></a>
-<span class="sourceLineNo">474</span><a name="line.474"></a>
-<span class="sourceLineNo">475</span><a name="line.475"></a>
-<span class="sourceLineNo">476</span>    public ResultHasher() {<a name="line.476"></a>
-<span class="sourceLineNo">477</span>      try {<a name="line.477"></a>
-<span class="sourceLineNo">478</span>        digest = MessageDigest.getInstance("MD5");<a name="line.478"></a>
-<span class="sourceLineNo">479</span>      } catch (NoSuchAlgorithmException e) {<a name="line.479"></a>
-<span class="sourceLineNo">480</span>        Throwables.propagate(e);<a name="line.480"></a>
-<span class="sourceLineNo">481</span>      }<a name="line.481"></a>
-<span class="sourceLineNo">482</span>    }<a name="line.482"></a>
-<span class="sourceLineNo">483</span><a name="line.483"></a>
-<span class="sourceLineNo">484</span>    public void startBatch(ImmutableBytesWritable row) {<a name="line.484"></a>
-<span class="sourceLineNo">485</span>      if (batchStarted) {<a name="line.485"></a>
-<span class="sourceLineNo">486</span>        throw new RuntimeException("Cannot start new batch without finishing existing one.");<a name="line.486"></a>
-<span class="sourceLineNo">487</span>      }<a name="line.487"></a>
-<span class="sourceLineNo">488</span>      batchStarted = true;<a name="line.488"></a>
-<span class="sourceLineNo">489</span>      batchSize = 0;<a name="line.489"></a>
-<span class="sourceLineNo">490</span>      batchStartKey = row;<a name="line.490"></a>
-<span class="sourceLineNo">491</span>      batchHash = null;<a name="line.491"></a>
-<span class="sourceLineNo">492</span>    }<a name="line.492"></a>
-<span class="sourceLineNo">493</span><a name="line.493"></a>
-<span class="sourceLineNo">494</span>    public void hashResult(Result result) {<a name="line.494"></a>
-<span class="sourceLineNo">495</span>      if (!batchStarted) {<a name="line.495"></a>
-<span class="sourceLineNo">496</span>        throw new RuntimeException("Cannot add to batch that has not been started.");<a name="line.496"></a>
-<span class="sourceLineNo">497</span>      }<a name="line.497"></a>
-<span class="sourceLineNo">498</span>      for (Cell cell : result.rawCells()) {<a name="line.498"></a>
-<span class="sourceLineNo">499</span>        int rowLength = cell.getRowLength();<a name="line.499"></a>
-<span class="sourceLineNo">500</span>        int familyLength = cell.getFamilyLength();<a name="line.500"></a>
-<span class="sourceLineNo">501</span>        int qualifierLength = cell.getQualifierLength();<a name="line.501"></a>
-<span class="sourceLineNo">502</span>        int valueLength = cell.getValueLength();<a name="line.502"></a>
-<span class="sourceLineNo">503</span>        digest.update(cell.getRowArray(), cell.getRowOffset(), rowLength);<a name="line.503"></a>
-<span class="sourceLineNo">504</span>        digest.update(cell.getFamilyArray(), cell.getFamilyOffset(), familyLength);<a name="line.504"></a>
-<span class="sourceLineNo">505</span>        digest.update(cell.getQualifierArray(), cell.getQualifierOffset(), qualifierLength);<a name="line.505"></a>
-<span class="sourceLineNo">506</span>        long ts = cell.getTimestamp();<a name="line.506"></a>
-<span class="sourceLineNo">507</span>        for (int i = 8; i &gt; 0; i--) {<a name="line.507"></a>
-<span class="sourceLineNo">508</span>          digest.update((byte) ts);<a name="line.508"></a>
-<span class="sourceLineNo">509</span>          ts &gt;&gt;&gt;= 8;<a name="line.509"></a>
-<span class="sourceLineNo">510</span>        }<a name="line.510"></a>
-<span class="sourceLineNo">511</span>        digest.update(cell.getValueArray(), cell.getValueOffset(), valueLength);<a name="line.511"></a>
-<span class="sourceLineNo">512</span><a name="line.512"></a>
-<span class="sourceLineNo">513</span>        batchSize += rowLength + familyLength + qualifierLength + 8 + valueLength;<a name="line.513"></a>
-<span class="sourceLineNo">514</span>      }<a name="line.514"></a>
-<span class="sourceLineNo">515</span>    }<a name="line.515"></a>
-<span class="sourceLineNo">516</span><a name="line.516"></a>
-<span class="sourceLineNo">517</span>    public void finishBatch() {<a name="line.517"></a>
-<span class="sourceLineNo">518</span>      if (!batchStarted) {<a name="line.518"></a>
-<span class="sourceLineNo">519</span>        throw new RuntimeException("Cannot finish batch that has not started.");<a name="line.519"></a>
-<span class="sourceLineNo">520</span>      }<a name="line.520"></a>
-<span class="sourceLineNo">521</span>      batchStarted = false;<a name="line.521"></a>
-<span class="sourceLineNo">522</span>      batchHash = new ImmutableBytesWritable(digest.digest());<a name="line.522"></a>
-<span class="sourceLineNo">523</span>    }<a name="line.523"></a>
-<span class="sourceLineNo">524</span><a name="line.524"></a>
-<span class="sourceLineNo">525</span>    public boolean isBatchStarted() {<a name="line.525"></a>
-<span class="sourceLineNo">526</span>      return batchStarted;<a name="line.526"></a>
-<span class="sourceLineNo">527</span>    }<a name="line.527"></a>
-<span class="sourceLineNo">528</span><a name="line.528"></a>
-<span class="sourceLineNo">529</span>    public ImmutableBytesWritable getBatchStartKey() {<a name="line.529"></a>
-<span class="sourceLineNo">530</span>      return batchStartKey;<a name="line.530"></a>
-<span class="sourceLineNo">531</span>    }<a name="line.531"></a>
-<span class="sourceLineNo">532</span><a name="line.532"></a>
-<span class="sourceLineNo">533</span>    public ImmutableBytesWritable getBatchHash() {<a name="line.533"></a>
-<span class="sourceLineNo">534</span>      return batchHash;<a name="line.534"></a>
-<span class="sourceLineNo">535</span>    }<a name="line.535"></a>
-<span class="sourceLineNo">536</span><a name="line.536"></a>
-<span class="sourceLineNo">537</span>    public long getBatchSize() {<a name="line.537"></a>
-<span class="sourceLineNo">538</span>      return batchSize;<a name="line.538"></a>
-<span class="sourceLineNo">539</span>    }<a name="line.539"></a>
-<span class="sourceLineNo">540</span>  }<a name="line.540"></a>
-<span class="sourceLineNo">541</span><a name="line.541"></a>
-<span class="sourceLineNo">542</span>  public static class HashMapper<a name="line.542"></a>
-<span class="sourceLineNo">543</span>    extends TableMapper&lt;ImmutableBytesWritable, ImmutableBytesWritable&gt; {<a name="line.543"></a>
-<span class="sourceLineNo">544</span><a name="line.544"></a>
-<span class="sourceLineNo">545</span>    private ResultHasher hasher;<a name="line.545"></a>
-<span class="sourceLineNo">546</span>    private long targetBatchSize;<a name="line.546"></a>
-<span class="sourceLineNo">547</span><a name="line.547"></a>
-<span class="sourceLineNo">548</span>    private ImmutableBytesWritable currentRow;<a name="line.548"></a>
-<span class="sourceLineNo">549</span><a name="line.549"></a>
-<span class="sourceLineNo">550</span>    @Override<a name="line.550"></a>
-<span class="sourceLineNo">551</span>    protected void setup(Context context) throws IOException, InterruptedException {<a name="line.551"></a>
-<span class="sourceLineNo">552</span>      targetBatchSize = context.getConfiguration()<a name="line.552"></a>
-<span class="sourceLineNo">553</span>          .getLong(HASH_BATCH_SIZE_CONF_KEY, DEFAULT_BATCH_SIZE);<a name="line.553"></a>
-<span class="sourceLineNo">554</span>      hasher = new ResultHasher();<a name="line.554"></a>
-<span class="sourceLineNo">555</span><a name="line.555"></a>
-<span class="sourceLineNo">556</span>      TableSplit split = (TableSplit) context.getInputSplit();<a name="line.556"></a>
-<span class="sourceLineNo">557</span>      hasher.startBatch(new ImmutableBytesWritable(split.getStartRow()));<a name="line.557"></a>
-<span class="sourceLineNo">558</span>    }<a name="line.558"></a>
-<span class="sourceLineNo">559</span><a name="line.559"></a>
-<span class="sourceLineNo">560</span>    @Override<a name="line.560"></a>
-<span class="sourceLineNo">561</span>    protected void map(ImmutableBytesWritable key, Result value, Context context)<a name="line.561"></a>
-<span class="sourceLineNo">562</span>        throws IOException, InterruptedException {<a name="line.562"></a>
-<span class="sourceLineNo">563</span><a name="line.563"></a>
-<span class="sourceLineNo">564</span>      if (currentRow == null || !currentRow.equals(key)) {<a name="line.564"></a>
-<span class="sourceLineNo">565</span>        currentRow = new ImmutableBytesWritable(key); // not immutable<a name="line.565"></a>
-<span class="sourceLineNo">566</span><a name="line.566"></a>
-<span class="sourceLineNo">567</span>        if (hasher.getBatchSize() &gt;= targetBatchSize) {<a name="line.567"></a>
-<span class="sourceLineNo">568</span>          hasher.finishBatch();<a name="line.568"></a>
-<span class="sourceLineNo">569</span>          context.write(hasher.getBatchStartKey(), hasher.getBatchHash());<a name="line.569"></a>
-<span class="sourceLineNo">570</span>          hasher.startBatch(currentRow);<a name="line.570"></a>
-<span class="sourceLineNo">571</span>        }<a name="line.571"></a>
-<span class="sourceLineNo">572</span>      }<a name="line.572"></a>
-<span class="sourceLineNo">573</span><a name="line.573"></a>
-<span class="sourceLineNo">574</span>      hasher.hashResult(value);<a name="line.574"></a>
-<span class="sourceLineNo">575</span>    }<a name="line.575"></a>
-<span class="sourceLineNo">576</span><a name="line.576"></a>
-<span class="sourceLineNo">577</span>    @Override<a name="line.577"></a>
-<span class="sourceLineNo">578</span>    protected void cleanup(Context context) throws IOException, InterruptedException {<a name="line.578"></a>
-<span class="sourceLineNo">579</span>      hasher.finishBatch();<a name="line.579"></a>
-<span class="sourceLineNo">580</span>      context.write(hasher.getBatchStartKey(), hasher.getBatchHash());<a name="line.580"></a>
-<span class="sourceLineNo">581</span>    }<a name="line.581"></a>
-<span class="sourceLineNo">582</span>  }<a name="line.582"></a>
-<span class="sourceLineNo">583</span><a name="line.583"></a>
-<span class="sourceLineNo">584</span>  private void writeTempManifestFile() throws IOException {<a name="line.584"></a>
-<span class="sourceLineNo">585</span>    Path tempManifestPath = new Path(destPath, TMP_MANIFEST_FILE_NAME);<a name="line.585"></a>
-<span class="sourceLineNo">586</span>    FileSystem fs = tempManifestPath.getFileSystem(getConf());<a name="line.586"></a>
-<span class="sourceLineNo">587</span>    tableHash.writePropertiesFile(fs, tempManifestPath);<a name="line.587"></a>
-<span class="sourceLineNo">588</span>  }<a name="line.588"></a>
-<span class="sourceLineNo">589</span><a name="line.589"></a>
-<span class="sourceLineNo">590</span>  private void completeManifest() throws IOException {<a name="line.590"></a>
-<span class="sourceLineNo">591</span>    Path tempManifestPath = new Path(destPath, TMP_MANIFEST_FILE_NAME);<a name="line.591"></a>
-<span class="sourceLineNo">592</span>    Path manifestPath = new Path(destPath, MANIFEST_FILE_NAME);<a name="line.592"></a>
-<span class="sourceLineNo">593</span>    FileSystem fs = tempManifestPath.getFileSystem(getConf());<a name="line.593"></a>
-<span class="sourceLineNo">594</span>    fs.rename(tempManifestPath, manifestPath);<a name="line.594"></a>
-<span class="sourceLineNo">595</span>  }<a name="line.595"></a>
-<span class="sourceLineNo">596</span><a name="line.596"></a>
-<span class="sourceLineNo">597</span>  private static final int NUM_ARGS = 2;<a name="line.597"></a>
-<span class="sourceLineNo">598</span>  private static void printUsage(final String errorMsg) {<a name="line.598"></a>
-<span class="sourceLineNo">599</span>    if (errorMsg != null &amp;&amp; errorMsg.length() &gt; 0) {<a name="line.599"></a>
-<span class="sourceLineNo">600</span>      System.err.println("ERROR: " + errorMsg);<a name="line.600"></a>
-<span class="sourceLineNo">601</span>      System.err.println();<a name="line.601"></a>
-<span class="sourceLineNo">602</span>    }<a name="line.602"></a>
-<span class="sourceLineNo">603</span>    System.err.println("Usage: HashTable [options] &lt;tablename&gt; &lt;outputpath&gt;");<a name="line.603"></a>
-<span class="sourceLineNo">604</span>    System.err.println();<a name="line.604"></a>
-<span class="sourceLineNo">605</span>    System.err.println("Options:");<a name="line.605"></a>
-<span class="sourceLineNo">606</span>    System.err.println(" batchsize     the target amount of bytes to hash in each batch");<a name="line.606"></a>
-<span class="sourceLineNo">607</span>    System.err.println("               rows are added to the batch until this size is reached");<a name="line.607"></a>
-<span class="sourceLineNo">608</span>    System.err.println("               (defaults to " + DEFAULT_BATCH_SIZE + " bytes)");<a name="line.608"></a>
-<span class="sourceLineNo">609</span>    System.err.println(" numhashfiles  the number of hash files to create");<a name="line.609"></a>
-<span class="sourceLineNo">610</span>    System.err.println("               if set to fewer than number of regions then");<a name="line.610"></a>
-<span class="sourceLineNo">611</span>    System.err.println("               the job will create this number of reducers");<a name="line.611"></a>
-<span class="sourceLineNo">612</span>    System.err.println("               (defaults to 1/100 of regions -- at least 1)");<a name="line.612"></a>
-<span class="sourceLineNo">613</span>    System.err.println(" startrow      the start row");<a name="line.613"></a>
-<span class="sourceLineNo">614</span>    System.err.println(" stoprow       the stop row");<a name="line.614"></a>
-<span class="sourceLineNo">615</span>    System.err.println(" starttime     beginning of the time range (unixtime in millis)");<a name="line.615"></a>
-<span class="sourceLineNo">616</span>    System.err.println("               without endtime means from starttime to forever");<a name="line.616"></a>
-<span class="sourceLineNo">617</span>    System.err.println(" endtime       end of the time range.  Ignored if no starttime specified.");<a name="line.617"></a>
-<span class="sourceLineNo">618</span>    System.err.println(" scanbatch     scanner batch size to support intra row scans");<a name="line.618"></a>
-<span class="sourceLineNo">619</span>    System.err.println(" versions      number of cell versions to include");<a name="line.619"></a>
-<span class="sourceLineNo">620</span>    System.err.println(" families      comma-separated list of families to include");<a name="line.620"></a>
-<span class="sourceLineNo">621</span>    System.err.println();<a name="line.621"></a>
-<span class="sourceLineNo">622</span>    System.err.println("Args:");<a name="line.622"></a>
-<span class="sourceLineNo">623</span>    System.err.println(" tablename     Name of the table to hash");<a name="line.623"></a>
-<span class="sourceLineNo">624</span>    System.err.println(" outputpath    Filesystem path to put the output data");<a name="line.624"></a>
-<span class="sourceLineNo">625</span>    System.err.println();<a name="line.625"></a>
-<span class="sourceLineNo">626</span>    System.err.println("Examples:");<a name="line.626"></a>
-<span class="sourceLineNo">627</span>    System.err.println(" To hash 'TestTable' in 32kB batches for a 1 hour window into 50 files:");<a name="line.627"></a>
-<span class="sourceLineNo">628</span>    System.err.println(" $ hbase " +<a name="line.628"></a>
-<span class="sourceLineNo">629</span>        "org.apache.hadoop.hbase.mapreduce.HashTable --batchsize=32000 --numhashfiles=50"<a name="line.629"></a>
-<span class="sourceLineNo">630</span>        + " --starttime=1265875194289 --endtime=1265878794289 --families=cf2,cf3"<a name="line.630"></a>
-<span class="sourceLineNo">631</span>        + " TestTable /hashes/testTable");<a name="line.631"></a>
-<span class="sourceLineNo">632</span>  }<a name="line.632"></a>
-<span class="sourceLineNo">633</span><a name="line.633"></a>
-<span class="sourceLineNo">634</span>  private boolean doCommandLine(final String[] args) {<a name="line.634"></a>
-<span class="sourceLineNo">635</span>    if (args.length &lt; NUM_ARGS) {<a name="line.635"></a>
-<span class="sourceLineNo">636</span>      printUsage(null);<a name="line.636"></a>
-<span class="sourceLineNo">637</span>      return false;<a name="line.637"></a>
-<span class="sourceLineNo">638</span>    }<a name="line.638"></a>
-<span class="sourceLineNo">639</span>    try {<a name="line.639"></a>
-<span class="sourceLineNo">640</span><a name="line.640"></a>
-<span class="sourceLineNo">641</span>      tableHash.tableName = args[args.length-2];<a name="line.641"></a>
-<span class="sourceLineNo">642</span>      destPath = new Path(args[args.length-1]);<a name="line.642"></a>
-<span class="sourceLineNo">643</span><a name="line.643"></a>
-<span class="sourceLineNo">644</span>      for (int i = 0; i &lt; args.length - NUM_ARGS; i++) {<a name="line.644"></a>
-<span class="sourceLineNo">645</span>        String cmd = args[i];<a name="line.645"></a>
-<span class="sourceLineNo">646</span>        if (cmd.equals("-h") || cmd.startsWith("--h")) {<a name="line.646"></a>
-<span class="sourceLineNo">647</span>          printUsage(null);<a name="line.647"></a>
-<span class="sourceLineNo">648</span>          return false;<a name="line.648"></a>
-<span class="sourceLineNo">649</span>        }<a name="line.649"></a>
-<span class="sourceLineNo">650</span><a name="line.650"></a>
-<span class="sourceLineNo">651</span>        final String batchSizeArgKey = "--batchsize=";<a name="line.651"></a>
-<span class="sourceLineNo">652</span>        if (cmd.startsWith(batchSizeArgKey)) {<a name="line.652"></a>
-<span class="sourceLineNo">653</span>          tableHash.batchSize = Long.parseLong(cmd.substring(batchSizeArgKey.length()));<a name="line.653"></a>
-<span class="sourceLineNo">654</span>          continue;<a name="line.654"></a>
-<span class="sourceLineNo">655</span>        }<a name="line.655"></a>
-<span class="sourceLineNo">656</span><a name="line.656"></a>
-<span class="sourceLineNo">657</span>        final String numHashFilesArgKey = "--numhashfiles=";<a name="line.657"></a>
-<span class="sourceLineNo">658</span>        if (cmd.startsWith(numHashFilesArgKey)) {<a name="line.658"></a>
-<span class="sourceLineNo">659</span>          tableHash.numHashFiles = Integer.parseInt(cmd.substring(numHashFilesArgKey.length()));<a name="line.659"></a>
-<span class="sourceLineNo">660</span>          continue;<a name="line.660"></a>
-<span class="sourceLineNo">661</span>        }<a name="line.661"></a>
-<span class="sourceLineNo">662</span><a name="line.662"></a>
-<span class="sourceLineNo">663</span>        final String startRowArgKey = "--startrow=";<a name="line.663"></a>
-<span class="sourceLineNo">664</span>        if (cmd.startsWith(startRowArgKey)) {<a name="line.664"></a>
-<span class="sourceLineNo">665</span>          tableHash.startRow = Bytes.fromHex(cmd.substring(startRowArgKey.length()));<a name="line.665"></a>
-<span class="sourceLineNo">666</span>          continue;<a name="line.666"></a>
-<span class="sourceLineNo">667</span>        }<a name="line.667"></a>
-<span class="sourceLineNo">668</span><a name="line.668"></a>
-<span class="sourceLineNo">669</span>        final String stopRowArgKey = "--stoprow=";<a name="line.669"></a>
-<span class="sourceLineNo">670</span>        if (cmd.startsWith(stopRowArgKey)) {<a name="line.670"></a>
-<span class="sourceLineNo">671</span>          tableHash.stopRow = Bytes.fromHex(cmd.substring(stopRowArgKey.length()));<a name="line.671"></a>
-<span class="sourceLineNo">672</span>          continue;<a name="line.672"></a>
-<span class="sourceLineNo">673</span>        }<a name="line.673"></a>
-<span class="sourceLineNo">674</span><a name="line.674"></a>
-<span class="sourceLineNo">675</span>        final String startTimeArgKey = "--starttime=";<a name="line.675"></a>
-<span class="sourceLineNo">676</span>        if (cmd.startsWith(startTimeArgKey)) {<a name="line.676"></a>
-<span class="sourceLineNo">677</span>          tableHash.startTime = Long.parseLong(cmd.substring(startTimeArgKey.length()));<a name="line.677"></a>
-<span class="sourceLineNo">678</span>          continue;<a name="line.678"></a>
-<span class="sourceLineNo">679</span>        }<a name="line.679"></a>
-<span class="sourceLineNo">680</span><a name="line.680"></a>
-<span class="sourceLineNo">681</span>        final String endTimeArgKey = "--endtime=";<a name="line.681"></a>
-<span class="sourceLineNo">682</span>        if (cmd.startsWith(endTimeArgKey)) {<a name="line.682"></a>
-<span class="sourceLineNo">683</span>          tableHash.endTime = Long.parseLong(cmd.substring(endTimeArgKey.length()));<a name="line.683"></a>
-<span class="sourceLineNo">684</span>          continue;<a name="line.684"></a>
-<span class="sourceLineNo">685</span>        }<a name="line.685"></a>
-<span class="sourceLineNo">686</span><a name="line.686"></a>
-<span class="sourceLineNo">687</span>        final String scanBatchArgKey = "--scanbatch=";<a name="line.687"></a>
-<span class="sourceLineNo">688</span>        if (cmd.startsWith(scanBatchArgKey)) {<a name="line.688"></a>
-<span class="sourceLineNo">689</span>          tableHash.scanBatch = Integer.parseInt(cmd.substring(scanBatchArgKey.length()));<a name="line.689"></a>
-<span class="sourceLineNo">690</span>          continue;<a name="line.690"></a>
-<span class="sourceLineNo">691</span>        }<a name="line.691"></a>
-<span class="sourceLineNo">692</span><a name="line.692"></a>
-<span class="sourceLineNo">693</span>        final String versionsArgKey = "--versions=";<a name="line.693"></a>
-<span class="sourceLineNo">694</span>        if (cmd.startsWith(versionsArgKey)) {<a name="line.694"></a>
-<span class="sourceLineNo">695</span>          tableHash.versions = Integer.parseInt(cmd.substring(versionsArgKey.length()));<a name="line.695"></a>
-<span class="sourceLineNo">696</span>          continue;<a name="line.696"></a>
-<span class="sourceLineNo">697</span>        }<a name="line.697"></a>
-<span class="sourceLineNo">698</span><a name="line.698"></a>
-<span class="sourceLineNo">699</span>        final String familiesArgKey = "--families=";<a name="line.699"></a>
-<span class="sourceLineNo">700</span>        if (cmd.startsWith(familiesArgKey)) {<a name="line.700"></a>
-<span class="sourceLineNo">701</span>          tableHash.families = cmd.substring(familiesArgKey.length());<a name="line.701"></a>
-<span class="sourceLineNo">702</span>          continue;<a name="line.702"></a>
-<span class="sourceLineNo">703</span>        }<a name="line.703"></a>
-<span class="sourceLineNo">704</span><a name="line.704"></a>
-<span class="sourceLineNo">705</span>        printUsage("Invalid argument '" + cmd + "'");<a name="line.705"></a>
-<span class="sourceLineNo">706</span>        return false;<a name="line.706"></a>
-<span class="sourceLineNo">707</span>      }<a name="line.707"></a>
-<span class="sourceLineNo">708</span>      if ((tableHash.startTime != 0 || tableHash.endTime != 0)<a name="line.708"></a>
-<span class="sourceLineNo">709</span>          &amp;&amp; (tableHash.startTime &gt;= tableHash.endTime)) {<a name="line.709"></a>
-<span class="sourceLineNo">710</span>        printUsage("Invalid time range filter: starttime="<a name="line.710"></a>
-<span class="sourceLineNo">711</span>            + tableHash.startTime + " &gt;=  endtime=" + tableHash.endTime);<a name="line.711"></a>
-<span class="sourceLineNo">712</span>        return false;<a name="line.712"></a>
-<span class="sourceLineNo">713</span>      }<a name="line.713"></a>
-<span class="sourceLineNo">714</span><a name="line.714"></a>
-<span class="sourceLineNo">715</span>    } catch (Exception e) {<a name="line.715"></a>
-<span class="sourceLineNo">716</span>      LOG.error("Failed to parse commandLine arguments", e);<a name="line.716"></a>
-<span class="sourceLineNo">717</span>      printUsage("Can't start because " + e.getMessage());<a name="line.717"></a>
-<span class="sourceLineNo">718</span>      return false;<a name="line.718"></a>
-<span class="sourceLineNo">719</span>    }<a name="line.719"></a>
-<span class="sourceLineNo">720</span>    return true;<a name="line.720"></a>
-<span class="sourceLineNo">721</span>  }<a name="line.721"></a>
+<span class="sourceLineNo">470</span>  static class ResultHasher {<a name="line.470"></a>
+<span class="sourceLineNo">471</span>    private MessageDigest digest;<a name="line.471"></a>
+<span class="sourceLineNo">472</span><a name="line.472"></a>
+<span class="sourceLineNo">473</span>    private boolean batchStarted = false;<a name="line.473"></a>
+<span class="sourceLineNo">474</span>    private ImmutableBytesWritable batchStartKey;<a name="line.474"></a>
+<span class="sourceLineNo">475</span>    private ImmutableBytesWritable batchHash;<a name="line.475"></a>
+<span class="sourceLineNo">476</span>    private long batchSize = 0;<a name="line.476"></a>
+<span class="sourceLineNo">477</span>    boolean ignoreTimestamps;<a name="line.477"></a>
+<span class="sourceLineNo">478</span><a name="line.478"></a>
+<span class="sourceLineNo">479</span><a name="line.479"></a>
+<span class="sourceLineNo">480</span>    public ResultHasher() {<a name="line.480"></a>
+<span class="sourceLineNo">481</span>      try {<a name="line.481"></a>
+<span class="sourceLineNo">482</span>        digest = MessageDigest.getInstance("MD5");<a name="line.482"></a>
+<span class="sourceLineNo">483</span>      } catch (NoSuchAlgorithmException e) {<a name="line.483"></a>
+<span class="sourceLineNo">484</span>        Throwables.propagate(e);<a name="line.484"></a>
+<span class="sourceLineNo">485</span>      }<a name="line.485"></a>
+<span class="sourceLineNo">486</span>    }<a name="line.486"></a>
+<span class="sourceLineNo">487</span><a name="line.487"></a>
+<span class="sourceLineNo">488</span>    public void startBatch(ImmutableBytesWritable row) {<a name="line.488"></a>
+<span class="sourceLineNo">489</span>      if (batchStarted) {<a name="line.489"></a>
+<span class="sourceLineNo">490</span>        throw new RuntimeException("Cannot start new batch without finishing existing one.");<a name="line.490"></a>
+<span class="sourceLineNo">491</span>      }<a name="line.491"></a>
+<span class="sourceLineNo">492</span>      batchStarted = true;<a name="line.492"></a>
+<span class="sourceLineNo">493</span>      batchSize = 0;<a name="line.493"></a>
+<span class="sourceLineNo">494</span>      batchStartKey = row;<a name="line.494"></a>
+<span class="sourceLineNo">495</span>      batchHash = null;<a name="line.495"></a>
+<span class="sourceLineNo">496</span>    }<a name="line.496"></a>
+<span class="sourceLineNo">497</span><a name="line.497"></a>
+<span class="sourceLineNo">498</span>    public void hashResult(Result result) {<a name="line.498"></a>
+<span class="sourceLineNo">499</span>      if (!batchStarted) {<a name="line.499"></a>
+<span class="sourceLineNo">500</span>        throw new RuntimeException("Cannot add to batch that has not been started.");<a name="line.500"></a>
+<span class="sourceLineNo">501</span>      }<a name="line.501"></a>
+<span class="sourceLineNo">502</span>      for (Cell cell : result.rawCells()) {<a name="line.502"></a>
+<span class="sourceLineNo">503</span>        int rowLength = cell.getRowLength();<a name="line.503"></a>
+<span class="sourceLineNo">504</span>        int familyLength = cell.getFamilyLength();<a name="line.504"></a>
+<span class="sourceLineNo">505</span>        int qualifierLength = cell.getQualifierLength();<a name="line.505"></a>
+<span class="sourceLineNo">506</span>        int valueLength = cell.getValueLength();<a name="line.506"></a>
+<span class="sourceLineNo">507</span>        digest.update(cell.getRowArray(), cell.getRowOffset(), rowLength);<a name="line.507"></a>
+<span class="sourceLineNo">508</span>        digest.update(cell.getFamilyArray(), cell.getFamilyOffset(), familyLength);<a name="line.508"></a>
+<span class="sourceLineNo">509</span>        digest.update(cell.getQualifierArray(), cell.getQualifierOffset(), qualifierLength);<a name="line.509"></a>
+<span class="sourceLineNo">510</span><a name="line.510"></a>
+<span class="sourceLineNo">511</span>        if (!ignoreTimestamps) {<a name="line.511"></a>
+<span class="sourceLineNo">512</span>          long ts = cell.getTimestamp();<a name="line.512"></a>
+<span class="sourceLineNo">513</span>          for (int i = 8; i &gt; 0; i--) {<a name="line.513"></a>
+<span class="sourceLineNo">514</span>            digest.update((byte) ts);<a name="line.514"></a>
+<span class="sourceLineNo">515</span>            ts &gt;&gt;&gt;= 8;<a name="line.515"></a>
+<span class="sourceLineNo">516</span>          }<a name="line.516"></a>
+<span class="sourceLineNo">517</span>        }<a name="line.517"></a>
+<span class="sourceLineNo">518</span>        digest.update(cell.getValueArray(), cell.getValueOffset(), valueLength);<a name="line.518"></a>
+<span class="sourceLineNo">519</span><a name="line.519"></a>
+<span class="sourceLineNo">520</span>        batchSize += rowLength + familyLength + qualifierLength + 8 + valueLength;<a name="line.520"></a>
+<span class="sourceLineNo">521</span>      }<a name="line.521"></a>
+<span class="sourceLineNo">522</span>    }<a name="line.522"></a>
+<span class="sourceLineNo">523</span><a name="line.523"></a>
+<span class="sourceLineNo">524</span>    public void finishBatch() {<a name="line.524"></a>
+<span class="sourceLineNo">525</span>      if (!batchStarted) {<a name="line.525"></a>
+<span class="sourceLineNo">526</span>        throw new RuntimeException("Cannot finish batch that has not started.");<a name="line.526"></a>
+<span class="sourceLineNo">527</span>      }<a name="line.527"></a>
+<span class="sourceLineNo">528</span>      batchStarted = false;<a name="line.528"></a>
+<span class="sourceLineNo">529</span>      batchHash = new ImmutableBytesWritable(digest.digest());<a name="line.529"></a>
+<span class="sourceLineNo">530</span>    }<a name="line.530"></a>
+<span class="sourceLineNo">531</span><a name="line.531"></a>
+<span class="sourceLineNo">532</span>    public boolean isBatchStarted() {<a name="line.532"></a>
+<span class="sourceLineNo">533</span>      return batchStarted;<a name="line.533"></a>
+<span class="sourceLineNo">534</span>    }<a name="line.534"></a>
+<span class="sourceLineNo">535</span><a name="line.535"></a>
+<span class="sourceLineNo">536</span>    public ImmutableBytesWritable getBatchStartKey() {<a name="line.536"></a>
+<span class="sourceLineNo">537</span>      return batchStartKey;<a name="line.537"></a>
+<span class="sourceLineNo">538</span>    }<a name="line.538"></a>
+<span class="sourceLineNo">539</span><a name="line.539"></a>
+<span class="sourceLineNo">540</span>    public ImmutableBytesWritable getBatchHash() {<a name="line.540"></a>
+<span class="sourceLineNo">541</span>      return batchHash;<a name="line.541"></a>
+<span class="sourceLineNo">542</span>    }<a name="line.542"></a>
+<span class="sourceLineNo">543</span><a name="line.543"></a>
+<span class="sourceLineNo">544</span>    public long getBatchSize() {<a name="line.544"></a>
+<span class="sourceLineNo">545</span>      return batchSize;<a name="line.545"></a>
+<span class="sourceLineNo">546</span>    }<a name="line.546"></a>
+<span class="sourceLineNo">547</span>  }<a name="line.547"></a>
+<span class="sourceLineNo">548</span><a name="line.548"></a>
+<span class="sourceLineNo">549</span>  public static class HashMapper<a name="line.549"></a>
+<span class="sourceLineNo">550</span>    extends TableMapper&lt;ImmutableBytesWritable, ImmutableBytesWritable&gt; {<a name="line.550"></a>
+<span class="sourceLineNo">551</span><a name="line.551"></a>
+<span class="sourceLineNo">552</span>    private ResultHasher hasher;<a name="line.552"></a>
+<span class="sourceLineNo">553</span>    private long targetBatchSize;<a name="line.553"></a>
+<span class="sourceLineNo">554</span><a name="line.554"></a>
+<span class="sourceLineNo">555</span>    private ImmutableBytesWritable currentRow;<a name="line.555"></a>
+<span class="sourceLineNo">556</span><a name="line.556"></a>
+<span class="sourceLineNo">557</span>    @Override<a name="line.557"></a>
+<span class="sourceLineNo">558</span>    protected void setup(Context context) throws IOException, InterruptedException {<a name="line.558"></a>
+<span class="sourceLineNo">559</span>      targetBatchSize = context.getConfiguration()<a name="line.559"></a>
+<span class="sourceLineNo">560</span>          .getLong(HASH_BATCH_SIZE_CONF_KEY, DEFAULT_BATCH_SIZE);<a name="line.560"></a>
+<span class="sourceLineNo">561</span>      hasher = new ResultHasher();<a name="line.561"></a>
+<span class="sourceLineNo">562</span>      hasher.ignoreTimestamps = context.getConfiguration().<a name="line.562"></a>
+<span class="sourceLineNo">563</span>        getBoolean(IGNORE_TIMESTAMPS, false);<a name="line.563"></a>
+<span class="sourceLineNo">564</span>      TableSplit split = (TableSplit) context.getInputSplit();<a name="line.564"></a>
+<span class="sourceLineNo">565</span>      hasher.startBatch(new ImmutableBytesWritable(split.getStartRow()));<a name="line.565"></a>
+<span class="sourceLineNo">566</span>    }<a name="line.566"></a>
+<span class="sourceLineNo">567</span><a name="line.567"></a>
+<span class="sourceLineNo">568</span>    @Override<a name="line.568"></a>
+<span class="sourceLineNo">569</span>    protected void map(ImmutableBytesWritable key, Result value, Context context)<a name="line.569"></a>
+<span class="sourceLineNo">570</span>        throws IOException, InterruptedException {<a name="line.570"></a>
+<span class="sourceLineNo">571</span><a name="line.571"></a>
+<span class="sourceLineNo">572</span>      if (currentRow == null || !currentRow.equals(key)) {<a name="line.572"></a>
+<span class="sourceLineNo">573</span>        currentRow = new ImmutableBytesWritable(key); // not immutable<a name="line.573"></a>
+<span class="sourceLineNo">574</span><a name="line.574"></a>
+<span class="sourceLineNo">575</span>        if (hasher.getBatchSize() &gt;= targetBatchSize) {<a name="line.575"></a>
+<span class="sourceLineNo">576</span>          hasher.finishBatch();<a name="line.576"></a>
+<span class="sourceLineNo">577</span>          context.write(hasher.getBatchStartKey(), hasher.getBatchHash());<a name="line.577"></a>
+<span class="sourceLineNo">578</span>          hasher.startBatch(currentRow);<a name="line.578"></a>
+<span class="sourceLineNo">579</span>        }<a name="line.579"></a>
+<span class="sourceLineNo">580</span>      }<a name="line.580"></a>
+<span class="sourceLineNo">581</span><a name="line.581"></a>
+<span class="sourceLineNo">582</span>      hasher.hashResult(value);<a name="line.582"></a>
+<span class="sourceLineNo">583</span>    }<a name="line.583"></a>
+<span class="sourceLineNo">584</span><a name="line.584"></a>
+<span class="sourceLineNo">585</span>    @Override<a name="line.585"></a>
+<span class="sourceLineNo">586</span>    protected void cleanup(Context context) throws IOException, InterruptedException {<a name="line.586"></a>
+<span class="sourceLineNo">587</span>      hasher.finishBatch();<a name="line.587"></a>
+<span class="sourceLineNo">588</span>      context.write(hasher.getBatchStartKey(), hasher.getBatchHash());<a name="line.588"></a>
+<span class="sourceLineNo">589</span>    }<a name="line.589"></a>
+<span class="sourceLineNo">590</span>  }<a name="line.590"></a>
+<span class="sourceLineNo">591</span><a name="line.591"></a>
+<span class="sourceLineNo">592</span>  private void writeTempManifestFile() throws IOException {<a name="line.592"></a>
+<span class="sourceLineNo">593</span>    Path tempManifestPath = new Path(destPath, TMP_MANIFEST_FILE_NAME);<a name="line.593"></a>
+<span class="sourceLineNo">594</span>    FileSystem fs = tempManifestPath.getFileSystem(getConf());<a name="line.594"></a>
+<span class="sourceLineNo">595</span>    tableHash.writePropertiesFile(fs, tempManifestPath);<a name="line.595"></a>
+<span class="sourceLineNo">596</span>  }<a name="line.596"></a>
+<span class="sourceLineNo">597</span><a name="line.597"></a>
+<span class="sourceLineNo">598</span>  private void completeManifest() throws IOException {<a name="line.598"></a>
+<span class="sourceLineNo">599</span>    Path tempManifestPath = new Path(destPath, TMP_MANIFEST_FILE_NAME);<a name="line.599"></a>
+<span class="sourceLineNo">600</span>    Path manifestPath = new Path(destPath, MANIFEST_FILE_NAME);<a name="line.600"></a>
+<span class="sourceLineNo">601</span>    FileSystem fs = tempManifestPath.getFileSystem(getConf());<a name="line.601"></a>
+<span class="sourceLineNo">602</span>    fs.rename(tempManifestPath, manifestPath);<a name="line.602"></a>
+<span class="sourceLineNo">603</span>  }<a name="line.603"></a>
+<span class="sourceLineNo">604</span><a name="line.604"></a>
+<span class="sourceLineNo">605</span>  private static final int NUM_ARGS = 2;<a name="line.605"></a>
+<span class="sourceLineNo">606</span>  private static void printUsage(final String errorMsg) {<a name="line.606"></a>
+<span class="sourceLineNo">607</span>    if (errorMsg != null &amp;&amp; errorMsg.length() &gt; 0) {<a name="line.607"></a>
+<span class="sourceLineNo">608</span>      System.err.println("ERROR: " + errorMsg);<a name="line.608"></a>
+<span class="sourceLineNo">609</span>      System.err.println();<a name="line.609"></a>
+<span class="sourceLineNo">610</span>    }<a name="line.610"></a>
+<span class="sourceLineNo">611</span>    System.err.println("Usage: HashTable [options] &lt;tablename&gt; &lt;outputpath&gt;");<a name="line.611"></a>
+<span class="sourceLineNo">612</span>    System.err.println();<a name="line.612"></a>
+<span class="sourceLineNo">613</span>    System.err.println("Options:");<a name="line.613"></a>
+<span class="sourceLineNo">614</span>    System.err.println(" batchsize         the target amount of bytes to hash in each batch");<a name="line.614"></a>
+<span class="sourceLineNo">615</span>    System.err.println("                   rows are added to the batch until this size is reached");<a name="line.615"></a>
+<span class="sourceLineNo">616</span>    System.err.println("                   (defaults to " + DEFAULT_BATCH_SIZE + " bytes)");<a name="line.616"></a>
+<span class="sourceLineNo">617</span>    System.err.println(" numhashfiles      the number of hash files to create");<a name="line.617"></a>
+<span class="sourceLineNo">618</span>    System.err.println("                   if set to fewer than number of regions then");<a name="line.618"></a>
+<span class="sourceLineNo">619</span>    System.err.println("                   the job will create this number of reducers");<a name="line.619"></a>
+<span class="sourceLineNo">620</span>    System.err.println("                   (defaults to 1/100 of regions -- at least 1)");<a name="line.620"></a>
+<span class="sourceLineNo">621</span>    System.err.println(" startrow          the start row");<a name="line.621"></a>
+<span class="sourceLineNo">622</span>    System.err.println(" stoprow           the stop row");<a name="line.622"></a>
+<span class="sourceLineNo">623</span>    System.err.println(" starttime         beginning of the time range (unixtime in millis)");<a name="line.623"></a>
+<span class="sourceLineNo">624</span>    System.err.println("                   without endtime means from starttime to forever");<a name="line.624"></a>
+<span class="sourceLineNo">625</span>    System.err.println(" endtime           end of the time range.");<a name="line.625"></a>
+<span class="sourceLineNo">626</span>    System.err.println("                   Ignored if no starttime specified.");<a name="line.626"></a>
+<span class="sourceLineNo">627</span>    System.err.println(" scanbatch         scanner batch size to support intra row scans");<a name="line.627"></a>
+<span class="sourceLineNo">628</span>    System.err.println(" versions          number of cell versions to include");<a name="line.628"></a>
+<span class="sourceLineNo">629</span>    System.err.println(" families          comma-separated list of families to include");<a name="line.629"></a>
+<span class="sourceLineNo">630</span>    System.err.println(" ignoreTimestamps  if true, ignores cell timestamps");<a name="line.630"></a>
+<span class="sourceLineNo">631</span>    System.err.println("                   when calculating hashes");<a name="line.631"></a>
+<span class="sourceLineNo">632</span>    System.err.println();<a name="line.632"></a>
+<span class="sourceLineNo">633</span>    System.err.println("Args:");<a name="line.633"></a>
+<span class="sourceLineNo">634</span>    System.err.println(" tablename     Name of the table to hash");<a name="line.634"></a>
+<span class="sourceLineNo">635</span>    System.err.println(" outputpath    Filesystem path to put the output data");<a name="line.635"></a>
+<span class="sourceLineNo">636</span>    System.err.println();<a name="line.636"></a>
+<span class="sourceLineNo">637</span>    System.err.println("Examples:");<a name="line.637"></a>
+<span class="sourceLineNo">638</span>    System.err.println(" To hash 'TestTable' in 32kB batches for a 1 hour window into 50 files:");<a name="line.638"></a>
+<span class="sourceLineNo">639</span>    System.err.println(" $ hbase " +<a name="line.639"></a>
+<span class="sourceLineNo">640</span>        "org.apache.hadoop.hbase.mapreduce.HashTable --batchsize=32000 --numhashfiles=50"<a name="line.640"></a>
+<span class="sourceLineNo">641</span>        + " --starttime=1265875194289 --endtime=1265878794289 --families=cf2,cf3"<a name="line.641"></a>
+<span class="sourceLineNo">642</span>        + " TestTable /hashes/testTable");<a name="line.642"></a>
+<span class="sourceLineNo">643</span>  }<a name="line.643"></a>
+<span class="sourceLineNo">644</span><a name="line.644"></a>
+<span class="sourceLineNo">645</span>  private boolean doCommandLine(final String[] args) {<a name="line.645"></a>
+<span class="sourceLineNo">646</span>    if (args.length &lt; NUM_ARGS) {<a name="line.646"></a>
+<span class="sourceLineNo">647</span>      printUsage(null);<a name="line.647"></a>
+<span class="sourceLineNo">648</span>      return false;<a name="line.648"></a>
+<span class="sourceLineNo">649</span>    }<a name="line.649"></a>
+<span class="sourceLineNo">650</span>    try {<a name="line.650"></a>
+<span class="sourceLineNo">651</span><a name="line.651"></a>
+<span class="sourceLineNo">652</span>      tableHash.tableName = args[args.length-2];<a name="line.652"></a>
+<span class="sourceLineNo">653</span>      destPath = new Path(args[args.length-1]);<a name="line.653"></a>
+<span class="sourceLineNo">654</span><a name="line.654"></a>
+<span class="sourceLineNo">655</span>      for (int i = 0; i &lt; args.length - NUM_ARGS; i++) {<a name="line.655"></a>
+<span class="sourceLineNo">656</span>        String cmd = args[i];<a name="line.656"></a>
+<span class="sourceLineNo">657</span>        if (cmd.equals("-h") || cmd.startsWith("--h")) {<a name="line.657"></a>
+<span class="sourceLineNo">658</span>          printUsage(null);<a name="line.658"></a>
+<span class="sourceLineNo">659</span>          return false;<a name="line.659"></a>
+<span class="sourceLineNo">660</span>        }<a name="line.660"></a>
+<span class="sourceLineNo">661</span><a name="line.661"></a>
+<span class="sourceLineNo">662</span>        final String batchSizeArgKey = "--batchsize=";<a name="line.662"></a>
+<span class="sourceLineNo">663</span>        if (cmd.startsWith(batchSizeArgKey)) {<a name="line.663"></a>
+<span class="sourceLineNo">664</span>          tableHash.batchSize = Long.parseLong(cmd.substring(batchSizeArgKey.length()));<a name="line.664"></a>
+<span class="sourceLineNo">665</span>          continue;<a name="line.665"></a>
+<span class="sourceLineNo">666</span>        }<a name="line.666"></a>
+<span class="sourceLineNo">667</span><a name="line.667"></a>
+<span class="sourceLineNo">668</span>        final String numHashFilesArgKey = "--numhashfiles=";<a name="line.668"></a>
+<span class="sourceLineNo">669</span>        if (cmd.startsWith(numHashFilesArgKey)) {<a name="line.669"></a>
+<span class="sourceLineNo">670</span>          tableHash.numHashFiles = Integer.parseInt(cmd.substring(numHashFilesArgKey.length()));<a name="line.670"></a>
+<span class="sourceLineNo">671</span>          continue;<a name="line.671"></a>
+<span class="sourceLineNo">672</span>        }<a name="line.672"></a>
+<span class="sourceLineNo">673</span><a name="line.673"></a>
+<span class="sourceLineNo">674</span>        final String startRowArgKey = "--startrow=";<a name="line.674"></a>
+<span class="sourceLineNo">675</span>        if (cmd.startsWith(startRowArgKey)) {<a name="line.675"></a>
+<span class="sourceLineNo">676</span>          tableHash.startRow = Bytes.fromHex(cmd.substring(startRowArgKey.length()));<a name="line.676"></a>
+<span class="sourceLineNo">677</span>          continue;<a name="line.677"></a>
+<span class="sourceLineNo">678</span>        }<a name="line.678"></a>
+<span class="sourceLineNo">679</span><a name="line.679"></a>
+<span class="sourceLineNo">680</span>        final String stopRowArgKey = "--stoprow=";<a name="line.680"></a>
+<span class="sourceLineNo">681</span>        if (cmd.startsWith(stopRowArgKey)) {<a name="line.681"></a>
+<span class="sourceLineNo">682</span>          tableHash.stopRow = Bytes.fromHex(cmd.substring(stopRowArgKey.length()));<a name="line.682"></a>
+<span class="sourceLineNo">683</span>          continue;<a name="line.683"></a>
+<span class="sourceLineNo">684</span>        }<a name="line.684"></a>
+<span class="sourceLineNo">685</span><a name="line.685"></a>
+<span class="sourceLineNo">686</span>        final String startTimeArgKey = "--starttime=";<a name="line.686"></a>
+<span class="sourceLineNo">687</span>        if (cmd.startsWith(startTimeArgKey)) {<a name="line.687"></a>
+<span class="sourceLineNo">688</span>          tableHash.startTime = Long.parseLong(cmd.substring(startTimeArgKey.length()));<a name="line.688"></a>
+<span class="sourceLineNo">689</span>          continue;<a name="line.689"></a>
+<span class="sourceLineNo">690</span>        }<a name="line.690"></a>
+<span class="sourceLineNo">691</span><a name="line.691"></a>
+<span class="sourceLineNo">692</span>        final String endTimeArgKey = "--endtime=";<a name="line.692"></a>
+<span class="sourceLineNo">693</span>        if (cmd.startsWith(endTimeArgKey)) {<a name="line.693"></a>
+<span class="sourceLineNo">694</span>          tableHash.endTime = Long.parseLong(cmd.substring(endTimeArgKey.length()));<a name="line.694"></a>
+<span class="sourceLineNo">695</span>          continue;<a name="line.695"></a>
+<span class="sourceLineNo">696</span>        }<a name="line.696"></a>
+<span class="sourceLineNo">697</span><a name="line.697"></a>
+<span class="sourceLineNo">698</span>        final String scanBatchArgKey = "--scanbatch=";<a name="line.698"></a>
+<span class="sourceLineNo">699</span>        if (cmd.startsWith(scanBatchArgKey)) {<a name="line.699"></a>
+<span class="sourceLineNo">700</span>          tableHash.scanBatch = Integer.parseInt(cmd.substring(scanBatchArgKey.length()));<a name="line.700"></a>
+<span class="sourceLineNo">701</span>          continue;<a name="line.701"></a>
+<span class="sourceLineNo">702</span>        }<a name="line.702"></a>
+<span class="sourceLineNo">703</span><a name="line.703"></a>
+<span class="sourceLineNo">704</span>        final String versionsArgKey = "--versions=";<a name="line.704"></a>
+<span class="sourceLineNo">705</span>        if (cmd.startsWith(versionsArgKey)) {<a name="line.705"></a>
+<span class="sourceLineNo">706</span>          tableHash.versions = Integer.parseInt(cmd.substring(versionsArgKey.length()));<a name="line.706"></a>
+<span class="sourceLineNo">707</span>          continue;<a name="line.707"></a>
+<span class="sourceLineNo">708</span>        }<a name="line.708"></a>
+<span class="sourceLineNo">709</span><a name="line.709"></a>
+<span class="sourceLineNo">710</span>        final String familiesArgKey = "--families=";<a name="line.710"></a>
+<span class="sourceLineNo">711</span>        if (cmd.startsWith(familiesArgKey)) {<a name="line.711"></a>
+<span class="sourceLineNo">712</span>          tableHash.families = cmd.substring(familiesArgKey.length());<a name="line.712"></a>
+<span class="sourceLineNo">713</span>          continue;<a name="line.713"></a>
+<span class="sourceLineNo">714</span>        }<a name="line.714"></a>
+<span class="sourceLineNo">715</span><a name="line.715"></a>
+<span class="sourceLineNo">716</span>        final String ignoreTimestampsKey = "--ignoreTimestamps=";<a name="line.716"></a>
+<span class="sourceLineNo">717</span>        if (cmd.startsWith(ignoreTimestampsKey)) {<a name="line.717"></a>
+<span class="sourceLineNo">718</span>          tableHash.ignoreTimestamps = Boolean.<a name="line.718"></a>
+<span class="sourceLineNo">719</span>            parseBoolean(cmd.substring(ignoreTimestampsKey.length()));<a name="line.719"></a>
+<span class="sourceLineNo">720</span>          continue;<a name="line.720"></a>
+<span class="sourceLineNo">721</span>        }<a name="line.721"></a>
 <span class="sourceLineNo">722</span><a name="line.722"></a>
-<span class="sourceLineNo">723</span>  /**<a name="line.723"></a>
-<span class="sourceLineNo">724</span>   * Main entry point.<a name="line.724"></a>
-<span class="sourceLineNo">725</span>   */<a name="line.725"></a>
-<span class="sourceLineNo">726</span>  public static void main(String[] args) throws Exception {<a name="line.726"></a>
-<span class="sourceLineNo">727</span>    int ret = ToolRunner.run(new HashTable(HBaseConfiguration.create()), args);<a name="line.727"></a>
-<span class="sourceLineNo">728</span>    System.exit(ret);<a name="line.728"></a>
-<span class="sourceLineNo">729</span>  }<a name="line.729"></a>
-<span class="sourceLineNo">730</span><a name="line.730"></a>
-<span class="sourceLineNo">731</span>  @Override<a name="line.731"></a>
-<span class="sourceLineNo">732</span>  public int run(String[] args) throws Exception {<a name="line.732"></a>
-<span class="sourceLineNo">733</span>    String[] otherArgs = new GenericOptionsParser(getConf(), args).getRemainingArgs();<a name="line.733"></a>
-<span class="sourceLineNo">734</span>    if (!doCommandLine(otherArgs)) {<a name="line.734"></a>
-<span class="sourceLineNo">735</span>      return 1;<a name="line.735"></a>
-<span class="sourceLineNo">736</span>    }<a name="line.736"></a>
-<span class="sourceLineNo">737</span><a name="line.737"></a>
-<span class="sourceLineNo">738</span>    Job job = createSubmittableJob(otherArgs);<a name="line.738"></a>
-<span class="sourceLineNo">739</span>    writeTempManifestFile();<a name="line.739"></a>
-<span class="sourceLineNo">740</span>    if (!job.waitForCompletion(true)) {<a name="line.740"></a>
-<span class="sourceLineNo">741</span>      LOG.info("Map-reduce job failed!");<a name="line.741"></a>
-<span class="sourceLineNo">742</span>      return 1;<a name="line.742"></a>
-<span class="sourceLineNo">743</span>    }<a name="line.743"></a>
-<span class="sourceLineNo">744</span>    completeManifest();<a name="line.744"></a>
-<span class="sourceLineNo">745</span>    return 0;<a name="line.745"></a>
-<span class="sourceLineNo">746</span>  }<a name="line.746"></a>
-<span class="sourceLineNo">747</span><a name="line.747"></a>
-<span class="sourceLineNo">748</span>}<a name="line.748"></a>
+<span class="sourceLineNo">723</span>        printUsage("Invalid argument '" + cmd + "'");<a name="line.723"></a>
+<span class="sourceLineNo">724</span>        return false;<a name="line.724"></a>
+<span class="sourceLineNo">725</span>      }<a name="line.725"></a>
+<span class="sourceLineNo">726</span>      if ((tableHash.startTime != 0 || tableHash.endTime != 0)<a name="line.726"></a>
+<span class="sourceLineNo">727</span>          &amp;&amp; (tableHash.startTime &gt;= tableHash.endTime)) {<a name="line.727"></a>
+<span class="sourceLineNo">728</span>        printUsage("Invalid time range filter: starttime="<a name="line.728"></a>
+<span class="sourceLineNo">729</span>            + tableHash.startTime + " &gt;=  endtime=" + tableHash.endTime);<a name="line.729"></a>
+<span class="sourceLineNo">730</span>        return false;<a name="line.730"></a>
+<span class="sourceLineNo">731</span>      }<a name="line.731"></a>
+<span class="sourceLineNo">732</span><a name="line.732"></a>
+<span class="sourceLineNo">733</span>    } catch (Exception e) {<a name="line.733"></a>
+<span class="sourceLineNo">734</span>      LOG.error("Failed to parse commandLine arguments", e);<a name="line.734"></a>
+<span class="sourceLineNo">735</span>      printUsage("Can't start because " + e.getMessage());<a name="line.735"></a>
+<span class="sourceLineNo">736</span>      return false;<a name="line.736"></a>
+<span class="sourceLineNo">737</span>    }<a name="line.737"></a>
+<span class="sourceLineNo">738</span>    return true;<a name="line.738"></a>
+<span class="sourceLineNo">739</span>  }<a name="line.739"></a>
+<span class="sourceLineNo">740</span><a name="line.740"></a>
+<span class="sourceLineNo">741</span>  /**<a name="line.741"></a>
+<span class="sourceLineNo">742</span>   * Main entry point.<a name="line.742"></a>
+<span class="sourceLineNo">743</span>   */<a name="line.743"></a>
+<span class="sourceLineNo">744</span>  public static void main(String[] args) throws Exception {<a name="line.744"></a>
+<span class="sourceLineNo">745</span>    int ret = ToolRunner.run(new HashTable(HBaseConfiguration.create()), args);<a name="line.745"></a>
+<span class="sourceLineNo">746</span>    System.exit(ret);<a name="line.746"></a>
+<span class="sourceLineNo">747</span>  }<a name="line.747"></a>
+<span class="sourceLineNo">748</span><a name="line.748"></a>
+<span class="sourceLineNo">749</span>  @Override<a name="line.749"></a>
+<span class="sourceLineNo">750</span>  public int run(String[] args) throws Exception {<a name="line.750"></a>
+<span class="sourceLineNo">751</span>    String[] otherArgs = new GenericOptionsParser(getConf(), args).getRemainingArgs();<a name="line.751"></a>
+<span class="sourceLineNo">752</span>    if (!doCommandLine(otherArgs)) {<a name="line.752"></a>
+<span class="sourceLineNo">753</span>      return 1;<a name="line.753"></a>
+<span class="sourceLineNo">754</span>    }<a name="line.754"></a>
+<span class="sourceLineNo">755</span><a name="line.755"></a>
+<span class="sourceLineNo">756</span>    Job job = createSubmittableJob(otherArgs);<a name="line.756"></a>
+<span class="sourceLineNo">757</span>    writeTempManifestFile();<a name="line.757"></a>
+<span class="sourceLineNo">758</span>    if (!job.waitForCompletion(true)) {<a name="line.758"></a>
+<span class="sourceLineNo">759</span>      LOG.info("Map-reduce job failed!");<a name="line.759"></a>
+<span class="sourceLineNo">760</span>      return 1;<a name="line.760"></a>
+<span class="sourceLineNo">761</span>    }<a name="line.761"></a>
+<span class="sourceLineNo">762</span>    completeManifest();<a name="line.762"></a>
+<span class="sourceLineNo">763</span>    return 0;<a name="line.763"></a>
+<span class="sourceLineNo">764</span>  }<a name="line.764"></a>
+<span class="sourceLineNo">765</span><a name="line.765"></a>
+<span class="sourceLineNo">766</span>}<a name="line.766"></a>
 
 
 
diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html
index 55fc1c9..32934c7 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HashTable.html
@@ -81,679 +81,697 @@
 <span class="sourceLineNo">073</span>  final static String MANIFEST_FILE_NAME = "manifest";<a name="line.73"></a>
 <span class="sourceLineNo">074</span>  final static String HASH_DATA_DIR = "hashes";<a name="line.74"></a>
 <span class="sourceLineNo">075</span>  final static String OUTPUT_DATA_FILE_PREFIX = "part-r-";<a name="line.75"></a>
-<span class="sourceLineNo">076</span>  private final static String TMP_MANIFEST_FILE_NAME = "manifest.tmp";<a name="line.76"></a>
-<span class="sourceLineNo">077</span><a name="line.77"></a>
-<span class="sourceLineNo">078</span>  TableHash tableHash = new TableHash();<a name="line.78"></a>
-<span class="sourceLineNo">079</span>  Path destPath;<a name="line.79"></a>
-<span class="sourceLineNo">080</span><a name="line.80"></a>
-<span class="sourceLineNo">081</span>  public HashTable(Configuration conf) {<a name="line.81"></a>
-<span class="sourceLineNo">082</span>    super(conf);<a name="line.82"></a>
-<span class="sourceLineNo">083</span>  }<a name="line.83"></a>
-<span class="sourceLineNo">084</span><a name="line.84"></a>
-<span class="sourceLineNo">085</span>  public static class TableHash {<a name="line.85"></a>
-<span class="sourceLineNo">086</span><a name="line.86"></a>
-<span class="sourceLineNo">087</span>    Path hashDir;<a name="line.87"></a>
-<span class="sourceLineNo">088</span><a name="line.88"></a>
-<span class="sourceLineNo">089</span>    String tableName;<a name="line.89"></a>
-<span class="sourceLineNo">090</span>    String families = null;<a name="line.90"></a>
-<span class="sourceLineNo">091</span>    long batchSize = DEFAULT_BATCH_SIZE;<a name="line.91"></a>
-<span class="sourceLineNo">092</span>    int numHashFiles = 0;<a name="line.92"></a>
-<span class="sourceLineNo">093</span>    byte[] startRow = HConstants.EMPTY_START_ROW;<a name="line.93"></a>
-<span class="sourceLineNo">094</span>    byte[] stopRow = HConstants.EMPTY_END_ROW;<a name="line.94"></a>
-<span class="sourceLineNo">095</span>    int scanBatch = 0;<a name="line.95"></a>
-<span class="sourceLineNo">096</span>    int versions = -1;<a name="line.96"></a>
-<span class="sourceLineNo">097</span>    long startTime = 0;<a name="line.97"></a>
-<span class="sourceLineNo">098</span>    long endTime = 0;<a name="line.98"></a>
-<span class="sourceLineNo">099</span><a name="line.99"></a>
-<span class="sourceLineNo">100</span>    List&lt;ImmutableBytesWritable&gt; partitions;<a name="line.100"></a>
+<span class="sourceLineNo">076</span>  final static String IGNORE_TIMESTAMPS = "ignoreTimestamps";<a name="line.76"></a>
+<span class="sourceLineNo">077</span>  private final static String TMP_MANIFEST_FILE_NAME = "manifest.tmp";<a name="line.77"></a>
+<span class="sourceLineNo">078</span><a name="line.78"></a>
+<span class="sourceLineNo">079</span>  TableHash tableHash = new TableHash();<a name="line.79"></a>
+<span class="sourceLineNo">080</span>  Path destPath;<a name="line.80"></a>
+<span class="sourceLineNo">081</span><a name="line.81"></a>
+<span class="sourceLineNo">082</span>  public HashTable(Configuration conf) {<a name="line.82"></a>
+<span class="sourceLineNo">083</span>    super(conf);<a name="line.83"></a>
+<span class="sourceLineNo">084</span>  }<a name="line.84"></a>
+<span class="sourceLineNo">085</span><a name="line.85"></a>
+<span class="sourceLineNo">086</span>  public static class TableHash {<a name="line.86"></a>
+<span class="sourceLineNo">087</span><a name="line.87"></a>
+<span class="sourceLineNo">088</span>    Path hashDir;<a name="line.88"></a>
+<span class="sourceLineNo">089</span><a name="line.89"></a>
+<span class="sourceLineNo">090</span>    String tableName;<a name="line.90"></a>
+<span class="sourceLineNo">091</span>    String families = null;<a name="line.91"></a>
+<span class="sourceLineNo">092</span>    long batchSize = DEFAULT_BATCH_SIZE;<a name="line.92"></a>
+<span class="sourceLineNo">093</span>    int numHashFiles = 0;<a name="line.93"></a>
+<span class="sourceLineNo">094</span>    byte[] startRow = HConstants.EMPTY_START_ROW;<a name="line.94"></a>
+<span class="sourceLineNo">095</span>    byte[] stopRow = HConstants.EMPTY_END_ROW;<a name="line.95"></a>
+<span class="sourceLineNo">096</span>    int scanBatch = 0;<a name="line.96"></a>
+<span class="sourceLineNo">097</span>    int versions = -1;<a name="line.97"></a>
+<span class="sourceLineNo">098</span>    long startTime = 0;<a name="line.98"></a>
+<span class="sourceLineNo">099</span>    long endTime = 0;<a name="line.99"></a>
+<span class="sourceLineNo">100</span>    boolean ignoreTimestamps;<a name="line.100"></a>
 <span class="sourceLineNo">101</span><a name="line.101"></a>
-<span class="sourceLineNo">102</span>    public static TableHash read(Configuration conf, Path hashDir) throws IOException {<a name="line.102"></a>
-<span class="sourceLineNo">103</span>      TableHash tableHash = new TableHash();<a name="line.103"></a>
-<span class="sourceLineNo">104</span>      FileSystem fs = hashDir.getFileSystem(conf);<a name="line.104"></a>
-<span class="sourceLineNo">105</span>      tableHash.hashDir = hashDir;<a name="line.105"></a>
-<span class="sourceLineNo">106</span>      tableHash.readPropertiesFile(fs, new Path(hashDir, MANIFEST_FILE_NAME));<a name="line.106"></a>
-<span class="sourceLineNo">107</span>      tableHash.readPartitionFile(fs, conf, new Path(hashDir, PARTITIONS_FILE_NAME));<a name="line.107"></a>
-<span class="sourceLineNo">108</span>      return tableHash;<a name="line.108"></a>
-<span class="sourceLineNo">109</span>    }<a name="line.109"></a>
-<span class="sourceLineNo">110</span><a name="line.110"></a>
-<span class="sourceLineNo">111</span>    void writePropertiesFile(FileSystem fs, Path path) throws IOException {<a name="line.111"></a>
-<span class="sourceLineNo">112</span>      Properties p = new Properties();<a name="line.112"></a>
-<span class="sourceLineNo">113</span>      p.setProperty("table", tableName);<a name="line.113"></a>
-<span class="sourceLineNo">114</span>      if (families != null) {<a name="line.114"></a>
-<span class="sourceLineNo">115</span>        p.setProperty("columnFamilies", families);<a name="line.115"></a>
-<span class="sourceLineNo">116</span>      }<a name="line.116"></a>
-<span class="sourceLineNo">117</span>      p.setProperty("targetBatchSize", Long.toString(batchSize));<a name="line.117"></a>
-<span class="sourceLineNo">118</span>      p.setProperty("numHashFiles", Integer.toString(numHashFiles));<a name="line.118"></a>
-<span class="sourceLineNo">119</span>      if (!isTableStartRow(startRow)) {<a name="line.119"></a>
-<span class="sourceLineNo">120</span>        p.setProperty("startRowHex", Bytes.toHex(startRow));<a name="line.120"></a>
-<span class="sourceLineNo">121</span>      }<a name="line.121"></a>
-<span class="sourceLineNo">122</span>      if (!isTableEndRow(stopRow)) {<a name="line.122"></a>
-<span class="sourceLineNo">123</span>        p.setProperty("stopRowHex", Bytes.toHex(stopRow));<a name="line.123"></a>
-<span class="sourceLineNo">124</span>      }<a name="line.124"></a>
-<span class="sourceLineNo">125</span>      if (scanBatch &gt; 0) {<a name="line.125"></a>
-<span class="sourceLineNo">126</span>        p.setProperty("scanBatch", Integer.toString(scanBatch));<a name="line.126"></a>
-<span class="sourceLineNo">127</span>      }<a name="line.127"></a>
-<span class="sourceLineNo">128</span>      if (versions &gt;= 0) {<a name="line.128"></a>
-<span class="sourceLineNo">129</span>        p.setProperty("versions", Integer.toString(versions));<a name="line.129"></a>
-<span class="sourceLineNo">130</span>      }<a name="line.130"></a>
-<span class="sourceLineNo">131</span>      if (startTime != 0) {<a name="line.131"></a>
-<span class="sourceLineNo">132</span>        p.setProperty("startTimestamp", Long.toString(startTime));<a name="line.132"></a>
-<span class="sourceLineNo">133</span>      }<a name="line.133"></a>
-<span class="sourceLineNo">134</span>      if (endTime != 0) {<a name="line.134"></a>
-<span class="sourceLineNo">135</span>        p.setProperty("endTimestamp", Long.toString(endTime));<a name="line.135"></a>
-<span class="sourceLineNo">136</span>      }<a name="line.136"></a>
-<span class="sourceLineNo">137</span><a name="line.137"></a>
-<span class="sourceLineNo">138</span>      try (OutputStreamWriter osw = new OutputStreamWriter(fs.create(path), Charsets.UTF_8)) {<a name="line.138"></a>
-<span class="sourceLineNo">139</span>        p.store(osw, null);<a name="line.139"></a>
-<span class="sourceLineNo">140</span>      }<a name="line.140"></a>
-<span class="sourceLineNo">141</span>    }<a name="line.141"></a>
-<span class="sourceLineNo">142</span><a name="line.142"></a>
-<span class="sourceLineNo">143</span>    void readPropertiesFile(FileSystem fs, Path path) throws IOException {<a name="line.143"></a>
-<span class="sourceLineNo">144</span>      Properties p = new Properties();<a name="line.144"></a>
-<span class="sourceLineNo">145</span>      try (FSDataInputStream in = fs.open(path)) {<a name="line.145"></a>
-<span class="sourceLineNo">146</span>        try (InputStreamReader isr = new InputStreamReader(in, Charsets.UTF_8)) {<a name="line.146"></a>
-<span class="sourceLineNo">147</span>          p.load(isr);<a name="line.147"></a>
-<span class="sourceLineNo">148</span>        }<a name="line.148"></a>
-<span class="sourceLineNo">149</span>      }<a name="line.149"></a>
-<span class="sourceLineNo">150</span>      tableName = p.getProperty("table");<a name="line.150"></a>
-<span class="sourceLineNo">151</span>      families = p.getProperty("columnFamilies");<a name="line.151"></a>
-<span class="sourceLineNo">152</span>      batchSize = Long.parseLong(p.getProperty("targetBatchSize"));<a name="line.152"></a>
-<span class="sourceLineNo">153</span>      numHashFiles = Integer.parseInt(p.getProperty("numHashFiles"));<a name="line.153"></a>
-<span class="sourceLineNo">154</span><a name="line.154"></a>
-<span class="sourceLineNo">155</span>      String startRowHex = p.getProperty("startRowHex");<a name="line.155"></a>
-<span class="sourceLineNo">156</span>      if (startRowHex != null) {<a name="line.156"></a>
-<span class="sourceLineNo">157</span>        startRow = Bytes.fromHex(startRowHex);<a name="line.157"></a>
-<span class="sourceLineNo">158</span>      }<a name="line.158"></a>
-<span class="sourceLineNo">159</span>      String stopRowHex = p.getProperty("stopRowHex");<a name="line.159"></a>
-<span class="sourceLineNo">160</span>      if (stopRowHex != null) {<a name="line.160"></a>
-<span class="sourceLineNo">161</span>        stopRow = Bytes.fromHex(stopRowHex);<a name="line.161"></a>
-<span class="sourceLineNo">162</span>      }<a name="line.162"></a>
-<span class="sourceLineNo">163</span><a name="line.163"></a>
-<span class="sourceLineNo">164</span>      String scanBatchString = p.getProperty("scanBatch");<a name="line.164"></a>
-<span class="sourceLineNo">165</span>      if (scanBatchString != null) {<a name="line.165"></a>
-<span class="sourceLineNo">166</span>        scanBatch = Integer.parseInt(scanBatchString);<a name="line.166"></a>
-<span class="sourceLineNo">167</span>      }<a name="line.167"></a>
-<span class="sourceLineNo">168</span><a name="line.168"></a>
-<span class="sourceLineNo">169</span>      String versionString = p.getProperty("versions");<a name="line.169"></a>
-<span class="sourceLineNo">170</span>      if (versionString != null) {<a name="line.170"></a>
-<span class="sourceLineNo">171</span>        versions = Integer.parseInt(versionString);<a name="line.171"></a>
-<span class="sourceLineNo">172</span>      }<a name="line.172"></a>
-<span class="sourceLineNo">173</span><a name="line.173"></a>
-<span class="sourceLineNo">174</span>      String startTimeString = p.getProperty("startTimestamp");<a name="line.174"></a>
-<span class="sourceLineNo">175</span>      if (startTimeString != null) {<a name="line.175"></a>
-<span class="sourceLineNo">176</span>        startTime = Long.parseLong(startTimeString);<a name="line.176"></a>
-<span class="sourceLineNo">177</span>      }<a name="line.177"></a>
-<span class="sourceLineNo">178</span><a name="line.178"></a>
-<span class="sourceLineNo">179</span>      String endTimeString = p.getProperty("endTimestamp");<a name="line.179"></a>
-<span class="sourceLineNo">180</span>      if (endTimeString != null) {<a name="line.180"></a>
-<span class="sourceLineNo">181</span>        endTime = Long.parseLong(endTimeString);<a name="line.181"></a>
-<span class="sourceLineNo">182</span>      }<a name="line.182"></a>
-<span class="sourceLineNo">183</span>    }<a name="line.183"></a>
-<span class="sourceLineNo">184</span><a name="line.184"></a>
-<span class="sourceLineNo">185</span>    Scan initScan() throws IOException {<a name="line.185"></a>
-<span class="sourceLineNo">186</span>      Scan scan = new Scan();<a name="line.186"></a>
-<span class="sourceLineNo">187</span>      scan.setCacheBlocks(false);<a name="line.187"></a>
-<span class="sourceLineNo">188</span>      if (startTime != 0 || endTime != 0) {<a name="line.188"></a>
-<span class="sourceLineNo">189</span>        scan.setTimeRange(startTime, endTime == 0 ? HConstants.LATEST_TIMESTAMP : endTime);<a name="line.189"></a>
-<span class="sourceLineNo">190</span>      }<a name="line.190"></a>
-<span class="sourceLineNo">191</span>      if (scanBatch &gt; 0) {<a name="line.191"></a>
-<span class="sourceLineNo">192</span>        scan.setBatch(scanBatch);<a name="line.192"></a>
-<span class="sourceLineNo">193</span>      }<a name="line.193"></a>
-<span class="sourceLineNo">194</span>      if (versions &gt;= 0) {<a name="line.194"></a>
-<span class="sourceLineNo">195</span>        scan.readVersions(versions);<a name="line.195"></a>
-<span class="sourceLineNo">196</span>      }<a name="line.196"></a>
-<span class="sourceLineNo">197</span>      if (!isTableStartRow(startRow)) {<a name="line.197"></a>
-<span class="sourceLineNo">198</span>        scan.withStartRow(startRow);<a name="line.198"></a>
-<span class="sourceLineNo">199</span>      }<a name="line.199"></a>
-<span class="sourceLineNo">200</span>      if (!isTableEndRow(stopRow)) {<a name="line.200"></a>
-<span class="sourceLineNo">201</span>        scan.withStopRow(stopRow);<a name="line.201"></a>
-<span class="sourceLineNo">202</span>      }<a name="line.202"></a>
-<span class="sourceLineNo">203</span>      if(families != null) {<a name="line.203"></a>
-<span class="sourceLineNo">204</span>        for(String fam : families.split(",")) {<a name="line.204"></a>
-<span class="sourceLineNo">205</span>          scan.addFamily(Bytes.toBytes(fam));<a name="line.205"></a>
-<span class="sourceLineNo">206</span>        }<a name="line.206"></a>
-<span class="sourceLineNo">207</span>      }<a name="line.207"></a>
-<span class="sourceLineNo">208</span>      return scan;<a name="line.208"></a>
-<span class="sourceLineNo">209</span>    }<a name="line.209"></a>
-<span class="sourceLineNo">210</span><a name="line.210"></a>
-<span class="sourceLineNo">211</span>    /**<a name="line.211"></a>
-<span class="sourceLineNo">212</span>     * Choose partitions between row ranges to hash to a single output file<a name="line.212"></a>
-<span class="sourceLineNo">213</span>     * Selects region boundaries that fall within the scan range, and groups them<a name="line.213"></a>
-<span class="sourceLineNo">214</span>     * into the desired number of partitions.<a name="line.214"></a>
-<span class="sourceLineNo">215</span>     */<a name="line.215"></a>
-<span class="sourceLineNo">216</span>    void selectPartitions(Pair&lt;byte[][], byte[][]&gt; regionStartEndKeys) {<a name="line.216"></a>
-<span class="sourceLineNo">217</span>      List&lt;byte[]&gt; startKeys = new ArrayList&lt;&gt;();<a name="line.217"></a>
-<span class="sourceLineNo">218</span>      for (int i = 0; i &lt; regionStartEndKeys.getFirst().length; i++) {<a name="line.218"></a>
-<span class="sourceLineNo">219</span>        byte[] regionStartKey = regionStartEndKeys.getFirst()[i];<a name="line.219"></a>
-<span class="sourceLineNo">220</span>        byte[] regionEndKey = regionStartEndKeys.getSecond()[i];<a name="line.220"></a>
-<span class="sourceLineNo">221</span><a name="line.221"></a>
-<span class="sourceLineNo">222</span>        // if scan begins after this region, or starts before this region, then drop this region<a name="line.222"></a>
-<span class="sourceLineNo">223</span>        // in other words:<a name="line.223"></a>
-<span class="sourceLineNo">224</span>        //   IF (scan begins before the end of this region<a name="line.224"></a>
-<span class="sourceLineNo">225</span>        //      AND scan ends before the start of this region)<a name="line.225"></a>
-<span class="sourceLineNo">226</span>        //   THEN include this region<a name="line.226"></a>
-<span class="sourceLineNo">227</span>        if ((isTableStartRow(startRow) || isTableEndRow(regionEndKey)<a name="line.227"></a>
-<span class="sourceLineNo">228</span>            || Bytes.compareTo(startRow, regionEndKey) &lt; 0)<a name="line.228"></a>
-<span class="sourceLineNo">229</span>          &amp;&amp; (isTableEndRow(stopRow) || isTableStartRow(regionStartKey)<a name="line.229"></a>
-<span class="sourceLineNo">230</span>            || Bytes.compareTo(stopRow, regionStartKey) &gt; 0)) {<a name="line.230"></a>
-<span class="sourceLineNo">231</span>          startKeys.add(regionStartKey);<a name="line.231"></a>
-<span class="sourceLineNo">232</span>        }<a name="line.232"></a>
-<span class="sourceLineNo">233</span>      }<a name="line.233"></a>
-<span class="sourceLineNo">234</span><a name="line.234"></a>
-<span class="sourceLineNo">235</span>      int numRegions = startKeys.size();<a name="line.235"></a>
-<span class="sourceLineNo">236</span>      if (numHashFiles == 0) {<a name="line.236"></a>
-<span class="sourceLineNo">237</span>        numHashFiles = numRegions / 100;<a name="line.237"></a>
-<span class="sourceLineNo">238</span>      }<a name="line.238"></a>
-<span class="sourceLineNo">239</span>      if (numHashFiles == 0) {<a name="line.239"></a>
-<span class="sourceLineNo">240</span>        numHashFiles = 1;<a name="line.240"></a>
-<span class="sourceLineNo">241</span>      }<a name="line.241"></a>
-<span class="sourceLineNo">242</span>      if (numHashFiles &gt; numRegions) {<a name="line.242"></a>
-<span class="sourceLineNo">243</span>        // can't partition within regions<a name="line.243"></a>
-<span class="sourceLineNo">244</span>        numHashFiles = numRegions;<a name="line.244"></a>
-<span class="sourceLineNo">245</span>      }<a name="line.245"></a>
-<span class="sourceLineNo">246</span><a name="line.246"></a>
-<span class="sourceLineNo">247</span>      // choose a subset of start keys to group regions into ranges<a name="line.247"></a>
-<span class="sourceLineNo">248</span>      partitions = new ArrayList&lt;&gt;(numHashFiles - 1);<a name="line.248"></a>
-<span class="sourceLineNo">249</span>      // skip the first start key as it is not a partition between ranges.<a name="line.249"></a>
-<span class="sourceLineNo">250</span>      for (long i = 1; i &lt; numHashFiles; i++) {<a name="line.250"></a>
-<span class="sourceLineNo">251</span>        int splitIndex = (int) (numRegions * i / numHashFiles);<a name="line.251"></a>
-<span class="sourceLineNo">252</span>        partitions.add(new ImmutableBytesWritable(startKeys.get(splitIndex)));<a name="line.252"></a>
-<span class="sourceLineNo">253</span>      }<a name="line.253"></a>
-<span class="sourceLineNo">254</span>    }<a name="line.254"></a>
-<span class="sourceLineNo">255</span><a name="line.255"></a>
-<span class="sourceLineNo">256</span>    void writePartitionFile(Configuration conf, Path path) throws IOException {<a name="line.256"></a>
-<span class="sourceLineNo">257</span>      FileSystem fs = path.getFileSystem(conf);<a name="line.257"></a>
-<span class="sourceLineNo">258</span>      @SuppressWarnings("deprecation")<a name="line.258"></a>
-<span class="sourceLineNo">259</span>      SequenceFile.Writer writer = SequenceFile.createWriter(<a name="line.259"></a>
-<span class="sourceLineNo">260</span>        fs, conf, path, ImmutableBytesWritable.class, NullWritable.class);<a name="line.260"></a>
-<span class="sourceLineNo">261</span><a name="line.261"></a>
-<span class="sourceLineNo">262</span>      for (int i = 0; i &lt; partitions.size(); i++) {<a name="line.262"></a>
-<span class="sourceLineNo">263</span>        writer.append(partitions.get(i), NullWritable.get());<a name="line.263"></a>
-<span class="sourceLineNo">264</span>      }<a name="line.264"></a>
-<span class="sourceLineNo">265</span>      writer.close();<a name="line.265"></a>
-<span class="sourceLineNo">266</span>    }<a name="line.266"></a>
-<span class="sourceLineNo">267</span><a name="line.267"></a>
-<span class="sourceLineNo">268</span>    private void readPartitionFile(FileSystem fs, Configuration conf, Path path)<a name="line.268"></a>
-<span class="sourceLineNo">269</span>         throws IOException {<a name="line.269"></a>
-<span class="sourceLineNo">270</span>      @SuppressWarnings("deprecation")<a name="line.270"></a>
-<span class="sourceLineNo">271</span>      SequenceFile.Reader reader = new SequenceFile.Reader(fs, path, conf);<a name="line.271"></a>
-<span class="sourceLineNo">272</span>      ImmutableBytesWritable key = new ImmutableBytesWritable();<a name="line.272"></a>
-<span class="sourceLineNo">273</span>      partitions = new ArrayList&lt;&gt;();<a name="line.273"></a>
-<span class="sourceLineNo">274</span>      while (reader.next(key)) {<a name="line.274"></a>
-<span class="sourceLineNo">275</span>        partitions.add(new ImmutableBytesWritable(key.copyBytes()));<a name="line.275"></a>
-<span class="sourceLineNo">276</span>      }<a name="line.276"></a>
-<span class="sourceLineNo">277</span>      reader.close();<a name="line.277"></a>
-<span class="sourceLineNo">278</span><a name="line.278"></a>
-<span class="sourceLineNo">279</span>      if (!Ordering.natural().isOrdered(partitions)) {<a name="line.279"></a>
-<span class="sourceLineNo">280</span>        throw new IOException("Partitions are not ordered!");<a name="line.280"></a>
-<span class="sourceLineNo">281</span>      }<a name="line.281"></a>
-<span class="sourceLineNo">282</span>    }<a name="line.282"></a>
-<span class="sourceLineNo">283</span><a name="line.283"></a>
-<span class="sourceLineNo">284</span>    @Override<a name="line.284"></a>
-<span class="sourceLineNo">285</span>    public String toString() {<a name="line.285"></a>
-<span class="sourceLineNo">286</span>      StringBuilder sb = new StringBuilder();<a name="line.286"></a>
-<span class="sourceLineNo">287</span>      sb.append("tableName=").append(tableName);<a name="line.287"></a>
-<span class="sourceLineNo">288</span>      if (families != null) {<a name="line.288"></a>
-<span class="sourceLineNo">289</span>        sb.append(", families=").append(families);<a name="line.289"></a>
-<span class="sourceLineNo">290</span>      }<a name="line.290"></a>
-<span class="sourceLineNo">291</span>      sb.append(", batchSize=").append(batchSize);<a name="line.291"></a>
-<span class="sourceLineNo">292</span>      sb.append(", numHashFiles=").append(numHashFiles);<a name="line.292"></a>
-<span class="sourceLineNo">293</span>      if (!isTableStartRow(startRow)) {<a name="line.293"></a>
-<span class="sourceLineNo">294</span>        sb.append(", startRowHex=").append(Bytes.toHex(startRow));<a name="line.294"></a>
-<span class="sourceLineNo">295</span>      }<a name="line.295"></a>
-<span class="sourceLineNo">296</span>      if (!isTableEndRow(stopRow)) {<a name="line.296"></a>
-<span class="sourceLineNo">297</span>        sb.append(", stopRowHex=").append(Bytes.toHex(stopRow));<a name="line.297"></a>
-<span class="sourceLineNo">298</span>      }<a name="line.298"></a>
-<span class="sourceLineNo">299</span>      if (scanBatch &gt;= 0) {<a name="line.299"></a>
-<span class="sourceLineNo">300</span>        sb.append(", scanBatch=").append(scanBatch);<a name="line.300"></a>
-<span class="sourceLineNo">301</span>      }<a name="line.301"></a>
-<span class="sourceLineNo">302</span>      if (versions &gt;= 0) {<a name="line.302"></a>
-<span class="sourceLineNo">303</span>        sb.append(", versions=").append(versions);<a name="line.303"></a>
-<span class="sourceLineNo">304</span>      }<a name="line.304"></a>
-<span class="sourceLineNo">305</span>      if (startTime != 0) {<a name="line.305"></a>
-<span class="sourceLineNo">306</span>        sb.append("startTime=").append(startTime);<a name="line.306"></a>
-<span class="sourceLineNo">307</span>      }<a name="line.307"></a>
-<span class="sourceLineNo">308</span>      if (endTime != 0) {<a name="line.308"></a>
-<span class="sourceLineNo">309</span>        sb.append("endTime=").append(endTime);<a name="line.309"></a>
-<span class="sourceLineNo">310</span>      }<a name="line.310"></a>
-<span class="sourceLineNo">311</span>      return sb.toString();<a name="line.311"></a>
-<span class="sourceLineNo">312</span>    }<a name="line.312"></a>
-<span class="sourceLineNo">313</span><a name="line.313"></a>
-<span class="sourceLineNo">314</span>    static String getDataFileName(int hashFileIndex) {<a name="line.314"></a>
-<span class="sourceLineNo">315</span>      return String.format(HashTable.OUTPUT_DATA_FILE_PREFIX + "%05d", hashFileIndex);<a name="line.315"></a>
-<span class="sourceLineNo">316</span>    }<a name="line.316"></a>
-<span class="sourceLineNo">317</span><a name="line.317"></a>
-<span class="sourceLineNo">318</span>    /**<a name="line.318"></a>
-<span class="sourceLineNo">319</span>     * Open a TableHash.Reader starting at the first hash at or after the given key.<a name="line.319"></a>
-<span class="sourceLineNo">320</span>     * @throws IOException<a name="line.320"></a>
-<span class="sourceLineNo">321</span>     */<a name="line.321"></a>
-<span class="sourceLineNo">322</span>    public Reader newReader(Configuration conf, ImmutableBytesWritable startKey)<a name="line.322"></a>
-<span class="sourceLineNo">323</span>        throws IOException {<a name="line.323"></a>
-<span class="sourceLineNo">324</span>      return new Reader(conf, startKey);<a name="line.324"></a>
-<span class="sourceLineNo">325</span>    }<a name="line.325"></a>
-<span class="sourceLineNo">326</span><a name="line.326"></a>
-<span class="sourceLineNo">327</span>    public class Reader implements java.io.Closeable {<a name="line.327"></a>
-<span class="sourceLineNo">328</span>      private final Configuration conf;<a name="line.328"></a>
-<span class="sourceLineNo">329</span><a name="line.329"></a>
-<span class="sourceLineNo">330</span>      private int hashFileIndex;<a name="line.330"></a>
-<span class="sourceLineNo">331</span>      private MapFile.Reader mapFileReader;<a name="line.331"></a>
-<span class="sourceLineNo">332</span><a name="line.332"></a>
-<span class="sourceLineNo">333</span>      private boolean cachedNext;<a name="line.333"></a>
-<span class="sourceLineNo">334</span>      private ImmutableBytesWritable key;<a name="line.334"></a>
-<span class="sourceLineNo">335</span>      private ImmutableBytesWritable hash;<a name="line.335"></a>
-<span class="sourceLineNo">336</span><a name="line.336"></a>
-<span class="sourceLineNo">337</span>      Reader(Configuration conf, ImmutableBytesWritable startKey) throws IOException {<a name="line.337"></a>
-<span class="sourceLineNo">338</span>        this.conf = conf;<a name="line.338"></a>
-<span class="sourceLineNo">339</span>        int partitionIndex = Collections.binarySearch(partitions, startKey);<a name="line.339"></a>
-<span class="sourceLineNo">340</span>        if (partitionIndex &gt;= 0) {<a name="line.340"></a>
-<span class="sourceLineNo">341</span>          // if the key is equal to a partition, then go the file after that partition<a name="line.341"></a>
-<span class="sourceLineNo">342</span>          hashFileIndex = partitionIndex+1;<a name="line.342"></a>
-<span class="sourceLineNo">343</span>        } else {<a name="line.343"></a>
-<span class="sourceLineNo">344</span>          // if the key is between partitions, then go to the file between those partitions<a name="line.344"></a>
-<span class="sourceLineNo">345</span>          hashFileIndex = -1-partitionIndex;<a name="line.345"></a>
-<span class="sourceLineNo">346</span>        }<a name="line.346"></a>
-<span class="sourceLineNo">347</span>        openHashFile();<a name="line.347"></a>
-<span class="sourceLineNo">348</span><a name="line.348"></a>
-<span class="sourceLineNo">349</span>        // MapFile's don't make it easy to seek() so that the subsequent next() returns<a name="line.349"></a>
-<span class="sourceLineNo">350</span>        // the desired key/value pair.  So we cache it for the first call of next().<a name="line.350"></a>
-<span class="sourceLineNo">351</span>        hash = new ImmutableBytesWritable();<a name="line.351"></a>
-<span class="sourceLineNo">352</span>        key = (ImmutableBytesWritable) mapFileReader.getClosest(startKey, hash);<a name="line.352"></a>
-<span class="sourceLineNo">353</span>        if (key == null) {<a name="line.353"></a>
-<span class="sourceLineNo">354</span>          cachedNext = false;<a name="line.354"></a>
-<span class="sourceLineNo">355</span>          hash = null;<a name="line.355"></a>
-<span class="sourceLineNo">356</span>        } else {<a name="line.356"></a>
-<span class="sourceLineNo">357</span>          cachedNext = true;<a name="line.357"></a>
-<span class="sourceLineNo">358</span>        }<a name="line.358"></a>
-<span class="sourceLineNo">359</span>      }<a name="line.359"></a>
-<span class="sourceLineNo">360</span><a name="line.360"></a>
-<span class="sourceLineNo">361</span>      /**<a name="line.361"></a>
-<span class="sourceLineNo">362</span>       * Read the next key/hash pair.<a name="line.362"></a>
-<span class="sourceLineNo">363</span>       * Returns true if such a pair exists and false when at the end of the data.<a name="line.363"></a>
-<span class="sourceLineNo">364</span>       */<a name="line.364"></a>
-<span class="sourceLineNo">365</span>      public boolean next() throws IOException {<a name="line.365"></a>
-<span class="sourceLineNo">366</span>        if (cachedNext) {<a name="line.366"></a>
-<span class="sourceLineNo">367</span>          cachedNext = false;<a name="line.367"></a>
-<span class="sourceLineNo">368</span>          return true;<a name="line.368"></a>
-<span class="sourceLineNo">369</span>        }<a name="line.369"></a>
-<span class="sourceLineNo">370</span>        key = new ImmutableBytesWritable();<a name="line.370"></a>
-<span class="sourceLineNo">371</span>        hash = new ImmutableBytesWritable();<a name="line.371"></a>
-<span class="sourceLineNo">372</span>        while (true) {<a name="line.372"></a>
-<span class="sourceLineNo">373</span>          boolean hasNext = mapFileReader.next(key, hash);<a name="line.373"></a>
-<span class="sourceLineNo">374</span>          if (hasNext) {<a name="line.374"></a>
-<span class="sourceLineNo">375</span>            return true;<a name="line.375"></a>
-<span class="sourceLineNo">376</span>          }<a name="line.376"></a>
-<span class="sourceLineNo">377</span>          hashFileIndex++;<a name="line.377"></a>
-<span class="sourceLineNo">378</span>          if (hashFileIndex &lt; TableHash.this.numHashFiles) {<a name="line.378"></a>
-<span class="sourceLineNo">379</span>            mapFileReader.close();<a name="line.379"></a>
-<span class="sourceLineNo">380</span>            openHashFile();<a name="line.380"></a>
-<span class="sourceLineNo">381</span>          } else {<a name="line.381"></a>
-<span class="sourceLineNo">382</span>            key = null;<a name="line.382"></a>
-<span class="sourceLineNo">383</span>            hash = null;<a name="line.383"></a>
-<span class="sourceLineNo">384</span>            return false;<a name="line.384"></a>
-<span class="sourceLineNo">385</span>          }<a name="line.385"></a>
-<span class="sourceLineNo">386</span>        }<a name="line.386"></a>
-<span class="sourceLineNo">387</span>      }<a name="line.387"></a>
-<span class="sourceLineNo">388</span><a name="line.388"></a>
-<span class="sourceLineNo">389</span>      /**<a name="line.389"></a>
-<span class="sourceLineNo">390</span>       * Get the current key<a name="line.390"></a>
-<span class="sourceLineNo">391</span>       * @return the current key or null if there is no current key<a name="line.391"></a>
-<span class="sourceLineNo">392</span>       */<a name="line.392"></a>
-<span class="sourceLineNo">393</span>      public ImmutableBytesWritable getCurrentKey() {<a name="line.393"></a>
-<span class="sourceLineNo">394</span>        return key;<a name="line.394"></a>
-<span class="sourceLineNo">395</span>      }<a name="line.395"></a>
-<span class="sourceLineNo">396</span><a name="line.396"></a>
-<span class="sourceLineNo">397</span>      /**<a name="line.397"></a>
-<span class="sourceLineNo">398</span>       * Get the current hash<a name="line.398"></a>
-<span class="sourceLineNo">399</span>       * @return the current hash or null if there is no current hash<a name="line.399"></a>
-<span class="sourceLineNo">400</span>       */<a name="line.400"></a>
-<span class="sourceLineNo">401</span>      public ImmutableBytesWritable getCurrentHash() {<a name="line.401"></a>
-<span class="sourceLineNo">402</span>        return hash;<a name="line.402"></a>
-<span class="sourceLineNo">403</span>      }<a name="line.403"></a>
-<span class="sourceLineNo">404</span><a name="line.404"></a>
-<span class="sourceLineNo">405</span>      private void openHashFile() throws IOException {<a name="line.405"></a>
-<span class="sourceLineNo">406</span>        if (mapFileReader != null) {<a name="line.406"></a>
-<span class="sourceLineNo">407</span>          mapFileReader.close();<a name="line.407"></a>
-<span class="sourceLineNo">408</span>        }<a name="line.408"></a>
-<span class="sourceLineNo">409</span>        Path dataDir = new Path(TableHash.this.hashDir, HASH_DATA_DIR);<a name="line.409"></a>
-<span class="sourceLineNo">410</span>        Path dataFile = new Path(dataDir, getDataFileName(hashFileIndex));<a name="line.410"></a>
-<span class="sourceLineNo">411</span>        mapFileReader = new MapFile.Reader(dataFile, conf);<a name="line.411"></a>
-<span class="sourceLineNo">412</span>      }<a name="line.412"></a>
-<span class="sourceLineNo">413</span><a name="line.413"></a>
-<span class="sourceLineNo">414</span>      @Override<a name="line.414"></a>
-<span class="sourceLineNo">415</span>      public void close() throws IOException {<a name="line.415"></a>
-<span class="sourceLineNo">416</span>        mapFileReader.close();<a name="line.416"></a>
-<span class="sourceLineNo">417</span>      }<a name="line.417"></a>
-<span class="sourceLineNo">418</span>    }<a name="line.418"></a>
-<span class="sourceLineNo">419</span>  }<a name="line.419"></a>
-<span class="sourceLineNo">420</span><a name="line.420"></a>
-<span class="sourceLineNo">421</span>  static boolean isTableStartRow(byte[] row) {<a name="line.421"></a>
-<span class="sourceLineNo">422</span>    return Bytes.equals(HConstants.EMPTY_START_ROW, row);<a name="line.422"></a>
-<span class="sourceLineNo">423</span>  }<a name="line.423"></a>
-<span class="sourceLineNo">424</span><a name="line.424"></a>
-<span class="sourceLineNo">425</span>  static boolean isTableEndRow(byte[] row) {<a name="line.425"></a>
-<span class="sourceLineNo">426</span>    return Bytes.equals(HConstants.EMPTY_END_ROW, row);<a name="line.426"></a>
-<span class="sourceLineNo">427</span>  }<a name="line.427"></a>
-<span class="sourceLineNo">428</span><a name="line.428"></a>
-<span class="sourceLineNo">429</span>  public Job createSubmittableJob(String[] args) throws IOException {<a name="line.429"></a>
-<span class="sourceLineNo">430</span>    Path partitionsPath = new Path(destPath, PARTITIONS_FILE_NAME);<a name="line.430"></a>
-<span class="sourceLineNo">431</span>    generatePartitions(partitionsPath);<a name="line.431"></a>
-<span class="sourceLineNo">432</span><a name="line.432"></a>
-<span class="sourceLineNo">433</span>    Job job = Job.getInstance(getConf(),<a name="line.433"></a>
-<span class="sourceLineNo">434</span>          getConf().get("mapreduce.job.name", "hashTable_" + tableHash.tableName));<a name="line.434"></a>
-<span class="sourceLineNo">435</span>    Configuration jobConf = job.getConfiguration();<a name="line.435"></a>
-<span class="sourceLineNo">436</span>    jobConf.setLong(HASH_BATCH_SIZE_CONF_KEY, tableHash.batchSize);<a name="line.436"></a>
-<span class="sourceLineNo">437</span>    job.setJarByClass(HashTable.class);<a name="line.437"></a>
-<span class="sourceLineNo">438</span><a name="line.438"></a>
-<span class="sourceLineNo">439</span>    TableMapReduceUtil.initTableMapperJob(tableHash.tableName, tableHash.initScan(),<a name="line.439"></a>
-<span class="sourceLineNo">440</span>        HashMapper.class, ImmutableBytesWritable.class, ImmutableBytesWritable.class, job);<a name="line.440"></a>
+<span class="sourceLineNo">102</span>    List&lt;ImmutableBytesWritable&gt; partitions;<a name="line.102"></a>
+<span class="sourceLineNo">103</span><a name="line.103"></a>
+<span class="sourceLineNo">104</span>    public static TableHash read(Configuration conf, Path hashDir) throws IOException {<a name="line.104"></a>
+<span class="sourceLineNo">105</span>      TableHash tableHash = new TableHash();<a name="line.105"></a>
+<span class="sourceLineNo">106</span>      FileSystem fs = hashDir.getFileSystem(conf);<a name="line.106"></a>
+<span class="sourceLineNo">107</span>      tableHash.hashDir = hashDir;<a name="line.107"></a>
+<span class="sourceLineNo">108</span>      tableHash.readPropertiesFile(fs, new Path(hashDir, MANIFEST_FILE_NAME));<a name="line.108"></a>
+<span class="sourceLineNo">109</span>      tableHash.readPartitionFile(fs, conf, new Path(hashDir, PARTITIONS_FILE_NAME));<a name="line.109"></a>
+<span class="sourceLineNo">110</span>      return tableHash;<a name="line.110"></a>
+<span class="sourceLineNo">111</span>    }<a name="line.111"></a>
+<span class="sourceLineNo">112</span><a name="line.112"></a>
+<span class="sourceLineNo">113</span>    void writePropertiesFile(FileSystem fs, Path path) throws IOException {<a name="line.113"></a>
+<span class="sourceLineNo">114</span>      Properties p = new Properties();<a name="line.114"></a>
+<span class="sourceLineNo">115</span>      p.setProperty("table", tableName);<a name="line.115"></a>
+<span class="sourceLineNo">116</span>      if (families != null) {<a name="line.116"></a>
+<span class="sourceLineNo">117</span>        p.setProperty("columnFamilies", families);<a name="line.117"></a>
+<span class="sourceLineNo">118</span>      }<a name="line.118"></a>
+<span class="sourceLineNo">119</span>      p.setProperty("targetBatchSize", Long.toString(batchSize));<a name="line.119"></a>
+<span class="sourceLineNo">120</span>      p.setProperty("numHashFiles", Integer.toString(numHashFiles));<a name="line.120"></a>
+<span class="sourceLineNo">121</span>      if (!isTableStartRow(startRow)) {<a name="line.121"></a>
+<span class="sourceLineNo">122</span>        p.setProperty("startRowHex", Bytes.toHex(startRow));<a name="line.122"></a>
+<span class="sourceLineNo">123</span>      }<a name="line.123"></a>
+<span class="sourceLineNo">124</span>      if (!isTableEndRow(stopRow)) {<a name="line.124"></a>
+<span class="sourceLineNo">125</span>        p.setProperty("stopRowHex", Bytes.toHex(stopRow));<a name="line.125"></a>
+<span class="sourceLineNo">126</span>      }<a name="line.126"></a>
+<span class="sourceLineNo">127</span>      if (scanBatch &gt; 0) {<a name="line.127"></a>
+<span class="sourceLineNo">128</span>        p.setProperty("scanBatch", Integer.toString(scanBatch));<a name="line.128"></a>
+<span class="sourceLineNo">129</span>      }<a name="line.129"></a>
+<span class="sourceLineNo">130</span>      if (versions &gt;= 0) {<a name="line.130"></a>
+<span class="sourceLineNo">131</span>        p.setProperty("versions", Integer.toString(versions));<a name="line.131"></a>
+<span class="sourceLineNo">132</span>      }<a name="line.132"></a>
+<span class="sourceLineNo">133</span>      if (startTime != 0) {<a name="line.133"></a>
+<span class="sourceLineNo">134</span>        p.setProperty("startTimestamp", Long.toString(startTime));<a name="line.134"></a>
+<span class="sourceLineNo">135</span>      }<a name="line.135"></a>
+<span class="sourceLineNo">136</span>      if (endTime != 0) {<a name="line.136"></a>
+<span class="sourceLineNo">137</span>        p.setProperty("endTimestamp", Long.toString(endTime));<a name="line.137"></a>
+<span class="sourceLineNo">138</span>      }<a name="line.138"></a>
+<span class="sourceLineNo">139</span><a name="line.139"></a>
+<span class="sourceLineNo">140</span>      try (OutputStreamWriter osw = new OutputStreamWriter(fs.create(path), Charsets.UTF_8)) {<a name="line.140"></a>
+<span class="sourceLineNo">141</span>        p.store(osw, null);<a name="line.141"></a>
+<span class="sourceLineNo">142</span>      }<a name="line.142"></a>
+<span class="sourceLineNo">143</span>    }<a name="line.143"></a>
+<span class="sourceLineNo">144</span><a name="line.144"></a>
+<span class="sourceLineNo">145</span>    void readPropertiesFile(FileSystem fs, Path path) throws IOException {<a name="line.145"></a>
+<span class="sourceLineNo">146</span>      Properties p = new Properties();<a name="line.146"></a>
+<span class="sourceLineNo">147</span>      try (FSDataInputStream in = fs.open(path)) {<a name="line.147"></a>
+<span class="sourceLineNo">148</span>        try (InputStreamReader isr = new InputStreamReader(in, Charsets.UTF_8)) {<a name="line.148"></a>
+<span class="sourceLineNo">149</span>          p.load(isr);<a name="line.149"></a>
+<span class="sourceLineNo">150</span>        }<a name="line.150"></a>
+<span class="sourceLineNo">151</span>      }<a name="line.151"></a>
+<span class="sourceLineNo">152</span>      tableName = p.getProperty("table");<a name="line.152"></a>
+<span class="sourceLineNo">153</span>      families = p.getProperty("columnFamilies");<a name="line.153"></a>
+<span class="sourceLineNo">154</span>      batchSize = Long.parseLong(p.getProperty("targetBatchSize"));<a name="line.154"></a>
+<span class="sourceLineNo">155</span>      numHashFiles = Integer.parseInt(p.getProperty("numHashFiles"));<a name="line.155"></a>
+<span class="sourceLineNo">156</span><a name="line.156"></a>
+<span class="sourceLineNo">157</span>      String startRowHex = p.getProperty("startRowHex");<a name="line.157"></a>
+<span class="sourceLineNo">158</span>      if (startRowHex != null) {<a name="line.158"></a>
+<span class="sourceLineNo">159</span>        startRow = Bytes.fromHex(startRowHex);<a name="line.159"></a>
+<span class="sourceLineNo">160</span>      }<a name="line.160"></a>
+<span class="sourceLineNo">161</span>      String stopRowHex = p.getProperty("stopRowHex");<a name="line.161"></a>
+<span class="sourceLineNo">162</span>      if (stopRowHex != null) {<a name="line.162"></a>
+<span class="sourceLineNo">163</span>        stopRow = Bytes.fromHex(stopRowHex);<a name="line.163"></a>
+<span class="sourceLineNo">164</span>      }<a name="line.164"></a>
+<span class="sourceLineNo">165</span><a name="line.165"></a>
+<span class="sourceLineNo">166</span>      String scanBatchString = p.getProperty("scanBatch");<a name="line.166"></a>
+<span class="sourceLineNo">167</span>      if (scanBatchString != null) {<a name="line.167"></a>
+<span class="sourceLineNo">168</span>        scanBatch = Integer.parseInt(scanBatchString);<a name="line.168"></a>
+<span class="sourceLineNo">169</span>      }<a name="line.169"></a>
+<span class="sourceLineNo">170</span><a name="line.170"></a>
+<span class="sourceLineNo">171</span>      String versionString = p.getProperty("versions");<a name="line.171"></a>
+<span class="sourceLineNo">172</span>      if (versionString != null) {<a name="line.172"></a>
+<span class="sourceLineNo">173</span>        versions = Integer.parseInt(versionString);<a name="line.173"></a>
+<span class="sourceLineNo">174</span>      }<a name="line.174"></a>
+<span class="sourceLineNo">175</span><a name="line.175"></a>
+<span class="sourceLineNo">176</span>      String startTimeString = p.getProperty("startTimestamp");<a name="line.176"></a>
+<span class="sourceLineNo">177</span>      if (startTimeString != null) {<a name="line.177"></a>
+<span class="sourceLineNo">178</span>        startTime = Long.parseLong(startTimeString);<a name="line.178"></a>
+<span class="sourceLineNo">179</span>      }<a name="line.179"></a>
+<span class="sourceLineNo">180</span><a name="line.180"></a>
+<span class="sourceLineNo">181</span>      String endTimeString = p.getProperty("endTimestamp");<a name="line.181"></a>
+<span class="sourceLineNo">182</span>      if (endTimeString != null) {<a name="line.182"></a>
+<span class="sourceLineNo">183</span>        endTime = Long.parseLong(endTimeString);<a name="line.183"></a>
+<span class="sourceLineNo">184</span>      }<a name="line.184"></a>
+<span class="sourceLineNo">185</span>    }<a name="line.185"></a>
+<span class="sourceLineNo">186</span><a name="line.186"></a>
+<span class="sourceLineNo">187</span>    Scan initScan() throws IOException {<a name="line.187"></a>
+<span class="sourceLineNo">188</span>      Scan scan = new Scan();<a name="line.188"></a>
+<span class="sourceLineNo">189</span>      scan.setCacheBlocks(false);<a name="line.189"></a>
+<span class="sourceLineNo">190</span>      if (startTime != 0 || endTime != 0) {<a name="line.190"></a>
+<span class="sourceLineNo">191</span>        scan.setTimeRange(startTime, endTime == 0 ? HConstants.LATEST_TIMESTAMP : endTime);<a name="line.191"></a>
+<span class="sourceLineNo">192</span>      }<a name="line.192"></a>
+<span class="sourceLineNo">193</span>      if (scanBatch &gt; 0) {<a name="line.193"></a>
+<span class="sourceLineNo">194</span>        scan.setBatch(scanBatch);<a name="line.194"></a>
+<span class="sourceLineNo">195</span>      }<a name="line.195"></a>
+<span class="sourceLineNo">196</span>      if (versions &gt;= 0) {<a name="line.196"></a>
+<span class="sourceLineNo">197</span>        scan.readVersions(versions);<a name="line.197"></a>
+<span class="sourceLineNo">198</span>      }<a name="line.198"></a>
+<span class="sourceLineNo">199</span>      if (!isTableStartRow(startRow)) {<a name="line.199"></a>
+<span class="sourceLineNo">200</span>        scan.withStartRow(startRow);<a name="line.200"></a>
+<span class="sourceLineNo">201</span>      }<a name="line.201"></a>
+<span class="sourceLineNo">202</span>      if (!isTableEndRow(stopRow)) {<a name="line.202"></a>
+<span class="sourceLineNo">203</span>        scan.withStopRow(stopRow);<a name="line.203"></a>
+<span class="sourceLineNo">204</span>      }<a name="line.204"></a>
+<span class="sourceLineNo">205</span>      if(families != null) {<a name="line.205"></a>
+<span class="sourceLineNo">206</span>        for(String fam : families.split(",")) {<a name="line.206"></a>
+<span class="sourceLineNo">207</span>          scan.addFamily(Bytes.toBytes(fam));<a name="line.207"></a>
+<span class="sourceLineNo">208</span>        }<a name="line.208"></a>
+<span class="sourceLineNo">209</span>      }<a name="line.209"></a>
+<span class="sourceLineNo">210</span>      return scan;<a name="line.210"></a>
+<span class="sourceLineNo">211</span>    }<a name="line.211"></a>
+<span class="sourceLineNo">212</span><a name="line.212"></a>
+<span class="sourceLineNo">213</span>    /**<a name="line.213"></a>
+<span class="sourceLineNo">214</span>     * Choose partitions between row ranges to hash to a single output file<a name="line.214"></a>
+<span class="sourceLineNo">215</span>     * Selects region boundaries that fall within the scan range, and groups them<a name="line.215"></a>
+<span class="sourceLineNo">216</span>     * into the desired number of partitions.<a name="line.216"></a>
+<span class="sourceLineNo">217</span>     */<a name="line.217"></a>
+<span class="sourceLineNo">218</span>    void selectPartitions(Pair&lt;byte[][], byte[][]&gt; regionStartEndKeys) {<a name="line.218"></a>
+<span class="sourceLineNo">219</span>      List&lt;byte[]&gt; startKeys = new ArrayList&lt;&gt;();<a name="line.219"></a>
+<span class="sourceLineNo">220</span>      for (int i = 0; i &lt; regionStartEndKeys.getFirst().length; i++) {<a name="line.220"></a>
+<span class="sourceLineNo">221</span>        byte[] regionStartKey = regionStartEndKeys.getFirst()[i];<a name="line.221"></a>
+<span class="sourceLineNo">222</span>        byte[] regionEndKey = regionStartEndKeys.getSecond()[i];<a name="line.222"></a>
+<span class="sourceLineNo">223</span><a name="line.223"></a>
+<span class="sourceLineNo">224</span>        // if scan begins after this region, or starts before this region, then drop this region<a name="line.224"></a>
+<span class="sourceLineNo">225</span>        // in other words:<a name="line.225"></a>
+<span class="sourceLineNo">226</span>        //   IF (scan begins before the end of this region<a name="line.226"></a>
+<span class="sourceLineNo">227</span>        //      AND scan ends before the start of this region)<a name="line.227"></a>
+<span class="sourceLineNo">228</span>        //   THEN include this region<a name="line.228"></a>
+<span class="sourceLineNo">229</span>        if ((isTableStartRow(startRow) || isTableEndRow(regionEndKey)<a name="line.229"></a>
+<span class="sourceLineNo">230</span>            || Bytes.compareTo(startRow, regionEndKey) &lt; 0)<a name="line.230"></a>
+<span class="sourceLineNo">231</span>          &amp;&amp; (isTableEndRow(stopRow) || isTableStartRow(regionStartKey)<a name="line.231"></a>
+<span class="sourceLineNo">232</span>            || Bytes.compareTo(stopRow, regionStartKey) &gt; 0)) {<a name="line.232"></a>
+<span class="sourceLineNo">233</span>          startKeys.add(regionStartKey);<a name="line.233"></a>
+<span class="sourceLineNo">234</span>        }<a name="line.234"></a>
+<span class="sourceLineNo">235</span>      }<a name="line.235"></a>
+<span class="sourceLineNo">236</span><a name="line.236"></a>
+<span class="sourceLineNo">237</span>      int numRegions = startKeys.size();<a name="line.237"></a>
+<span class="sourceLineNo">238</span>      if (numHashFiles == 0) {<a name="line.238"></a>
+<span class="sourceLineNo">239</span>        numHashFiles = numRegions / 100;<a name="line.239"></a>
+<span class="sourceLineNo">240</span>      }<a name="line.240"></a>
+<span class="sourceLineNo">241</span>      if (numHashFiles == 0) {<a name="line.241"></a>
+<span class="sourceLineNo">242</span>        numHashFiles = 1;<a name="line.242"></a>
+<span class="sourceLineNo">243</span>      }<a name="line.243"></a>
+<span class="sourceLineNo">244</span>      if (numHashFiles &gt; numRegions) {<a name="line.244"></a>
+<span class="sourceLineNo">245</span>        // can't partition within regions<a name="line.245"></a>
+<span class="sourceLineNo">246</span>        numHashFiles = numRegions;<a name="line.246"></a>
+<span class="sourceLineNo">247</span>      }<a name="line.247"></a>
+<span class="sourceLineNo">248</span><a name="line.248"></a>
+<span class="sourceLineNo">249</span>      // choose a subset of start keys to group regions into ranges<a name="line.249"></a>
+<span class="sourceLineNo">250</span>      partitions = new ArrayList&lt;&gt;(numHashFiles - 1);<a name="line.250"></a>
+<span class="sourceLineNo">251</span>      // skip the first start key as it is not a partition between ranges.<a name="line.251"></a>
+<span class="sourceLineNo">252</span>      for (long i = 1; i &lt; numHashFiles; i++) {<a name="line.252"></a>
+<span class="sourceLineNo">253</span>        int splitIndex = (int) (numRegions * i / numHashFiles);<a name="line.253"></a>
+<span class="sourceLineNo">254</span>        partitions.add(new ImmutableBytesWritable(startKeys.get(splitIndex)));<a name="line.254"></a>
+<span class="sourceLineNo">255</span>      }<a name="line.255"></a>
+<span class="sourceLineNo">256</span>    }<a name="line.256"></a>
+<span class="sourceLineNo">257</span><a name="line.257"></a>
+<span class="sourceLineNo">258</span>    void writePartitionFile(Configuration conf, Path path) throws IOException {<a name="line.258"></a>
+<span class="sourceLineNo">259</span>      FileSystem fs = path.getFileSystem(conf);<a name="line.259"></a>
+<span class="sourceLineNo">260</span>      @SuppressWarnings("deprecation")<a name="line.260"></a>
+<span class="sourceLineNo">261</span>      SequenceFile.Writer writer = SequenceFile.createWriter(<a name="line.261"></a>
+<span class="sourceLineNo">262</span>        fs, conf, path, ImmutableBytesWritable.class, NullWritable.class);<a name="line.262"></a>
+<span class="sourceLineNo">263</span><a name="line.263"></a>
+<span class="sourceLineNo">264</span>      for (int i = 0; i &lt; partitions.size(); i++) {<a name="line.264"></a>
+<span class="sourceLineNo">265</span>        writer.append(partitions.get(i), NullWritable.get());<a name="line.265"></a>
+<span class="sourceLineNo">266</span>      }<a name="line.266"></a>
+<span class="sourceLineNo">267</span>      writer.close();<a name="line.267"></a>
+<span class="sourceLineNo">268</span>    }<a name="line.268"></a>
+<span class="sourceLineNo">269</span><a name="line.269"></a>
+<span class="sourceLineNo">270</span>    private void readPartitionFile(FileSystem fs, Configuration conf, Path path)<a name="line.270"></a>
+<span class="sourceLineNo">271</span>         throws IOException {<a name="line.271"></a>
+<span class="sourceLineNo">272</span>      @SuppressWarnings("deprecation")<a name="line.272"></a>
+<span class="sourceLineNo">273</span>      SequenceFile.Reader reader = new SequenceFile.Reader(fs, path, conf);<a name="line.273"></a>
+<span class="sourceLineNo">274</span>      ImmutableBytesWritable key = new ImmutableBytesWritable();<a name="line.274"></a>
+<span class="sourceLineNo">275</span>      partitions = new ArrayList&lt;&gt;();<a name="line.275"></a>
+<span class="sourceLineNo">276</span>      while (reader.next(key)) {<a name="line.276"></a>
+<span class="sourceLineNo">277</span>        partitions.add(new ImmutableBytesWritable(key.copyBytes()));<a name="line.277"></a>
+<span class="sourceLineNo">278</span>      }<a name="line.278"></a>
+<span class="sourceLineNo">279</span>      reader.close();<a name="line.279"></a>
+<span class="sourceLineNo">280</span><a name="line.280"></a>
+<span class="sourceLineNo">281</span>      if (!Ordering.natural().isOrdered(partitions)) {<a name="line.281"></a>
+<span class="sourceLineNo">282</span>        throw new IOException("Partitions are not ordered!");<a name="line.282"></a>
+<span class="sourceLineNo">283</span>      }<a name="line.283"></a>
+<span class="sourceLineNo">284</span>    }<a name="line.284"></a>
+<span class="sourceLineNo">285</span><a name="line.285"></a>
+<span class="sourceLineNo">286</span>    @Override<a name="line.286"></a>
+<span class="sourceLineNo">287</span>    public String toString() {<a name="line.287"></a>
+<span class="sourceLineNo">288</span>      StringBuilder sb = new StringBuilder();<a name="line.288"></a>
+<span class="sourceLineNo">289</span>      sb.append("tableName=").append(tableName);<a name="line.289"></a>
+<span class="sourceLineNo">290</span>      if (families != null) {<a name="line.290"></a>
+<span class="sourceLineNo">291</span>        sb.append(", families=").append(families);<a name="line.291"></a>
+<span class="sourceLineNo">292</span>      }<a name="line.292"></a>
+<span class="sourceLineNo">293</span>      sb.append(", batchSize=").append(batchSize);<a name="line.293"></a>
+<span class="sourceLineNo">294</span>      sb.append(", numHashFiles=").append(numHashFiles);<a name="line.294"></a>
+<span class="sourceLineNo">295</span>      if (!isTableStartRow(startRow)) {<a name="line.295"></a>
+<span class="sourceLineNo">296</span>        sb.append(", startRowHex=").append(Bytes.toHex(startRow));<a name="line.296"></a>
+<span class="sourceLineNo">297</span>      }<a name="line.297"></a>
+<span class="sourceLineNo">298</span>      if (!isTableEndRow(stopRow)) {<a name="line.298"></a>
+<span class="sourceLineNo">299</span>        sb.append(", stopRowHex=").append(Bytes.toHex(stopRow));<a name="line.299"></a>
+<span class="sourceLineNo">300</span>      }<a name="line.300"></a>
+<span class="sourceLineNo">301</span>      if (scanBatch &gt;= 0) {<a name="line.301"></a>
+<span class="sourceLineNo">302</span>        sb.append(", scanBatch=").append(scanBatch);<a name="line.302"></a>
+<span class="sourceLineNo">303</span>      }<a name="line.303"></a>
+<span class="sourceLineNo">304</span>      if (versions &gt;= 0) {<a name="line.304"></a>
+<span class="sourceLineNo">305</span>        sb.append(", versions=").append(versions);<a name="line.305"></a>
+<span class="sourceLineNo">306</span>      }<a name="line.306"></a>
+<span class="sourceLineNo">307</span>      if (startTime != 0) {<a name="line.307"></a>
+<span class="sourceLineNo">308</span>        sb.append("startTime=").append(startTime);<a name="line.308"></a>
+<span class="sourceLineNo">309</span>      }<a name="line.309"></a>
+<span class="sourceLineNo">310</span>      if (endTime != 0) {<a name="line.310"></a>
+<span class="sourceLineNo">311</span>        sb.append("endTime=").append(endTime);<a name="line.311"></a>
+<span class="sourceLineNo">312</span>      }<a name="line.312"></a>
+<span class="sourceLineNo">313</span>      return sb.toString();<a name="line.313"></a>
+<span class="sourceLineNo">314</span>    }<a name="line.314"></a>
+<span class="sourceLineNo">315</span><a name="line.315"></a>
+<span class="sourceLineNo">316</span>    static String getDataFileName(int hashFileIndex) {<a name="line.316"></a>
+<span class="sourceLineNo">317</span>      return String.format(HashTable.OUTPUT_DATA_FILE_PREFIX + "%05d", hashFileIndex);<a name="line.317"></a>
+<span class="sourceLineNo">318</span>    }<a name="line.318"></a>
+<span class="sourceLineNo">319</span><a name="line.319"></a>
+<span class="sourceLineNo">320</span>    /**<a name="line.320"></a>
+<span class="sourceLineNo">321</span>     * Open a TableHash.Reader starting at the first hash at or after the given key.<a name="line.321"></a>
+<span class="sourceLineNo">322</span>     * @throws IOException<a name="line.322"></a>
+<span class="sourceLineNo">323</span>     */<a name="line.323"></a>
+<span class="sourceLineNo">324</span>    public Reader newReader(Configuration conf, ImmutableBytesWritable startKey)<a name="line.324"></a>
+<span class="sourceLineNo">325</span>        throws IOException {<a name="line.325"></a>
+<span class="sourceLineNo">326</span>      return new Reader(conf, startKey);<a name="line.326"></a>
+<span class="sourceLineNo">327</span>    }<a name="line.327"></a>
+<span class="sourceLineNo">328</span><a name="line.328"></a>
+<span class="sourceLineNo">329</span>    public class Reader implements java.io.Closeable {<a name="line.329"></a>
+<span class="sourceLineNo">330</span>      private final Configuration conf;<a name="line.330"></a>
+<span class="sourceLineNo">331</span><a name="line.331"></a>
+<span class="sourceLineNo">332</span>      private int hashFileIndex;<a name="line.332"></a>
+<span class="sourceLineNo">333</span>      private MapFile.Reader mapFileReader;<a name="line.333"></a>
+<span class="sourceLineNo">334</span><a name="line.334"></a>
+<span class="sourceLineNo">335</span>      private boolean cachedNext;<a name="line.335"></a>
+<span class="sourceLineNo">336</span>      private ImmutableBytesWritable key;<a name="line.336"></a>
+<span class="sourceLineNo">337</span>      private ImmutableBytesWritable hash;<a name="line.337"></a>
+<span class="sourceLineNo">338</span><a name="line.338"></a>
+<span class="sourceLineNo">339</span>      Reader(Configuration conf, ImmutableBytesWritable startKey) throws IOException {<a name="line.339"></a>
+<span class="sourceLineNo">340</span>        this.conf = conf;<a name="line.340"></a>
+<span class="sourceLineNo">341</span>        int partitionIndex = Collections.binarySearch(partitions, startKey);<a name="line.341"></a>
+<span class="sourceLineNo">342</span>        if (partitionIndex &gt;= 0) {<a name="line.342"></a>
+<span class="sourceLineNo">343</span>          // if the key is equal to a partition, then go the file after that partition<a name="line.343"></a>
+<span class="sourceLineNo">344</span>          hashFileIndex = partitionIndex+1;<a name="line.344"></a>
+<span class="sourceLineNo">345</span>        } else {<a name="line.345"></a>
+<span class="sourceLineNo">346</span>          // if the key is between partitions, then go to the file between those partitions<a name="line.346"></a>
+<span class="sourceLineNo">347</span>          hashFileIndex = -1-partitionIndex;<a name="line.347"></a>
+<span class="sourceLineNo">348</span>        }<a name="line.348"></a>
+<span class="sourceLineNo">349</span>        openHashFile();<a name="line.349"></a>
+<span class="sourceLineNo">350</span><a name="line.350"></a>
+<span class="sourceLineNo">351</span>        // MapFile's don't make it easy to seek() so that the subsequent next() returns<a name="line.351"></a>
+<span class="sourceLineNo">352</span>        // the desired key/value pair.  So we cache it for the first call of next().<a name="line.352"></a>
+<span class="sourceLineNo">353</span>        hash = new ImmutableBytesWritable();<a name="line.353"></a>
+<span class="sourceLineNo">354</span>        key = (ImmutableBytesWritable) mapFileReader.getClosest(startKey, hash);<a name="line.354"></a>
+<span class="sourceLineNo">355</span>        if (key == null) {<a name="line.355"></a>
+<span class="sourceLineNo">356</span>          cachedNext = false;<a name="line.356"></a>
+<span class="sourceLineNo">357</span>          hash = null;<a name="line.357"></a>
+<span class="sourceLineNo">358</span>        } else {<a name="line.358"></a>
+<span class="sourceLineNo">359</span>          cachedNext = true;<a name="line.359"></a>
+<span class="sourceLineNo">360</span>        }<a name="line.360"></a>
+<span class="sourceLineNo">361</span>      }<a name="line.361"></a>
+<span class="sourceLineNo">362</span><a name="line.362"></a>
+<span class="sourceLineNo">363</span>      /**<a name="line.363"></a>
+<span class="sourceLineNo">364</span>       * Read the next key/hash pair.<a name="line.364"></a>
+<span class="sourceLineNo">365</span>       * Returns true if such a pair exists and false when at the end of the data.<a name="line.365"></a>
+<span class="sourceLineNo">366</span>       */<a name="line.366"></a>
+<span class="sourceLineNo">367</span>      public boolean next() throws IOException {<a name="line.367"></a>
+<span class="sourceLineNo">368</span>        if (cachedNext) {<a name="line.368"></a>
+<span class="sourceLineNo">369</span>          cachedNext = false;<a name="line.369"></a>
+<span class="sourceLineNo">370</span>          return true;<a name="line.370"></a>
+<span class="sourceLineNo">371</span>        }<a name="line.371"></a>
+<span class="sourceLineNo">372</span>        key = new ImmutableBytesWritable();<a name="line.372"></a>
+<span class="sourceLineNo">373</span>        hash = new ImmutableBytesWritable();<a name="line.373"></a>
+<span class="sourceLineNo">374</span>        while (true) {<a name="line.374"></a>
+<span class="sourceLineNo">375</span>          boolean hasNext = mapFileReader.next(key, hash);<a name="line.375"></a>
+<span class="sourceLineNo">376</span>          if (hasNext) {<a name="line.376"></a>
+<span class="sourceLineNo">377</span>            return true;<a name="line.377"></a>
+<span class="sourceLineNo">378</span>          }<a name="line.378"></a>
+<span class="sourceLineNo">379</span>          hashFileIndex++;<a name="line.379"></a>
+<span class="sourceLineNo">380</span>          if (hashFileIndex &lt; TableHash.this.numHashFiles) {<a name="line.380"></a>
+<span class="sourceLineNo">381</span>            mapFileReader.close();<a name="line.381"></a>
+<span class="sourceLineNo">382</span>            openHashFile();<a name="line.382"></a>
+<span class="sourceLineNo">383</span>          } else {<a name="line.383"></a>
+<span class="sourceLineNo">384</span>            key = null;<a name="line.384"></a>
+<span class="sourceLineNo">385</span>            hash = null;<a name="line.385"></a>
+<span class="sourceLineNo">386</span>            return false;<a name="line.386"></a>
+<span class="sourceLineNo">387</span>          }<a name="line.387"></a>
+<span class="sourceLineNo">388</span>        }<a name="line.388"></a>
+<span class="sourceLineNo">389</span>      }<a name="line.389"></a>
+<span class="sourceLineNo">390</span><a name="line.390"></a>
+<span class="sourceLineNo">391</span>      /**<a name="line.391"></a>
+<span class="sourceLineNo">392</span>       * Get the current key<a name="line.392"></a>
+<span class="sourceLineNo">393</span>       * @return the current key or null if there is no current key<a name="line.393"></a>
+<span class="sourceLineNo">394</span>       */<a name="line.394"></a>
+<span class="sourceLineNo">395</span>      public ImmutableBytesWritable getCurrentKey() {<a name="line.395"></a>
+<span class="sourceLineNo">396</span>        return key;<a name="line.396"></a>
+<span class="sourceLineNo">397</span>      }<a name="line.397"></a>
+<span class="sourceLineNo">398</span><a name="line.398"></a>
+<span class="sourceLineNo">399</span>      /**<a name="line.399"></a>
+<span class="sourceLineNo">400</span>       * Get the current hash<a name="line.400"></a>
+<span class="sourceLineNo">401</span>       * @return the current hash or null if there is no current hash<a name="line.401"></a>
+<span class="sourceLineNo">402</span>       */<a name="line.402"></a>
+<span class="sourceLineNo">403</span>      public ImmutableBytesWritable getCurrentHash() {<a name="line.403"></a>
+<span class="sourceLineNo">404</span>        return hash;<a name="line.404"></a>
+<span class="sourceLineNo">405</span>      }<a name="line.405"></a>
+<span class="sourceLineNo">406</span><a name="line.406"></a>
+<span class="sourceLineNo">407</span>      private void openHashFile() throws IOException {<a name="line.407"></a>
+<span class="sourceLineNo">408</span>        if (mapFileReader != null) {<a name="line.408"></a>
+<span class="sourceLineNo">409</span>          mapFileReader.close();<a name="line.409"></a>
+<span class="sourceLineNo">410</span>        }<a name="line.410"></a>
+<span class="sourceLineNo">411</span>        Path dataDir = new Path(TableHash.this.hashDir, HASH_DATA_DIR);<a name="line.411"></a>
+<span class="sourceLineNo">412</span>        Path dataFile = new Path(dataDir, getDataFileName(hashFileIndex));<a name="line.412"></a>
+<span class="sourceLineNo">413</span>        mapFileReader = new MapFile.Reader(dataFile, conf);<a name="line.413"></a>
+<span class="sourceLineNo">414</span>      }<a name="line.414"></a>
+<span class="sourceLineNo">415</span><a name="line.415"></a>
+<span class="sourceLineNo">416</span>      @Override<a name="line.416"></a>
+<span class="sourceLineNo">417</span>      public void close() throws IOException {<a name="line.417"></a>
+<span class="sourceLineNo">418</span>        mapFileReader.close();<a name="line.418"></a>
+<span class="sourceLineNo">419</span>      }<a name="line.419"></a>
+<span class="sourceLineNo">420</span>    }<a name="line.420"></a>
+<span class="sourceLineNo">421</span>  }<a name="line.421"></a>
+<span class="sourceLineNo">422</span><a name="line.422"></a>
+<span class="sourceLineNo">423</span>  static boolean isTableStartRow(byte[] row) {<a name="line.423"></a>
+<span class="sourceLineNo">424</span>    return Bytes.equals(HConstants.EMPTY_START_ROW, row);<a name="line.424"></a>
+<span class="sourceLineNo">425</span>  }<a name="line.425"></a>
+<span class="sourceLineNo">426</span><a name="line.426"></a>
+<span class="sourceLineNo">427</span>  static boolean isTableEndRow(byte[] row) {<a name="line.427"></a>
+<span class="sourceLineNo">428</span>    return Bytes.equals(HConstants.EMPTY_END_ROW, row);<a name="line.428"></a>
+<span class="sourceLineNo">429</span>  }<a name="line.429"></a>
+<span class="sourceLineNo">430</span><a name="line.430"></a>
+<span class="sourceLineNo">431</span>  public Job createSubmittableJob(String[] args) throws IOException {<a name="line.431"></a>
+<span class="sourceLineNo">432</span>    Path partitionsPath = new Path(destPath, PARTITIONS_FILE_NAME);<a name="line.432"></a>
+<span class="sourceLineNo">433</span>    generatePartitions(partitionsPath);<a name="line.433"></a>
+<span class="sourceLineNo">434</span><a name="line.434"></a>
+<span class="sourceLineNo">435</span>    Job job = Job.getInstance(getConf(),<a name="line.435"></a>
+<span class="sourceLineNo">436</span>          getConf().get("mapreduce.job.name", "hashTable_" + tableHash.tableName));<a name="line.436"></a>
+<span class="sourceLineNo">437</span>    Configuration jobConf = job.getConfiguration();<a name="line.437"></a>
+<span class="sourceLineNo">438</span>    jobConf.setLong(HASH_BATCH_SIZE_CONF_KEY, tableHash.batchSize);<a name="line.438"></a>
+<span class="sourceLineNo">439</span>    jobConf.setBoolean(IGNORE_TIMESTAMPS, tableHash.ignoreTimestamps);<a name="line.439"></a>
+<span class="sourceLineNo">440</span>    job.setJarByClass(HashTable.class);<a name="line.440"></a>
 <span class="sourceLineNo">441</span><a name="line.441"></a>
-<span class="sourceLineNo">442</span>    // use a TotalOrderPartitioner and reducers to group region output into hash files<a name="line.442"></a>
-<span class="sourceLineNo">443</span>    job.setPartitionerClass(TotalOrderPartitioner.class);<a name="line.443"></a>
-<span class="sourceLineNo">444</span>    TotalOrderPartitioner.setPartitionFile(jobConf, partitionsPath);<a name="line.444"></a>
-<span class="sourceLineNo">445</span>    job.setReducerClass(Reducer.class);  // identity reducer<a name="line.445"></a>
-<span class="sourceLineNo">446</span>    job.setNumReduceTasks(tableHash.numHashFiles);<a name="line.446"></a>
-<span class="sourceLineNo">447</span>    job.setOutputKeyClass(ImmutableBytesWritable.class);<a name="line.447"></a>
-<span class="sourceLineNo">448</span>    job.setOutputValueClass(ImmutableBytesWritable.class);<a name="line.448"></a>
-<span class="sourceLineNo">449</span>    job.setOutputFormatClass(MapFileOutputFormat.class);<a name="line.449"></a>
-<span class="sourceLineNo">450</span>    FileOutputFormat.setOutputPath(job, new Path(destPath, HASH_DATA_DIR));<a name="line.450"></a>
-<span class="sourceLineNo">451</span><a name="line.451"></a>
-<span class="sourceLineNo">452</span>    return job;<a name="line.452"></a>
-<span class="sourceLineNo">453</span>  }<a name="line.453"></a>
+<span class="sourceLineNo">442</span>    TableMapReduceUtil.initTableMapperJob(tableHash.tableName, tableHash.initScan(),<a name="line.442"></a>
+<span class="sourceLineNo">443</span>        HashMapper.class, ImmutableBytesWritable.class, ImmutableBytesWritable.class, job);<a name="line.443"></a>
+<span class="sourceLineNo">444</span><a name="line.444"></a>
+<span class="sourceLineNo">445</span>    // use a TotalOrderPartitioner and reducers to group region output into hash files<a name="line.445"></a>
+<span class="sourceLineNo">446</span>    job.setPartitionerClass(TotalOrderPartitioner.class);<a name="line.446"></a>
+<span class="sourceLineNo">447</span>    TotalOrderPartitioner.setPartitionFile(jobConf, partitionsPath);<a name="line.447"></a>
+<span class="sourceLineNo">448</span>    job.setReducerClass(Reducer.class);  // identity reducer<a name="line.448"></a>
+<span class="sourceLineNo">449</span>    job.setNumReduceTasks(tableHash.numHashFiles);<a name="line.449"></a>
+<span class="sourceLineNo">450</span>    job.setOutputKeyClass(ImmutableBytesWritable.class);<a name="line.450"></a>
+<span class="sourceLineNo">451</span>    job.setOutputValueClass(ImmutableBytesWritable.class);<a name="line.451"></a>
+<span class="sourceLineNo">452</span>    job.setOutputFormatClass(MapFileOutputFormat.class);<a name="line.452"></a>
+<span class="sourceLineNo">453</span>    FileOutputFormat.setOutputPath(job, new Path(destPath, HASH_DATA_DIR));<a name="line.453"></a>
 <span class="sourceLineNo">454</span><a name="line.454"></a>
-<span class="sourceLineNo">455</span>  private void generatePartitions(Path partitionsPath) throws IOException {<a name="line.455"></a>
-<span class="sourceLineNo">456</span>    Connection connection = ConnectionFactory.createConnection(getConf());<a name="line.456"></a>
-<span class="sourceLineNo">457</span>    Pair&lt;byte[][], byte[][]&gt; regionKeys<a name="line.457"></a>
-<span class="sourceLineNo">458</span>      = connection.getRegionLocator(TableName.valueOf(tableHash.tableName)).getStartEndKeys();<a name="line.458"></a>
-<span class="sourceLineNo">459</span>    connection.close();<a name="line.459"></a>
-<span class="sourceLineNo">460</span><a name="line.460"></a>
-<span class="sourceLineNo">461</span>    tableHash.selectPartitions(regionKeys);<a name="line.461"></a>
-<span class="sourceLineNo">462</span>    LOG.info("Writing " + tableHash.partitions.size() + " partition keys to " + partitionsPath);<a name="line.462"></a>
+<span class="sourceLineNo">455</span>    return job;<a name="line.455"></a>
+<span class="sourceLineNo">456</span>  }<a name="line.456"></a>
+<span class="sourceLineNo">457</span><a name="line.457"></a>
+<span class="sourceLineNo">458</span>  private void generatePartitions(Path partitionsPath) throws IOException {<a name="line.458"></a>
+<span class="sourceLineNo">459</span>    Connection connection = ConnectionFactory.createConnection(getConf());<a name="line.459"></a>
+<span class="sourceLineNo">460</span>    Pair&lt;byte[][], byte[][]&gt; regionKeys<a name="line.460"></a>
+<span class="sourceLineNo">461</span>      = connection.getRegionLocator(TableName.valueOf(tableHash.tableName)).getStartEndKeys();<a name="line.461"></a>
+<span class="sourceLineNo">462</span>    connection.close();<a name="line.462"></a>
 <span class="sourceLineNo">463</span><a name="line.463"></a>
-<span class="sourceLineNo">464</span>    tableHash.writePartitionFile(getConf(), partitionsPath);<a name="line.464"></a>
-<span class="sourceLineNo">465</span>  }<a name="line.465"></a>
+<span class="sourceLineNo">464</span>    tableHash.selectPartitions(regionKeys);<a name="line.464"></a>
+<span class="sourceLineNo">465</span>    LOG.info("Writing " + tableHash.partitions.size() + " partition keys to " + partitionsPath);<a name="line.465"></a>
 <span class="sourceLineNo">466</span><a name="line.466"></a>
-<span class="sourceLineNo">467</span>  static class ResultHasher {<a name="line.467"></a>
-<span class="sourceLineNo">468</span>    private MessageDigest digest;<a name="line.468"></a>
+<span class="sourceLineNo">467</span>    tableHash.writePartitionFile(getConf(), partitionsPath);<a name="line.467"></a>
+<span class="sourceLineNo">468</span>  }<a name="line.468"></a>
 <span class="sourceLineNo">469</span><a name="line.469"></a>
-<span class="sourceLineNo">470</span>    private boolean batchStarted = false;<a name="line.470"></a>
-<span class="sourceLineNo">471</span>    private ImmutableBytesWritable batchStartKey;<a name="line.471"></a>
-<span class="sourceLineNo">472</span>    private ImmutableBytesWritable batchHash;<a name="line.472"></a>
-<span class="sourceLineNo">473</span>    private long batchSize = 0;<a name="line.473"></a>
-<span class="sourceLineNo">474</span><a name="line.474"></a>
-<span class="sourceLineNo">475</span><a name="line.475"></a>
-<span class="sourceLineNo">476</span>    public ResultHasher() {<a name="line.476"></a>
-<span class="sourceLineNo">477</span>      try {<a name="line.477"></a>
-<span class="sourceLineNo">478</span>        digest = MessageDigest.getInstance("MD5");<a name="line.478"></a>
-<span class="sourceLineNo">479</span>      } catch (NoSuchAlgorithmException e) {<a name="line.479"></a>
-<span class="sourceLineNo">480</span>        Throwables.propagate(e);<a name="line.480"></a>
-<span class="sourceLineNo">481</span>      }<a name="line.481"></a>
-<span class="sourceLineNo">482</span>    }<a name="line.482"></a>
-<span class="sourceLineNo">483</span><a name="line.483"></a>
-<span class="sourceLineNo">484</span>    public void startBatch(ImmutableBytesWritable row) {<a name="line.484"></a>
-<span class="sourceLineNo">485</span>      if (batchStarted) {<a name="line.485"></a>
-<span class="sourceLineNo">486</span>        throw new RuntimeException("Cannot start new batch without finishing existing one.");<a name="line.486"></a>
-<span class="sourceLineNo">487</span>      }<a name="line.487"></a>
-<span class="sourceLineNo">488</span>      batchStarted = true;<a name="line.488"></a>
-<span class="sourceLineNo">489</span>      batchSize = 0;<a name="line.489"></a>
-<span class="sourceLineNo">490</span>      batchStartKey = row;<a name="line.490"></a>
-<span class="sourceLineNo">491</span>      batchHash = null;<a name="line.491"></a>
-<span class="sourceLineNo">492</span>    }<a name="line.492"></a>
-<span class="sourceLineNo">493</span><a name="line.493"></a>
-<span class="sourceLineNo">494</span>    public void hashResult(Result result) {<a name="line.494"></a>
-<span class="sourceLineNo">495</span>      if (!batchStarted) {<a name="line.495"></a>
-<span class="sourceLineNo">496</span>        throw new RuntimeException("Cannot add to batch that has not been started.");<a name="line.496"></a>
-<span class="sourceLineNo">497</span>      }<a name="line.497"></a>
-<span class="sourceLineNo">498</span>      for (Cell cell : result.rawCells()) {<a name="line.498"></a>
-<span class="sourceLineNo">499</span>        int rowLength = cell.getRowLength();<a name="line.499"></a>
-<span class="sourceLineNo">500</span>        int familyLength = cell.getFamilyLength();<a name="line.500"></a>
-<span class="sourceLineNo">501</span>        int qualifierLength = cell.getQualifierLength();<a name="line.501"></a>
-<span class="sourceLineNo">502</span>        int valueLength = cell.getValueLength();<a name="line.502"></a>
-<span class="sourceLineNo">503</span>        digest.update(cell.getRowArray(), cell.getRowOffset(), rowLength);<a name="line.503"></a>
-<span class="sourceLineNo">504</span>        digest.update(cell.getFamilyArray(), cell.getFamilyOffset(), familyLength);<a name="line.504"></a>
-<span class="sourceLineNo">505</span>        digest.update(cell.getQualifierArray(), cell.getQualifierOffset(), qualifierLength);<a name="line.505"></a>
-<span class="sourceLineNo">506</span>        long ts = cell.getTimestamp();<a name="line.506"></a>
-<span class="sourceLineNo">507</span>        for (int i = 8; i &gt; 0; i--) {<a name="line.507"></a>
-<span class="sourceLineNo">508</span>          digest.update((byte) ts);<a name="line.508"></a>
-<span class="sourceLineNo">509</span>          ts &gt;&gt;&gt;= 8;<a name="line.509"></a>
-<span class="sourceLineNo">510</span>        }<a name="line.510"></a>
-<span class="sourceLineNo">511</span>        digest.update(cell.getValueArray(), cell.getValueOffset(), valueLength);<a name="line.511"></a>
-<span class="sourceLineNo">512</span><a name="line.512"></a>
-<span class="sourceLineNo">513</span>        batchSize += rowLength + familyLength + qualifierLength + 8 + valueLength;<a name="line.513"></a>
-<span class="sourceLineNo">514</span>      }<a name="line.514"></a>
-<span class="sourceLineNo">515</span>    }<a name="line.515"></a>
-<span class="sourceLineNo">516</span><a name="line.516"></a>
-<span class="sourceLineNo">517</span>    public void finishBatch() {<a name="line.517"></a>
-<span class="sourceLineNo">518</span>      if (!batchStarted) {<a name="line.518"></a>
-<span class="sourceLineNo">519</span>        throw new RuntimeException("Cannot finish batch that has not started.");<a name="line.519"></a>
-<span class="sourceLineNo">520</span>      }<a name="line.520"></a>
-<span class="sourceLineNo">521</span>      batchStarted = false;<a name="line.521"></a>
-<span class="sourceLineNo">522</span>      batchHash = new ImmutableBytesWritable(digest.digest());<a name="line.522"></a>
-<span class="sourceLineNo">523</span>    }<a name="line.523"></a>
-<span class="sourceLineNo">524</span><a name="line.524"></a>
-<span class="sourceLineNo">525</span>    public boolean isBatchStarted() {<a name="line.525"></a>
-<span class="sourceLineNo">526</span>      return batchStarted;<a name="line.526"></a>
-<span class="sourceLineNo">527</span>    }<a name="line.527"></a>
-<span class="sourceLineNo">528</span><a name="line.528"></a>
-<span class="sourceLineNo">529</span>    public ImmutableBytesWritable getBatchStartKey() {<a name="line.529"></a>
-<span class="sourceLineNo">530</span>      return batchStartKey;<a name="line.530"></a>
-<span class="sourceLineNo">531</span>    }<a name="line.531"></a>
-<span class="sourceLineNo">532</span><a name="line.532"></a>
-<span class="sourceLineNo">533</span>    public ImmutableBytesWritable getBatchHash() {<a name="line.533"></a>
-<span class="sourceLineNo">534</span>      return batchHash;<a name="line.534"></a>
-<span class="sourceLineNo">535</span>    }<a name="line.535"></a>
-<span class="sourceLineNo">536</span><a name="line.536"></a>
-<span class="sourceLineNo">537</span>    public long getBatchSize() {<a name="line.537"></a>
-<span class="sourceLineNo">538</span>      return batchSize;<a name="line.538"></a>
-<span class="sourceLineNo">539</span>    }<a name="line.539"></a>
-<span class="sourceLineNo">540</span>  }<a name="line.540"></a>
-<span class="sourceLineNo">541</span><a name="line.541"></a>
-<span class="sourceLineNo">542</span>  public static class HashMapper<a name="line.542"></a>
-<span class="sourceLineNo">543</span>    extends TableMapper&lt;ImmutableBytesWritable, ImmutableBytesWritable&gt; {<a name="line.543"></a>
-<span class="sourceLineNo">544</span><a name="line.544"></a>
-<span class="sourceLineNo">545</span>    private ResultHasher hasher;<a name="line.545"></a>
-<span class="sourceLineNo">546</span>    private long targetBatchSize;<a name="line.546"></a>
-<span class="sourceLineNo">547</span><a name="line.547"></a>
-<span class="sourceLineNo">548</span>    private ImmutableBytesWritable currentRow;<a name="line.548"></a>
-<span class="sourceLineNo">549</span><a name="line.549"></a>
-<span class="sourceLineNo">550</span>    @Override<a name="line.550"></a>
-<span class="sourceLineNo">551</span>    protected void setup(Context context) throws IOException, InterruptedException {<a name="line.551"></a>
-<span class="sourceLineNo">552</span>      targetBatchSize = context.getConfiguration()<a name="line.552"></a>
-<span class="sourceLineNo">553</span>          .getLong(HASH_BATCH_SIZE_CONF_KEY, DEFAULT_BATCH_SIZE);<a name="line.553"></a>
-<span class="sourceLineNo">554</span>      hasher = new ResultHasher();<a name="line.554"></a>
-<span class="sourceLineNo">555</span><a name="line.555"></a>
-<span class="sourceLineNo">556</span>      TableSplit split = (TableSplit) context.getInputSplit();<a name="line.556"></a>
-<span class="sourceLineNo">557</span>      hasher.startBatch(new ImmutableBytesWritable(split.getStartRow()));<a name="line.557"></a>
-<span class="sourceLineNo">558</span>    }<a name="line.558"></a>
-<span class="sourceLineNo">559</span><a name="line.559"></a>
-<span class="sourceLineNo">560</span>    @Override<a name="line.560"></a>
-<span class="sourceLineNo">561</span>    protected void map(ImmutableBytesWritable key, Result value, Context context)<a name="line.561"></a>
-<span class="sourceLineNo">562</span>        throws IOException, InterruptedException {<a name="line.562"></a>
-<span class="sourceLineNo">563</span><a name="line.563"></a>
-<span class="sourceLineNo">564</span>      if (currentRow == null || !currentRow.equals(key)) {<a name="line.564"></a>
-<span class="sourceLineNo">565</span>        currentRow = new ImmutableBytesWritable(key); // not immutable<a name="line.565"></a>
-<span class="sourceLineNo">566</span><a name="line.566"></a>
-<span class="sourceLineNo">567</span>        if (hasher.getBatchSize() &gt;= targetBatchSize) {<a name="line.567"></a>
-<span class="sourceLineNo">568</span>          hasher.finishBatch();<a name="line.568"></a>
-<span class="sourceLineNo">569</span>          context.write(hasher.getBatchStartKey(), hasher.getBatchHash());<a name="line.569"></a>
-<span class="sourceLineNo">570</span>          hasher.startBatch(currentRow);<a name="line.570"></a>
-<span class="sourceLineNo">571</span>        }<a name="line.571"></a>
-<span class="sourceLineNo">572</span>      }<a name="line.572"></a>
-<span class="sourceLineNo">573</span><a name="line.573"></a>
-<span class="sourceLineNo">574</span>      hasher.hashResult(value);<a name="line.574"></a>
-<span class="sourceLineNo">575</span>    }<a name="line.575"></a>
-<span class="sourceLineNo">576</span><a name="line.576"></a>
-<span class="sourceLineNo">577</span>    @Override<a name="line.577"></a>
-<span class="sourceLineNo">578</span>    protected void cleanup(Context context) throws IOException, InterruptedException {<a name="line.578"></a>
-<span class="sourceLineNo">579</span>      hasher.finishBatch();<a name="line.579"></a>
-<span class="sourceLineNo">580</span>      context.write(hasher.getBatchStartKey(), hasher.getBatchHash());<a name="line.580"></a>
-<span class="sourceLineNo">581</span>    }<a name="line.581"></a>
-<span class="sourceLineNo">582</span>  }<a name="line.582"></a>
-<span class="sourceLineNo">583</span><a name="line.583"></a>
-<span class="sourceLineNo">584</span>  private void writeTempManifestFile() throws IOException {<a name="line.584"></a>
-<span class="sourceLineNo">585</span>    Path tempManifestPath = new Path(destPath, TMP_MANIFEST_FILE_NAME);<a name="line.585"></a>
-<span class="sourceLineNo">586</span>    FileSystem fs = tempManifestPath.getFileSystem(getConf());<a name="line.586"></a>
-<span class="sourceLineNo">587</span>    tableHash.writePropertiesFile(fs, tempManifestPath);<a name="line.587"></a>
-<span class="sourceLineNo">588</span>  }<a name="line.588"></a>
-<span class="sourceLineNo">589</span><a name="line.589"></a>
-<span class="sourceLineNo">590</span>  private void completeManifest() throws IOException {<a name="line.590"></a>
-<span class="sourceLineNo">591</span>    Path tempManifestPath = new Path(destPath, TMP_MANIFEST_FILE_NAME);<a name="line.591"></a>
-<span class="sourceLineNo">592</span>    Path manifestPath = new Path(destPath, MANIFEST_FILE_NAME);<a name="line.592"></a>
-<span class="sourceLineNo">593</span>    FileSystem fs = tempManifestPath.getFileSystem(getConf());<a name="line.593"></a>
-<span class="sourceLineNo">594</span>    fs.rename(tempManifestPath, manifestPath);<a name="line.594"></a>
-<span class="sourceLineNo">595</span>  }<a name="line.595"></a>
-<span class="sourceLineNo">596</span><a name="line.596"></a>
-<span class="sourceLineNo">597</span>  private static final int NUM_ARGS = 2;<a name="line.597"></a>
-<span class="sourceLineNo">598</span>  private static void printUsage(final String errorMsg) {<a name="line.598"></a>
-<span class="sourceLineNo">599</span>    if (errorMsg != null &amp;&amp; errorMsg.length() &gt; 0) {<a name="line.599"></a>
-<span class="sourceLineNo">600</span>      System.err.println("ERROR: " + errorMsg);<a name="line.600"></a>
-<span class="sourceLineNo">601</span>      System.err.println();<a name="line.601"></a>
-<span class="sourceLineNo">602</span>    }<a name="line.602"></a>
-<span class="sourceLineNo">603</span>    System.err.println("Usage: HashTable [options] &lt;tablename&gt; &lt;outputpath&gt;");<a name="line.603"></a>
-<span class="sourceLineNo">604</span>    System.err.println();<a name="line.604"></a>
-<span class="sourceLineNo">605</span>    System.err.println("Options:");<a name="line.605"></a>
-<span class="sourceLineNo">606</span>    System.err.println(" batchsize     the target amount of bytes to hash in each batch");<a name="line.606"></a>
-<span class="sourceLineNo">607</span>    System.err.println("               rows are added to the batch until this size is reached");<a name="line.607"></a>
-<span class="sourceLineNo">608</span>    System.err.println("               (defaults to " + DEFAULT_BATCH_SIZE + " bytes)");<a name="line.608"></a>
-<span class="sourceLineNo">609</span>    System.err.println(" numhashfiles  the number of hash files to create");<a name="line.609"></a>
-<span class="sourceLineNo">610</span>    System.err.println("               if set to fewer than number of regions then");<a name="line.610"></a>
-<span class="sourceLineNo">611</span>    System.err.println("               the job will create this number of reducers");<a name="line.611"></a>
-<span class="sourceLineNo">612</span>    System.err.println("               (defaults to 1/100 of regions -- at least 1)");<a name="line.612"></a>
-<span class="sourceLineNo">613</span>    System.err.println(" startrow      the start row");<a name="line.613"></a>
-<span class="sourceLineNo">614</span>    System.err.println(" stoprow       the stop row");<a name="line.614"></a>
-<span class="sourceLineNo">615</span>    System.err.println(" starttime     beginning of the time range (unixtime in millis)");<a name="line.615"></a>
-<span class="sourceLineNo">616</span>    System.err.println("               without endtime means from starttime to forever");<a name="line.616"></a>
-<span class="sourceLineNo">617</span>    System.err.println(" endtime       end of the time range.  Ignored if no starttime specified.");<a name="line.617"></a>
-<span class="sourceLineNo">618</span>    System.err.println(" scanbatch     scanner batch size to support intra row scans");<a name="line.618"></a>
-<span class="sourceLineNo">619</span>    System.err.println(" versions      number of cell versions to include");<a name="line.619"></a>
-<span class="sourceLineNo">620</span>    System.err.println(" families      comma-separated list of families to include");<a name="line.620"></a>
-<span class="sourceLineNo">621</span>    System.err.println();<a name="line.621"></a>
-<span class="sourceLineNo">622</span>    System.err.println("Args:");<a name="line.622"></a>
-<span class="sourceLineNo">623</span>    System.err.println(" tablename     Name of the table to hash");<a name="line.623"></a>
-<span class="sourceLineNo">624</span>    System.err.println(" outputpath    Filesystem path to put the output data");<a name="line.624"></a>
-<span class="sourceLineNo">625</span>    System.err.println();<a name="line.625"></a>
-<span class="sourceLineNo">626</span>    System.err.println("Examples:");<a name="line.626"></a>
-<span class="sourceLineNo">627</span>    System.err.println(" To hash 'TestTable' in 32kB batches for a 1 hour window into 50 files:");<a name="line.627"></a>
-<span class="sourceLineNo">628</span>    System.err.println(" $ hbase " +<a name="line.628"></a>
-<span class="sourceLineNo">629</span>        "org.apache.hadoop.hbase.mapreduce.HashTable --batchsize=32000 --numhashfiles=50"<a name="line.629"></a>
-<span class="sourceLineNo">630</span>        + " --starttime=1265875194289 --endtime=1265878794289 --families=cf2,cf3"<a name="line.630"></a>
-<span class="sourceLineNo">631</span>        + " TestTable /hashes/testTable");<a name="line.631"></a>
-<span class="sourceLineNo">632</span>  }<a name="line.632"></a>
-<span class="sourceLineNo">633</span><a name="line.633"></a>
-<span class="sourceLineNo">634</span>  private boolean doCommandLine(final String[] args) {<a name="line.634"></a>
-<span class="sourceLineNo">635</span>    if (args.length &lt; NUM_ARGS) {<a name="line.635"></a>
-<span class="sourceLineNo">636</span>      printUsage(null);<a name="line.636"></a>
-<span class="sourceLineNo">637</span>      return false;<a name="line.637"></a>
-<span class="sourceLineNo">638</span>    }<a name="line.638"></a>
-<span class="sourceLineNo">639</span>    try {<a name="line.639"></a>
-<span class="sourceLineNo">640</span><a name="line.640"></a>
-<span class="sourceLineNo">641</span>      tableHash.tableName = args[args.length-2];<a name="line.641"></a>
-<span class="sourceLineNo">642</span>      destPath = new Path(args[args.length-1]);<a name="line.642"></a>
-<span class="sourceLineNo">643</span><a name="line.643"></a>
-<span class="sourceLineNo">644</span>      for (int i = 0; i &lt; args.length - NUM_ARGS; i++) {<a name="line.644"></a>
-<span class="sourceLineNo">645</span>        String cmd = args[i];<a name="line.645"></a>
-<span class="sourceLineNo">646</span>        if (cmd.equals("-h") || cmd.startsWith("--h")) {<a name="line.646"></a>
-<span class="sourceLineNo">647</span>          printUsage(null);<a name="line.647"></a>
-<span class="sourceLineNo">648</span>          return false;<a name="line.648"></a>
-<span class="sourceLineNo">649</span>        }<a name="line.649"></a>
-<span class="sourceLineNo">650</span><a name="line.650"></a>
-<span class="sourceLineNo">651</span>        final String batchSizeArgKey = "--batchsize=";<a name="line.651"></a>
-<span class="sourceLineNo">652</span>        if (cmd.startsWith(batchSizeArgKey)) {<a name="line.652"></a>
-<span class="sourceLineNo">653</span>          tableHash.batchSize = Long.parseLong(cmd.substring(batchSizeArgKey.length()));<a name="line.653"></a>
-<span class="sourceLineNo">654</span>          continue;<a name="line.654"></a>
-<span class="sourceLineNo">655</span>        }<a name="line.655"></a>
-<span class="sourceLineNo">656</span><a name="line.656"></a>
-<span class="sourceLineNo">657</span>        final String numHashFilesArgKey = "--numhashfiles=";<a name="line.657"></a>
-<span class="sourceLineNo">658</span>        if (cmd.startsWith(numHashFilesArgKey)) {<a name="line.658"></a>
-<span class="sourceLineNo">659</span>          tableHash.numHashFiles = Integer.parseInt(cmd.substring(numHashFilesArgKey.length()));<a name="line.659"></a>
-<span class="sourceLineNo">660</span>          continue;<a name="line.660"></a>
-<span class="sourceLineNo">661</span>        }<a name="line.661"></a>
-<span class="sourceLineNo">662</span><a name="line.662"></a>
-<span class="sourceLineNo">663</span>        final String startRowArgKey = "--startrow=";<a name="line.663"></a>
-<span class="sourceLineNo">664</span>        if (cmd.startsWith(startRowArgKey)) {<a name="line.664"></a>
-<span class="sourceLineNo">665</span>          tableHash.startRow = Bytes.fromHex(cmd.substring(startRowArgKey.length()));<a name="line.665"></a>
-<span class="sourceLineNo">666</span>          continue;<a name="line.666"></a>
-<span class="sourceLineNo">667</span>        }<a name="line.667"></a>
-<span class="sourceLineNo">668</span><a name="line.668"></a>
-<span class="sourceLineNo">669</span>        final String stopRowArgKey = "--stoprow=";<a name="line.669"></a>
-<span class="sourceLineNo">670</span>        if (cmd.startsWith(stopRowArgKey)) {<a name="line.670"></a>
-<span class="sourceLineNo">671</span>          tableHash.stopRow = Bytes.fromHex(cmd.substring(stopRowArgKey.length()));<a name="line.671"></a>
-<span class="sourceLineNo">672</span>          continue;<a name="line.672"></a>
-<span class="sourceLineNo">673</span>        }<a name="line.673"></a>
-<span class="sourceLineNo">674</span><a name="line.674"></a>
-<span class="sourceLineNo">675</span>        final String startTimeArgKey = "--starttime=";<a name="line.675"></a>
-<span class="sourceLineNo">676</span>        if (cmd.startsWith(startTimeArgKey)) {<a name="line.676"></a>
-<span class="sourceLineNo">677</span>          tableHash.startTime = Long.parseLong(cmd.substring(startTimeArgKey.length()));<a name="line.677"></a>
-<span class="sourceLineNo">678</span>          continue;<a name="line.678"></a>
-<span class="sourceLineNo">679</span>        }<a name="line.679"></a>
-<span class="sourceLineNo">680</span><a name="line.680"></a>
-<span class="sourceLineNo">681</span>        final String endTimeArgKey = "--endtime=";<a name="line.681"></a>
-<span class="sourceLineNo">682</span>        if (cmd.startsWith(endTimeArgKey)) {<a name="line.682"></a>
-<span class="sourceLineNo">683</span>          tableHash.endTime = Long.parseLong(cmd.substring(endTimeArgKey.length()));<a name="line.683"></a>
-<span class="sourceLineNo">684</span>          continue;<a name="line.684"></a>
-<span class="sourceLineNo">685</span>        }<a name="line.685"></a>
-<span class="sourceLineNo">686</span><a name="line.686"></a>
-<span class="sourceLineNo">687</span>        final String scanBatchArgKey = "--scanbatch=";<a name="line.687"></a>
-<span class="sourceLineNo">688</span>        if (cmd.startsWith(scanBatchArgKey)) {<a name="line.688"></a>
-<span class="sourceLineNo">689</span>          tableHash.scanBatch = Integer.parseInt(cmd.substring(scanBatchArgKey.length()));<a name="line.689"></a>
-<span class="sourceLineNo">690</span>          continue;<a name="line.690"></a>
-<span class="sourceLineNo">691</span>        }<a name="line.691"></a>
-<span class="sourceLineNo">692</span><a name="line.692"></a>
-<span class="sourceLineNo">693</span>        final String versionsArgKey = "--versions=";<a name="line.693"></a>
-<span class="sourceLineNo">694</span>        if (cmd.startsWith(versionsArgKey)) {<a name="line.694"></a>
-<span class="sourceLineNo">695</span>          tableHash.versions = Integer.parseInt(cmd.substring(versionsArgKey.length()));<a name="line.695"></a>
-<span class="sourceLineNo">696</span>          continue;<a name="line.696"></a>
-<span class="sourceLineNo">697</span>        }<a name="line.697"></a>
-<span class="sourceLineNo">698</span><a name="line.698"></a>
-<span class="sourceLineNo">699</span>        final String familiesArgKey = "--families=";<a name="line.699"></a>
-<span class="sourceLineNo">700</span>        if (cmd.startsWith(familiesArgKey)) {<a name="line.700"></a>
-<span class="sourceLineNo">701</span>          tableHash.families = cmd.substring(familiesArgKey.length());<a name="line.701"></a>
-<span class="sourceLineNo">702</span>          continue;<a name="line.702"></a>
-<span class="sourceLineNo">703</span>        }<a name="line.703"></a>
-<span class="sourceLineNo">704</span><a name="line.704"></a>
-<span class="sourceLineNo">705</span>        printUsage("Invalid argument '" + cmd + "'");<a name="line.705"></a>
-<span class="sourceLineNo">706</span>        return false;<a name="line.706"></a>
-<span class="sourceLineNo">707</span>      }<a name="line.707"></a>
-<span class="sourceLineNo">708</span>      if ((tableHash.startTime != 0 || tableHash.endTime != 0)<a name="line.708"></a>
-<span class="sourceLineNo">709</span>          &amp;&amp; (tableHash.startTime &gt;= tableHash.endTime)) {<a name="line.709"></a>
-<span class="sourceLineNo">710</span>        printUsage("Invalid time range filter: starttime="<a name="line.710"></a>
-<span class="sourceLineNo">711</span>            + tableHash.startTime + " &gt;=  endtime=" + tableHash.endTime);<a name="line.711"></a>
-<span class="sourceLineNo">712</span>        return false;<a name="line.712"></a>
-<span class="sourceLineNo">713</span>      }<a name="line.713"></a>
-<span class="sourceLineNo">714</span><a name="line.714"></a>
-<span class="sourceLineNo">715</span>    } catch (Exception e) {<a name="line.715"></a>
-<span class="sourceLineNo">716</span>      LOG.error("Failed to parse commandLine arguments", e);<a name="line.716"></a>
-<span class="sourceLineNo">717</span>      printUsage("Can't start because " + e.getMessage());<a name="line.717"></a>
-<span class="sourceLineNo">718</span>      return false;<a name="line.718"></a>
-<span class="sourceLineNo">719</span>    }<a name="line.719"></a>
-<span class="sourceLineNo">720</span>    return true;<a name="line.720"></a>
-<span class="sourceLineNo">721</span>  }<a name="line.721"></a>
+<span class="sourceLineNo">470</span>  static class ResultHasher {<a name="line.470"></a>
+<span class="sourceLineNo">471</span>    private MessageDigest digest;<a name="line.471"></a>
+<span class="sourceLineNo">472</span><a name="line.472"></a>
+<span class="sourceLineNo">473</span>    private boolean batchStarted = false;<a name="line.473"></a>
+<span class="sourceLineNo">474</span>    private ImmutableBytesWritable batchStartKey;<a name="line.474"></a>
+<span class="sourceLineNo">475</span>    private ImmutableBytesWritable batchHash;<a name="line.475"></a>
+<span class="sourceLineNo">476</span>    private long batchSize = 0;<a name="line.476"></a>
+<span class="sourceLineNo">477</span>    boolean ignoreTimestamps;<a name="line.477"></a>
+<span class="sourceLineNo">478</span><a name="line.478"></a>
+<span class="sourceLineNo">479</span><a name="line.479"></a>
+<span class="sourceLineNo">480</span>    public ResultHasher() {<a name="line.480"></a>
+<span class="sourceLineNo">481</span>      try {<a name="line.481"></a>
+<span class="sourceLineNo">482</span>        digest = MessageDigest.getInstance("MD5");<a name="line.482"></a>
+<span class="sourceLineNo">483</span>      } catch (NoSuchAlgorithmException e) {<a name="line.483"></a>
+<span class="sourceLineNo">484</span>        Throwables.propagate(e);<a name="line.484"></a>
+<span class="sourceLineNo">485</span>      }<a name="line.485"></a>
+<span class="sourceLineNo">486</span>    }<a name="line.486"></a>
+<span class="sourceLineNo">487</span><a name="line.487"></a>
+<span class="sourceLineNo">488</span>    public void startBatch(ImmutableBytesWritable row) {<a name="line.488"></a>
+<span class="sourceLineNo">489</span>      if (batchStarted) {<a name="line.489"></a>
+<span class="sourceLineNo">490</span>        throw new RuntimeException("Cannot start new batch without finishing existing one.");<a name="line.490"></a>
+<span class="sourceLineNo">491</span>      }<a name="line.491"></a>
+<span class="sourceLineNo">492</span>      batchStarted = true;<a name="line.492"></a>
+<span class="sourceLineNo">493</span>      batchSize = 0;<a name="line.493"></a>
+<span class="sourceLineNo">494</span>      batchStartKey = row;<a name="line.494"></a>
+<span class="sourceLineNo">495</span>      batchHash = null;<a name="line.495"></a>
+<span class="sourceLineNo">496</span>    }<a name="line.496"></a>
+<span class="sourceLineNo">497</span><a name="line.497"></a>
+<span class="sourceLineNo">498</span>    public void hashResult(Result result) {<a name="line.498"></a>
+<span class="sourceLineNo">499</span>      if (!batchStarted) {<a name="line.499"></a>
+<span class="sourceLineNo">500</span>        throw new RuntimeException("Cannot add to batch that has not been started.");<a name="line.500"></a>
+<span class="sourceLineNo">501</span>      }<a name="line.501"></a>
+<span class="sourceLineNo">502</span>      for (Cell cell : result.rawCells()) {<a name="line.502"></a>
+<span class="sourceLineNo">503</span>        int rowLength = cell.getRowLength();<a name="line.503"></a>
+<span class="sourceLineNo">504</span>        int familyLength = cell.getFamilyLength();<a name="line.504"></a>
+<span class="sourceLineNo">505</span>        int qualifierLength = cell.getQualifierLength();<a name="line.505"></a>
+<span class="sourceLineNo">506</span>        int valueLength = cell.getValueLength();<a name="line.506"></a>
+<span class="sourceLineNo">507</span>        digest.update(cell.getRowArray(), cell.getRowOffset(), rowLength);<a name="line.507"></a>
+<span class="sourceLineNo">508</span>        digest.update(cell.getFamilyArray(), cell.getFamilyOffset(), familyLength);<a name="line.508"></a>
+<span class="sourceLineNo">509</span>        digest.update(cell.getQualifierArray(), cell.getQualifierOffset(), qualifierLength);<a name="line.509"></a>
+<span class="sourceLineNo">510</span><a name="line.510"></a>
+<span class="sourceLineNo">511</span>        if (!ignoreTimestamps) {<a name="line.511"></a>
+<span class="sourceLineNo">512</span>          long ts = cell.getTimestamp();<a name="line.512"></a>
+<span class="sourceLineNo">513</span>          for (int i = 8; i &gt; 0; i--) {<a name="line.513"></a>
+<span class="sourceLineNo">514</span>            digest.update((byte) ts);<a name="line.514"></a>
+<span class="sourceLineNo">515</span>            ts &gt;&gt;&gt;= 8;<a name="line.515"></a>
+<span class="sourceLineNo">516</span>          }<a name="line.516"></a>
+<span class="sourceLineNo">517</span>        }<a name="line.517"></a>
+<span class="sourceLineNo">518</span>        digest.update(cell.getValueArray(), cell.getValueOffset(), valueLength);<a name="line.518"></a>
+<span class="sourceLineNo">519</span><a name="line.519"></a>
+<span class="sourceLineNo">520</span>        batchSize += rowLength + familyLength + qualifierLength + 8 + valueLength;<a name="line.520"></a>
+<span class="sourceLineNo">521</span>      }<a name="line.521"></a>
+<span class="sourceLineNo">522</span>    }<a name="line.522"></a>
+<span class="sourceLineNo">523</span><a name="line.523"></a>
+<span class="sourceLineNo">524</span>    public void finishBatch() {<a name="line.524"></a>
+<span class="sourceLineNo">525</span>      if (!batchStarted) {<a name="line.525"></a>
+<span class="sourceLineNo">526</span>        throw new RuntimeException("Cannot finish batch that has not started.");<a name="line.526"></a>
+<span class="sourceLineNo">527</span>      }<a name="line.527"></a>
+<span class="sourceLineNo">528</span>      batchStarted = false;<a name="line.528"></a>
+<span class="sourceLineNo">529</span>      batchHash = new ImmutableBytesWritable(digest.digest());<a name="line.529"></a>
+<span class="sourceLineNo">530</span>    }<a name="line.530"></a>
+<span class="sourceLineNo">531</span><a name="line.531"></a>
+<span class="sourceLineNo">532</span>    public boolean isBatchStarted() {<a name="line.532"></a>
+<span class="sourceLineNo">533</span>      return batchStarted;<a name="line.533"></a>
+<span class="sourceLineNo">534</span>    }<a name="line.534"></a>
+<span class="sourceLineNo">535</span><a name="line.535"></a>
+<span class="sourceLineNo">536</span>    public ImmutableBytesWritable getBatchStartKey() {<a name="line.536"></a>
+<span class="sourceLineNo">537</span>      return batchStartKey;<a name="line.537"></a>
+<span class="sourceLineNo">538</span>    }<a name="line.538"></a>
+<span class="sourceLineNo">539</span><a name="line.539"></a>
+<span class="sourceLineNo">540</span>    public ImmutableBytesWritable getBatchHash() {<a name="line.540"></a>
+<span class="sourceLineNo">541</span>      return batchHash;<a name="line.541"></a>
+<span class="sourceLineNo">542</span>    }<a name="line.542"></a>
+<span class="sourceLineNo">543</span><a name="line.543"></a>
+<span class="sourceLineNo">544</span>    public long getBatchSize() {<a name="line.544"></a>
+<span class="sourceLineNo">545</span>      return batchSize;<a name="line.545"></a>
+<span class="sourceLineNo">546</span>    }<a name="line.546"></a>
+<span class="sourceLineNo">547</span>  }<a name="line.547"></a>
+<span class="sourceLineNo">548</span><a name="line.548"></a>
+<span class="sourceLineNo">549</span>  public static class HashMapper<a name="line.549"></a>
+<span class="sourceLineNo">550</span>    extends TableMapper&lt;ImmutableBytesWritable, ImmutableBytesWritable&gt; {<a name="line.550"></a>
+<span class="sourceLineNo">551</span><a name="line.551"></a>
+<span class="sourceLineNo">552</span>    private ResultHasher hasher;<a name="line.552"></a>
+<span class="sourceLineNo">553</span>    private long targetBatchSize;<a name="line.553"></a>
+<span class="sourceLineNo">554</span><a name="line.554"></a>
+<span class="sourceLineNo">555</span>    private ImmutableBytesWritable currentRow;<a name="line.555"></a>
+<span class="sourceLineNo">556</span><a name="line.556"></a>
+<span class="sourceLineNo">557</span>    @Override<a name="line.557"></a>
+<span class="sourceLineNo">558</span>    protected void setup(Context context) throws IOException, InterruptedException {<a name="line.558"></a>
+<span class="sourceLineNo">559</span>      targetBatchSize = context.getConfiguration()<a name="line.559"></a>
+<span class="sourceLineNo">560</span>          .getLong(HASH_BATCH_SIZE_CONF_KEY, DEFAULT_BATCH_SIZE);<a name="line.560"></a>
+<span class="sourceLineNo">561</span>      hasher = new ResultHasher();<a name="line.561"></a>
+<span class="sourceLineNo">562</span>      hasher.ignoreTimestamps = context.getConfiguration().<a name="line.562"></a>
+<span class="sourceLineNo">563</span>        getBoolean(IGNORE_TIMESTAMPS, false);<a name="line.563"></a>
+<span class="sourceLineNo">564</span>      TableSplit split = (TableSplit) context.getInputSplit();<a name="line.564"></a>
+<span class="sourceLineNo">565</span>      hasher.startBatch(new ImmutableBytesWritable(split.getStartRow()));<a name="line.565"></a>
+<span class="sourceLineNo">566</span>    }<a name="line.566"></a>
+<span class="sourceLineNo">567</span><a name="line.567"></a>
+<span class="sourceLineNo">568</span>    @Override<a name="line.568"></a>
+<span class="sourceLineNo">569</span>    protected void map(ImmutableBytesWritable key, Result value, Context context)<a name="line.569"></a>
+<span class="sourceLineNo">570</span>        throws IOException, InterruptedException {<a name="line.570"></a>
+<span class="sourceLineNo">571</span><a name="line.571"></a>
+<span class="sourceLineNo">572</span>      if (currentRow == null || !currentRow.equals(key)) {<a name="line.572"></a>
+<span class="sourceLineNo">573</span>        currentRow = new ImmutableBytesWritable(key); // not immutable<a name="line.573"></a>
+<span class="sourceLineNo">574</span><a name="line.574"></a>
+<span class="sourceLineNo">575</span>        if (hasher.getBatchSize() &gt;= targetBatchSize) {<a name="line.575"></a>
+<span class="sourceLineNo">576</span>          hasher.finishBatch();<a name="line.576"></a>
+<span class="sourceLineNo">577</span>          context.write(hasher.getBatchStartKey(), hasher.getBatchHash());<a name="line.577"></a>
+<span class="sourceLineNo">578</span>          hasher.startBatch(currentRow);<a name="line.578"></a>
+<span class="sourceLineNo">579</span>        }<a name="line.579"></a>
+<span class="sourceLineNo">580</span>      }<a name="line.580"></a>
+<span class="sourceLineNo">581</span><a name="line.581"></a>
+<span class="sourceLineNo">582</span>      hasher.hashResult(value);<a name="line.582"></a>
+<span class="sourceLineNo">583</span>    }<a name="line.583"></a>
+<span class="sourceLineNo">584</span><a name="line.584"></a>
+<span class="sourceLineNo">585</span>    @Override<a name="line.585"></a>
+<span class="sourceLineNo">586</span>    protected void cleanup(Context context) throws IOException, InterruptedException {<a name="line.586"></a>
+<span class="sourceLineNo">587</span>      hasher.finishBatch();<a name="line.587"></a>
+<span class="sourceLineNo">588</span>      context.write(hasher.getBatchStartKey(), hasher.getBatchHash());<a name="line.588"></a>
+<span class="sourceLineNo">589</span>    }<a name="line.589"></a>
+<span class="sourceLineNo">590</span>  }<a name="line.590"></a>
+<span class="sourceLineNo">591</span><a name="line.591"></a>
+<span class="sourceLineNo">592</span>  private void writeTempManifestFile() throws IOException {<a name="line.592"></a>
+<span class="sourceLineNo">593</span>    Path tempManifestPath = new Path(destPath, TMP_MANIFEST_FILE_NAME);<a name="line.593"></a>
+<span class="sourceLineNo">594</span>    FileSystem fs = tempManifestPath.getFileSystem(getConf());<a name="line.594"></a>
+<span class="sourceLineNo">595</span>    tableHash.writePropertiesFile(fs, tempManifestPath);<a name="line.595"></a>
+<span class="sourceLineNo">596</span>  }<a name="line.596"></a>
+<span class="sourceLineNo">597</span><a name="line.597"></a>
+<span class="sourceLineNo">598</span>  private void completeManifest() throws IOException {<a name="line.598"></a>
+<span class="sourceLineNo">599</span>    Path tempManifestPath = new Path(destPath, TMP_MANIFEST_FILE_NAME);<a name="line.599"></a>
+<span class="sourceLineNo">600</span>    Path manifestPath = new Path(destPath, MANIFEST_FILE_NAME);<a name="line.600"></a>
+<span class="sourceLineNo">601</span>    FileSystem fs = tempManifestPath.getFileSystem(getConf());<a name="line.601"></a>
+<span class="sourceLineNo">602</span>    fs.rename(tempManifestPath, manifestPath);<a name="line.602"></a>
+<span class="sourceLineNo">603</span>  }<a name="line.603"></a>
+<span class="sourceLineNo">604</span><a name="line.604"></a>
+<span class="sourceLineNo">605</span>  private static final int NUM_ARGS = 2;<a name="line.605"></a>
+<span class="sourceLineNo">606</span>  private static void printUsage(final String errorMsg) {<a name="line.606"></a>
+<span class="sourceLineNo">607</span>    if (errorMsg != null &amp;&amp; errorMsg.length() &gt; 0) {<a name="line.607"></a>
+<span class="sourceLineNo">608</span>      System.err.println("ERROR: " + errorMsg);<a name="line.608"></a>
+<span class="sourceLineNo">609</span>      System.err.println();<a name="line.609"></a>
+<span class="sourceLineNo">610</span>    }<a name="line.610"></a>
+<span class="sourceLineNo">611</span>    System.err.println("Usage: HashTable [options] &lt;tablename&gt; &lt;outputpath&gt;");<a name="line.611"></a>
+<span class="sourceLineNo">612</span>    System.err.println();<a name="line.612"></a>
+<span class="sourceLineNo">613</span>    System.err.println("Options:");<a name="line.613"></a>
+<span class="sourceLineNo">614</span>    System.err.println(" batchsize         the target amount of bytes to hash in each batch");<a name="line.614"></a>
+<span class="sourceLineNo">615</span>    System.err.println("                   rows are added to the batch until this size is reached");<a name="line.615"></a>
... 14468 lines suppressed ...