You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by en...@apache.org on 2014/09/09 08:40:51 UTC

[8/8] git commit: HBASE-11679 Replace HTable with HTableInterface where backwards-compatible (Carter)

HBASE-11679 Replace HTable with HTableInterface where backwards-compatible (Carter)


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/4995ed8a
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/4995ed8a
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/4995ed8a

Branch: refs/heads/master
Commit: 4995ed8a029feb8ccac8054f56f23261a6918add
Parents: 71e6ff4
Author: Enis Soztutar <en...@apache.org>
Authored: Mon Sep 8 23:39:34 2014 -0700
Committer: Enis Soztutar <en...@apache.org>
Committed: Mon Sep 8 23:39:34 2014 -0700

----------------------------------------------------------------------
 .../apache/hadoop/hbase/MetaTableAccessor.java  |  29 +--
 .../org/apache/hadoop/hbase/client/HTable.java  |   3 +-
 .../apache/hadoop/hbase/client/HTableUtil.java  |   5 +-
 .../client/coprocessor/AggregationClient.java   |  35 ++--
 .../coprocessor/SecureBulkLoadClient.java       |   6 +-
 .../security/access/AccessControlClient.java    |  10 +-
 .../security/visibility/VisibilityClient.java   |   7 +-
 .../hbase/client/TestClientNoCluster.java       |  10 +-
 .../hbase/client/TestSnapshotFromAdmin.java     |   6 +-
 .../example/TestBulkDeleteProtocol.java         |  23 ++-
 .../example/TestRowCountEndpoint.java           |   7 +-
 .../TestZooKeeperScanPolicyObserver.java        |   3 +-
 .../hadoop/hbase/DistributedHBaseCluster.java   |   3 +-
 .../hbase/IntegrationTestLazyCfLoading.java     |   3 +-
 .../mapreduce/IntegrationTestImportTsv.java     |   3 +-
 .../hadoop/hbase/mttr/IntegrationTestMTTR.java  |   8 +-
 .../test/IntegrationTestBigLinkedList.java      |  14 +-
 ...egrationTestBigLinkedListWithVisibility.java |  10 +-
 .../test/IntegrationTestLoadAndVerify.java      |   2 +-
 ...stTimeBoundedRequestsWithRegionReplicas.java |   4 +-
 ...tionTestWithCellVisibilityLoadAndVerify.java |   3 +-
 .../trace/IntegrationTestSendTraceRequests.java |   5 +-
 .../apache/hadoop/hbase/LocalHBaseCluster.java  |   3 +-
 .../hbase/coprocessor/CoprocessorHost.java      |   1 +
 .../hadoop/hbase/mapred/HRegionPartitioner.java |   3 +-
 .../hadoop/hbase/mapred/TableOutputFormat.java  |   5 +-
 .../hadoop/hbase/mapred/TableRecordReader.java  |   3 +-
 .../hbase/mapred/TableRecordReaderImpl.java     |   5 +-
 .../DefaultVisibilityExpressionResolver.java    |   3 +-
 .../hbase/mapreduce/HFileOutputFormat.java      |   9 +-
 .../hbase/mapreduce/HFileOutputFormat2.java     |  12 +-
 .../hbase/mapreduce/HRegionPartitioner.java     |   3 +-
 .../hadoop/hbase/mapreduce/ImportTsv.java       |   3 +-
 .../hbase/mapreduce/LoadIncrementalHFiles.java  |  11 +-
 .../mapreduce/MultiTableInputFormatBase.java    |   3 +-
 .../hbase/mapreduce/TableInputFormatBase.java   |   2 -
 .../hbase/mapreduce/TableOutputFormat.java      |   5 +-
 .../hbase/mapreduce/TableRecordReader.java      |   3 +-
 .../hbase/mapreduce/TableRecordReaderImpl.java  |   5 +-
 .../replication/VerifyReplication.java          |   3 +-
 .../hadoop/hbase/master/HMasterCommandLine.java |   4 +-
 .../hbase/master/RegionPlacementMaintainer.java |   5 +-
 .../hbase/master/TableNamespaceManager.java     |  11 +-
 .../balancer/FavoredNodeAssignmentHelper.java   |   3 +-
 .../master/handler/ModifyTableHandler.java      |   3 +-
 .../regionserver/ReplicationSink.java           |   4 +-
 .../apache/hadoop/hbase/rest/RESTServlet.java   |   4 +-
 .../apache/hadoop/hbase/rest/RowResource.java   |  12 +-
 .../hadoop/hbase/rest/RowResultGenerator.java   |   4 +-
 .../hbase/rest/ScannerResultGenerator.java      |   4 +-
 .../hadoop/hbase/rest/SchemaResource.java       |   4 +-
 .../apache/hadoop/hbase/rest/TableResource.java |   4 +-
 .../security/access/AccessControlLists.java     |  15 +-
 .../hadoop/hbase/security/token/TokenUtil.java  |   3 +-
 .../org/apache/hadoop/hbase/tool/Canary.java    |   9 +-
 .../org/apache/hadoop/hbase/util/HBaseFsck.java |   8 +-
 .../hadoop/hbase/util/HBaseFsckRepair.java      |  12 +-
 .../org/apache/hadoop/hbase/util/HMerge.java    |   6 +-
 .../hadoop/hbase/util/RegionSizeCalculator.java |   4 +-
 .../org/apache/hadoop/hbase/HBaseTestCase.java  |   5 +-
 .../hadoop/hbase/HBaseTestingUtility.java       |  33 +--
 .../hadoop/hbase/PerformanceEvaluation.java     |   6 +-
 .../hadoop/hbase/ScanPerformanceEvaluation.java |   3 +-
 .../apache/hadoop/hbase/TestAcidGuarantees.java |   7 +-
 .../hadoop/hbase/TestHBaseTestingUtility.java   |   6 +-
 .../hadoop/hbase/TestMetaTableAccessor.java     |   5 +-
 .../apache/hadoop/hbase/TestMultiVersions.java  |   8 +-
 .../org/apache/hadoop/hbase/TestNamespace.java  |   3 +-
 .../hadoop/hbase/TestRegionRebalancing.java     |   6 +-
 .../org/apache/hadoop/hbase/TestZooKeeper.java  |  16 +-
 .../apache/hadoop/hbase/client/TestAdmin.java   |  20 +-
 .../client/TestClientOperationInterrupt.java    |   6 +-
 .../client/TestClientScannerRPCTimeout.java     |   4 +-
 .../hadoop/hbase/client/TestClientTimeouts.java |   2 +-
 .../hadoop/hbase/client/TestFromClientSide.java | 180 ++++++++---------
 .../hbase/client/TestFromClientSide3.java       |  10 +-
 .../hbase/client/TestHBaseAdminNoCluster.java   |   3 +-
 .../org/apache/hadoop/hbase/client/TestHCM.java |  12 +-
 .../hadoop/hbase/client/TestMultiParallel.java  |  22 +-
 .../hbase/client/TestMultipleTimestamps.java    |  32 +--
 .../hadoop/hbase/client/TestPutWithDelete.java  |   2 +-
 .../hbase/client/TestReplicaWithCluster.java    |  12 +-
 .../hadoop/hbase/client/TestReplicasClient.java |   2 +-
 .../hbase/client/TestRpcControllerFactory.java  |   4 +-
 .../hadoop/hbase/client/TestScannerTimeout.java |  10 +-
 .../client/TestScannersFromClientSide.java      |   8 +-
 .../client/TestSnapshotCloneIndependence.java   |   2 +-
 .../hbase/client/TestSnapshotMetadata.java      |   6 +-
 .../hbase/client/TestTimestampsFilter.java      |  16 +-
 .../hadoop/hbase/constraint/TestConstraint.java |  11 +-
 .../TestBatchCoprocessorEndpoint.java           |  16 +-
 .../coprocessor/TestCoprocessorEndpoint.java    |  11 +-
 .../hbase/coprocessor/TestHTableWrapper.java    |   4 +-
 .../coprocessor/TestOpenTableInCoprocessor.java |  13 +-
 .../coprocessor/TestRegionObserverBypass.java   |   9 +-
 .../TestRegionObserverInterface.java            |  21 +-
 .../TestRegionObserverScannerOpenHook.java      |   3 +-
 .../coprocessor/TestRegionServerObserver.java   |   3 +-
 .../coprocessor/TestRowProcessorEndpoint.java   |   8 +-
 .../hbase/filter/TestColumnRangeFilter.java     |   7 +-
 .../hbase/filter/TestFilterWithScanLimits.java  |   5 +-
 .../hadoop/hbase/filter/TestFilterWrapper.java  |   6 +-
 .../TestFuzzyRowAndColumnRangeFilter.java       |   8 +-
 .../hadoop/hbase/fs/TestBlockReorder.java       |   4 +-
 .../hbase/io/encoding/TestChangingEncoding.java |   5 +-
 .../hbase/io/encoding/TestPrefixTree.java       |   3 +-
 .../hbase/mapred/TestTableInputFormat.java      |  38 ++--
 .../hbase/mapred/TestTableMapReduceUtil.java    |  10 +-
 .../hadoop/hbase/mapreduce/TestCellCounter.java |   4 +-
 .../hadoop/hbase/mapreduce/TestCopyTable.java   |  14 +-
 .../hbase/mapreduce/TestHFileOutputFormat.java  |  28 +--
 .../hbase/mapreduce/TestHFileOutputFormat2.java |  28 +--
 .../hbase/mapreduce/TestImportExport.java       |  19 +-
 .../TestImportTSVWithOperationAttributes.java   |   6 +-
 .../TestImportTSVWithVisibilityLabels.java      |   8 +-
 .../hadoop/hbase/mapreduce/TestImportTsv.java   |   3 +-
 .../mapreduce/TestLoadIncrementalHFiles.java    |   4 +-
 .../TestLoadIncrementalHFilesSplitRecovery.java |   5 +-
 .../mapreduce/TestMultithreadedTableMapper.java |   5 +-
 .../hadoop/hbase/mapreduce/TestRowCounter.java  |   6 +-
 .../hbase/mapreduce/TestTableMapReduceBase.java |   5 +-
 .../hbase/mapreduce/TestTimeRangeMapRed.java    |  12 +-
 .../hadoop/hbase/mapreduce/TestWALPlayer.java   |   6 +-
 .../hbase/master/TestAssignmentListener.java    |   3 +-
 .../master/TestAssignmentManagerOnCluster.java  |  27 +--
 .../master/TestDistributedLogSplitting.java     |  25 +--
 .../hadoop/hbase/master/TestMasterFailover.java |   4 +-
 .../TestMasterOperationsForRegionReplicas.java  |   6 +-
 .../TestMasterRestartAfterDisablingTable.java   |   3 +-
 .../hbase/master/TestMasterTransitions.java     |   5 +-
 .../hbase/master/TestRegionPlacement.java       |   4 +-
 .../hbase/regionserver/TestCompactionState.java |   8 +-
 .../regionserver/TestEncryptionKeyRotation.java |   3 +-
 .../TestEncryptionRandomKeying.java             |   3 +-
 .../TestEndToEndSplitTransaction.java           |  15 +-
 .../hbase/regionserver/TestFSErrorsExposed.java |   3 +-
 .../hadoop/hbase/regionserver/TestHRegion.java  |   4 +-
 .../regionserver/TestHRegionOnCluster.java      |   4 +-
 .../hbase/regionserver/TestJoinedScanners.java  |   5 +-
 .../TestRegionMergeTransactionOnCluster.java    |  18 +-
 .../hbase/regionserver/TestRegionReplicas.java  |   3 +-
 .../regionserver/TestRegionServerMetrics.java   |  10 +-
 .../regionserver/TestSCVFWithMiniCluster.java   |   4 +-
 .../regionserver/TestScannerWithBulkload.java   |   3 +-
 .../regionserver/TestServerCustomProtocol.java  |  22 +-
 .../TestSplitTransactionOnCluster.java          |  20 +-
 .../hadoop/hbase/regionserver/TestTags.java     |  11 +-
 .../regionserver/wal/TestHLogFiltering.java     |   4 +-
 .../regionserver/wal/TestLogRollAbort.java      |   4 +-
 .../regionserver/wal/TestLogRollPeriod.java     |  10 +-
 .../hbase/regionserver/wal/TestLogRolling.java  |  23 ++-
 .../hbase/regionserver/wal/TestWALReplay.java   |   3 +-
 .../replication/TestMasterReplication.java      |  26 +--
 .../replication/TestMultiSlaveReplication.java  |  21 +-
 .../replication/TestPerTableCFReplication.java  |  31 +--
 .../hbase/replication/TestReplicationBase.java  |   8 +-
 .../replication/TestReplicationSyncUpTool.java  |   6 +-
 .../replication/TestReplicationWithTags.java    |   8 +-
 .../TestRegionReplicaReplicationEndpoint.java   |  12 +-
 .../regionserver/TestReplicationSink.java       |   6 +-
 .../hadoop/hbase/rest/TestGzipFilter.java       |   5 +-
 .../hadoop/hbase/rest/TestScannerResource.java  |   4 +-
 .../hbase/rest/TestScannersWithFilters.java     |   4 +-
 .../hbase/rest/TestScannersWithLabels.java      |   4 +-
 .../hbase/rest/client/TestRemoteTable.java      |   6 +-
 .../hbase/security/access/SecureTestUtil.java   |  15 +-
 .../access/TestAccessControlFilter.java         |  11 +-
 .../security/access/TestAccessController.java   |  65 +++---
 .../security/access/TestAccessController2.java  |   3 +-
 .../access/TestCellACLWithMultipleVersions.java |  60 +++---
 .../hbase/security/access/TestCellACLs.java     |  34 ++--
 .../security/access/TestNamespaceCommands.java  |   7 +-
 .../access/TestScanEarlyTermination.java        |  12 +-
 .../security/access/TestTablePermissions.java   |   3 +-
 .../ExpAsStringVisibilityLabelServiceImpl.java  |   3 +-
 .../TestEnforcingScanLabelGenerator.java        |   5 +-
 .../visibility/TestVisibilityLabels.java        |  39 ++--
 .../visibility/TestVisibilityLabelsWithACL.java |   9 +-
 ...ibilityLabelsWithDefaultVisLabelService.java |   4 +-
 .../TestVisibilityLabelsWithDeletes.java        | 199 ++++++++++---------
 .../TestVisibilityLabelsWithSLGStack.java       |   4 +-
 .../TestVisibilityWithCheckAuths.java           |  13 +-
 .../hbase/snapshot/SnapshotTestingUtils.java    |   6 +-
 .../hbase/snapshot/TestExportSnapshot.java      |  10 +-
 .../hadoop/hbase/trace/TestHTraceHooks.java     |   4 +-
 .../apache/hadoop/hbase/util/LoadTestTool.java  |   3 +-
 .../hadoop/hbase/util/MultiThreadedReader.java  |   9 +-
 .../hbase/util/MultiThreadedReaderWithACL.java  |   7 +-
 .../hadoop/hbase/util/MultiThreadedUpdater.java |  11 +-
 .../hbase/util/MultiThreadedUpdaterWithACL.java |  13 +-
 .../hadoop/hbase/util/MultiThreadedWriter.java  |   5 +-
 .../hbase/util/MultiThreadedWriterBase.java     |   4 +-
 .../hbase/util/MultiThreadedWriterWithACL.java  |   7 +-
 .../hadoop/hbase/util/RestartMetaTest.java      |   3 +-
 .../hbase/util/TestCoprocessorScanPolicy.java   |   5 +-
 .../apache/hadoop/hbase/util/TestHBaseFsck.java |  37 ++--
 .../hbase/util/TestHBaseFsckEncryption.java     |   3 +-
 .../hadoop/hbase/util/TestMergeTable.java       |   3 +-
 .../util/TestMiniClusterLoadSequential.java     |   5 +-
 .../hbase/util/TestProcessBasedCluster.java     |   3 +-
 .../hbase/util/TestRegionSizeCalculator.java    |  11 +-
 .../util/hbck/OfflineMetaRebuildTestCore.java   |  13 +-
 .../hadoop/hbase/thrift/IncrementCoalescer.java |   4 +-
 .../thrift2/ThriftHBaseServiceHandler.java      |  33 +--
 .../hadoop/hbase/thrift2/TestHTablePool.java    |  53 ++---
 .../thrift2/TestThriftHBaseServiceHandler.java  |   3 +-
 ...TestThriftHBaseServiceHandlerWithLabels.java |   3 +-
 207 files changed, 1226 insertions(+), 1105 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java
index c63e4c6..af9b587 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java
@@ -34,6 +34,7 @@ import org.apache.hadoop.hbase.client.RegionReplicaUtil;
 import org.apache.hadoop.hbase.client.Result;
 import org.apache.hadoop.hbase.client.ResultScanner;
 import org.apache.hadoop.hbase.client.Scan;
+import org.apache.hadoop.hbase.client.Table;
 import org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel;
 import org.apache.hadoop.hbase.protobuf.ProtobufUtil;
 import org.apache.hadoop.hbase.protobuf.generated.ClientProtos;
@@ -174,7 +175,7 @@ public class MetaTableAccessor {
    * @throws IOException
    * @SuppressWarnings("deprecation")
    */
-  private static HTable getHTable(final HConnection hConnection,
+  private static Table getHTable(final HConnection hConnection,
       final TableName tableName)
   throws IOException {
     // We used to pass whole CatalogTracker in here, now we just pass in HConnection
@@ -190,7 +191,7 @@ public class MetaTableAccessor {
    * @return An {@link HTable} for <code>hbase:meta</code>
    * @throws IOException
    */
-  static HTable getMetaHTable(final HConnection hConnection)
+  static Table getMetaHTable(final HConnection hConnection)
   throws IOException {
     return getHTable(hConnection, TableName.META_TABLE_NAME);
   }
@@ -200,7 +201,7 @@ public class MetaTableAccessor {
    * @param g Get to run
    * @throws IOException
    */
-  private static Result get(final HTable t, final Get g) throws IOException {
+  private static Result get(final Table t, final Get g) throws IOException {
     try {
       return t.get(g);
     } finally {
@@ -599,7 +600,7 @@ public class MetaTableAccessor {
       scan.setCaching(caching);
     }
     scan.addFamily(HConstants.CATALOG_FAMILY);
-    HTable metaTable = getMetaHTable(hConnection);
+    Table metaTable = getMetaHTable(hConnection);
     ResultScanner scanner = null;
     try {
       scanner = metaTable.getScanner(scan);
@@ -948,7 +949,7 @@ public class MetaTableAccessor {
    * @param p put to make
    * @throws IOException
    */
-  private static void put(final HTable t, final Put p) throws IOException {
+  private static void put(final Table t, final Put p) throws IOException {
     try {
       t.put(p);
     } finally {
@@ -964,7 +965,7 @@ public class MetaTableAccessor {
    */
   public static void putsToMetaTable(final HConnection hConnection, final List<Put> ps)
     throws IOException {
-    HTable t = getMetaHTable(hConnection);
+    Table t = getMetaHTable(hConnection);
     try {
       t.put(ps);
     } finally {
@@ -993,7 +994,7 @@ public class MetaTableAccessor {
    */
   public static void deleteFromMetaTable(final HConnection hConnection, final List<Delete> deletes)
     throws IOException {
-    HTable t = getMetaHTable(hConnection);
+    Table t = getMetaHTable(hConnection);
     try {
       t.delete(deletes);
     } finally {
@@ -1036,7 +1037,7 @@ public class MetaTableAccessor {
   public static void mutateMetaTable(final HConnection hConnection,
                                      final List<Mutation> mutations)
     throws IOException {
-    HTable t = getMetaHTable(hConnection);
+    Table t = getMetaHTable(hConnection);
     try {
       t.batch(mutations);
     } catch (InterruptedException e) {
@@ -1068,7 +1069,7 @@ public class MetaTableAccessor {
    * @param regionInfo region information
    * @throws IOException if problem connecting or updating meta
    */
-  public static void addRegionToMeta(HTable meta, HRegionInfo regionInfo) throws IOException {
+  public static void addRegionToMeta(Table meta, HRegionInfo regionInfo) throws IOException {
     addRegionToMeta(meta, regionInfo, null, null);
   }
 
@@ -1085,7 +1086,7 @@ public class MetaTableAccessor {
    * @param splitB second split daughter of the parent regionInfo
    * @throws IOException if problem connecting or updating meta
    */
-  public static void addRegionToMeta(HTable meta, HRegionInfo regionInfo,
+  public static void addRegionToMeta(Table meta, HRegionInfo regionInfo,
                                      HRegionInfo splitA, HRegionInfo splitB) throws IOException {
     Put put = makePutFromRegionInfo(regionInfo);
     addDaughtersToPut(put, splitA, splitB);
@@ -1109,7 +1110,7 @@ public class MetaTableAccessor {
    */
   public static void addRegionToMeta(HConnection hConnection, HRegionInfo regionInfo,
                                      HRegionInfo splitA, HRegionInfo splitB) throws IOException {
-    HTable meta = getMetaHTable(hConnection);
+    Table meta = getMetaHTable(hConnection);
     try {
       addRegionToMeta(meta, regionInfo, splitA, splitB);
     } finally {
@@ -1168,7 +1169,7 @@ public class MetaTableAccessor {
    */
   public static void mergeRegions(final HConnection hConnection, HRegionInfo mergedRegion,
       HRegionInfo regionA, HRegionInfo regionB, ServerName sn) throws IOException {
-    HTable meta = getMetaHTable(hConnection);
+    Table meta = getMetaHTable(hConnection);
     try {
       HRegionInfo copyOfMerged = new HRegionInfo(mergedRegion);
 
@@ -1208,7 +1209,7 @@ public class MetaTableAccessor {
   public static void splitRegion(final HConnection hConnection,
                                  HRegionInfo parent, HRegionInfo splitA, HRegionInfo splitB,
                                  ServerName sn) throws IOException {
-    HTable meta = getMetaHTable(hConnection);
+    Table meta = getMetaHTable(hConnection);
     try {
       HRegionInfo copyOfParent = new HRegionInfo(parent);
       copyOfParent.setOffline(true);
@@ -1235,7 +1236,7 @@ public class MetaTableAccessor {
   /**
    * Performs an atomic multi-Mutate operation against the given table.
    */
-  private static void multiMutate(HTable table, byte[] row, Mutation... mutations)
+  private static void multiMutate(Table table, byte[] row, Mutation... mutations)
       throws IOException {
     CoprocessorRpcChannel channel = table.coprocessorService(row);
     MultiRowMutationProtos.MutateRowsRequest.Builder mmrBuilder

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java
index 8a68c56..40d46c4 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java
@@ -22,7 +22,6 @@ import java.io.Closeable;
 import java.io.IOException;
 import java.io.InterruptedIOException;
 import java.util.ArrayList;
-import java.util.Arrays;
 import java.util.Collections;
 import java.util.LinkedList;
 import java.util.List;
@@ -1779,7 +1778,7 @@ public class HTable implements HTableInterface, RegionLocator {
    * @throws IOException
    */
   public static void main(String[] args) throws IOException {
-    HTable t = new HTable(HBaseConfiguration.create(), args[0]);
+    Table t = new HTable(HBaseConfiguration.create(), args[0]);
     try {
       System.out.println(t.get(new Get(Bytes.toBytes(args[1]))));
     } finally {

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableUtil.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableUtil.java
index 27aec10..04b3fea 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableUtil.java
@@ -19,7 +19,6 @@
 package org.apache.hadoop.hbase.client;
 
 import org.apache.hadoop.classification.InterfaceAudience;
-import org.apache.hadoop.classification.InterfaceStability;
 import org.apache.hadoop.hbase.HRegionLocation;
 
 import java.io.IOException;
@@ -102,7 +101,7 @@ public class HTableUtil {
 		
   }
 
-  private static Map<String,List<Put>> createRsPutMap(HTable htable, List<Put> puts) throws IOException {
+  private static Map<String,List<Put>> createRsPutMap(RegionLocator htable, List<Put> puts) throws IOException {
 
     Map<String, List<Put>> putMap = new HashMap<String, List<Put>>();
     for (Put put: puts) {
@@ -118,7 +117,7 @@ public class HTableUtil {
     return putMap;
   }
 
-  private static Map<String,List<Row>> createRsRowMap(HTable htable, List<Row> rows) throws IOException {
+  private static Map<String,List<Row>> createRsRowMap(RegionLocator htable, List<Row> rows) throws IOException {
 
     Map<String, List<Row>> rowMap = new HashMap<String, List<Row>>();
     for (Row row: rows) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java
index 3b37238..9058289 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java
@@ -40,6 +40,7 @@ import org.apache.hadoop.hbase.client.HTable;
 import org.apache.hadoop.hbase.client.Result;
 import org.apache.hadoop.hbase.client.ResultScanner;
 import org.apache.hadoop.hbase.client.Scan;
+import org.apache.hadoop.hbase.client.Table;
 import org.apache.hadoop.hbase.coprocessor.ColumnInterpreter;
 import org.apache.hadoop.hbase.ipc.BlockingRpcCallback;
 import org.apache.hadoop.hbase.ipc.ServerRpcController;
@@ -101,7 +102,7 @@ public class AggregationClient {
   public <R, S, P extends Message, Q extends Message, T extends Message> R max(
       final TableName tableName, final ColumnInterpreter<R, S, P, Q, T> ci, final Scan scan)
       throws Throwable {
-    HTable table = null;
+    Table table = null;
     try {
       table = new HTable(conf, tableName);
       return max(table, ci, scan);
@@ -125,7 +126,7 @@ public class AggregationClient {
    *           & propagated to it.
    */
   public <R, S, P extends Message, Q extends Message, T extends Message> 
-  R max(final HTable table, final ColumnInterpreter<R, S, P, Q, T> ci,
+  R max(final Table table, final ColumnInterpreter<R, S, P, Q, T> ci,
       final Scan scan) throws Throwable {
     final AggregateRequest requestArg = validateArgAndGetPB(scan, ci, false);
     class MaxCallBack implements Batch.Callback<R> {
@@ -196,7 +197,7 @@ public class AggregationClient {
   public <R, S, P extends Message, Q extends Message, T extends Message> R min(
       final TableName tableName, final ColumnInterpreter<R, S, P, Q, T> ci, final Scan scan)
       throws Throwable {
-    HTable table = null;
+    Table table = null;
     try {
       table = new HTable(conf, tableName);
       return min(table, ci, scan);
@@ -218,7 +219,7 @@ public class AggregationClient {
    * @throws Throwable
    */
   public <R, S, P extends Message, Q extends Message, T extends Message> 
-  R min(final HTable table, final ColumnInterpreter<R, S, P, Q, T> ci,
+  R min(final Table table, final ColumnInterpreter<R, S, P, Q, T> ci,
       final Scan scan) throws Throwable {
     final AggregateRequest requestArg = validateArgAndGetPB(scan, ci, false);
     class MinCallBack implements Batch.Callback<R> {
@@ -276,7 +277,7 @@ public class AggregationClient {
   public <R, S, P extends Message, Q extends Message, T extends Message> long rowCount(
       final TableName tableName, final ColumnInterpreter<R, S, P, Q, T> ci, final Scan scan)
       throws Throwable {
-    HTable table = null;
+    Table table = null;
     try {
       table = new HTable(conf, tableName);
       return rowCount(table, ci, scan);
@@ -301,7 +302,7 @@ public class AggregationClient {
    * @throws Throwable
    */
   public <R, S, P extends Message, Q extends Message, T extends Message> 
-  long rowCount(final HTable table,
+  long rowCount(final Table table,
       final ColumnInterpreter<R, S, P, Q, T> ci, final Scan scan) throws Throwable {
     final AggregateRequest requestArg = validateArgAndGetPB(scan, ci, true);
     class RowNumCallback implements Batch.Callback<Long> {
@@ -350,7 +351,7 @@ public class AggregationClient {
   public <R, S, P extends Message, Q extends Message, T extends Message> S sum(
       final TableName tableName, final ColumnInterpreter<R, S, P, Q, T> ci, final Scan scan)
       throws Throwable {
-    HTable table = null;
+    Table table = null;
     try {
       table = new HTable(conf, tableName);
       return sum(table, ci, scan);
@@ -371,7 +372,7 @@ public class AggregationClient {
    * @throws Throwable
    */
   public <R, S, P extends Message, Q extends Message, T extends Message> 
-  S sum(final HTable table, final ColumnInterpreter<R, S, P, Q, T> ci,
+  S sum(final Table table, final ColumnInterpreter<R, S, P, Q, T> ci,
       final Scan scan) throws Throwable {
     final AggregateRequest requestArg = validateArgAndGetPB(scan, ci, false);
     
@@ -423,7 +424,7 @@ public class AggregationClient {
   private <R, S, P extends Message, Q extends Message, T extends Message> Pair<S, Long> getAvgArgs(
       final TableName tableName, final ColumnInterpreter<R, S, P, Q, T> ci, final Scan scan)
       throws Throwable {
-    HTable table = null;
+    Table table = null;
     try {
       table = new HTable(conf, tableName);
       return getAvgArgs(table, ci, scan);
@@ -443,7 +444,7 @@ public class AggregationClient {
    * @throws Throwable
    */
   private <R, S, P extends Message, Q extends Message, T extends Message>
-  Pair<S, Long> getAvgArgs(final HTable table,
+  Pair<S, Long> getAvgArgs(final Table table,
       final ColumnInterpreter<R, S, P, Q, T> ci, final Scan scan) throws Throwable {
     final AggregateRequest requestArg = validateArgAndGetPB(scan, ci, false);
     class AvgCallBack implements Batch.Callback<Pair<S, Long>> {
@@ -523,7 +524,7 @@ public class AggregationClient {
    * @throws Throwable
    */
   public <R, S, P extends Message, Q extends Message, T extends Message> double avg(
-      final HTable table, final ColumnInterpreter<R, S, P, Q, T> ci, Scan scan) throws Throwable {
+      final Table table, final ColumnInterpreter<R, S, P, Q, T> ci, Scan scan) throws Throwable {
     Pair<S, Long> p = getAvgArgs(table, ci, scan);
     return ci.divideForAvg(p.getFirst(), p.getSecond());
   }
@@ -540,7 +541,7 @@ public class AggregationClient {
    * @throws Throwable
    */
   private <R, S, P extends Message, Q extends Message, T extends Message>
-  Pair<List<S>, Long> getStdArgs(final HTable table,
+  Pair<List<S>, Long> getStdArgs(final Table table,
       final ColumnInterpreter<R, S, P, Q, T> ci, final Scan scan) throws Throwable {
     final AggregateRequest requestArg = validateArgAndGetPB(scan, ci, false);
     class StdCallback implements Batch.Callback<Pair<List<S>, Long>> {
@@ -614,7 +615,7 @@ public class AggregationClient {
   public <R, S, P extends Message, Q extends Message, T extends Message>
   double std(final TableName tableName, ColumnInterpreter<R, S, P, Q, T> ci,
       Scan scan) throws Throwable {
-    HTable table = null;
+    Table table = null;
     try {
       table = new HTable(conf, tableName);
       return std(table, ci, scan);
@@ -638,7 +639,7 @@ public class AggregationClient {
    * @throws Throwable
    */
   public <R, S, P extends Message, Q extends Message, T extends Message> double std(
-      final HTable table, ColumnInterpreter<R, S, P, Q, T> ci, Scan scan) throws Throwable {
+      final Table table, ColumnInterpreter<R, S, P, Q, T> ci, Scan scan) throws Throwable {
     Pair<List<S>, Long> p = getStdArgs(table, ci, scan);
     double res = 0d;
     double avg = ci.divideForAvg(p.getFirst().get(0), p.getSecond());
@@ -662,7 +663,7 @@ public class AggregationClient {
    */
   private <R, S, P extends Message, Q extends Message, T extends Message>
   Pair<NavigableMap<byte[], List<S>>, List<S>>
-  getMedianArgs(final HTable table,
+  getMedianArgs(final Table table,
       final ColumnInterpreter<R, S, P, Q, T> ci, final Scan scan) throws Throwable {
     final AggregateRequest requestArg = validateArgAndGetPB(scan, ci, false);
     final NavigableMap<byte[], List<S>> map =
@@ -727,7 +728,7 @@ public class AggregationClient {
   public <R, S, P extends Message, Q extends Message, T extends Message>
   R median(final TableName tableName, ColumnInterpreter<R, S, P, Q, T> ci,
       Scan scan) throws Throwable {
-    HTable table = null;
+    Table table = null;
     try {
       table = new HTable(conf, tableName);
       return median(table, ci, scan);
@@ -749,7 +750,7 @@ public class AggregationClient {
    * @throws Throwable
    */
   public <R, S, P extends Message, Q extends Message, T extends Message>
-  R median(final HTable table, ColumnInterpreter<R, S, P, Q, T> ci,
+  R median(final Table table, ColumnInterpreter<R, S, P, Q, T> ci,
       Scan scan) throws Throwable {
     Pair<NavigableMap<byte[], List<S>>, List<S>> p = getMedianArgs(table, ci, scan);
     byte[] startRow = null;

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/SecureBulkLoadClient.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/SecureBulkLoadClient.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/SecureBulkLoadClient.java
index 48986b1..c4ce866 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/SecureBulkLoadClient.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/SecureBulkLoadClient.java
@@ -21,12 +21,12 @@ package org.apache.hadoop.hbase.client.coprocessor;
 import static org.apache.hadoop.hbase.HConstants.EMPTY_START_ROW;
 import static org.apache.hadoop.hbase.HConstants.LAST_ROW;
 
+import org.apache.hadoop.hbase.client.Table;
 import org.apache.hadoop.hbase.util.ByteStringer;
 
 import org.apache.hadoop.classification.InterfaceAudience;
 import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.hbase.TableName;
-import org.apache.hadoop.hbase.client.HTable;
 import org.apache.hadoop.hbase.ipc.BlockingRpcCallback;
 import org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel;
 import org.apache.hadoop.hbase.ipc.ServerRpcController;
@@ -47,9 +47,9 @@ import java.util.List;
  */
 @InterfaceAudience.Private
 public class SecureBulkLoadClient {
-  private HTable table;
+  private Table table;
 
-  public SecureBulkLoadClient(HTable table) {
+  public SecureBulkLoadClient(Table table) {
     this.table = table;
   }
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlClient.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlClient.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlClient.java
index 35c1412..87db23d 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlClient.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlClient.java
@@ -32,8 +32,10 @@ import org.apache.hadoop.hbase.MasterNotRunningException;
 import org.apache.hadoop.hbase.NamespaceDescriptor;
 import org.apache.hadoop.hbase.TableName;
 import org.apache.hadoop.hbase.ZooKeeperConnectionException;
+import org.apache.hadoop.hbase.client.Admin;
 import org.apache.hadoop.hbase.client.HBaseAdmin;
 import org.apache.hadoop.hbase.client.HTable;
+import org.apache.hadoop.hbase.client.Table;
 import org.apache.hadoop.hbase.client.coprocessor.Batch;
 import org.apache.hadoop.hbase.ipc.BlockingRpcCallback;
 import org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel;
@@ -70,7 +72,7 @@ public class AccessControlClient {
   public static GrantResponse grant(Configuration conf, final TableName tableName,
       final String userName, final byte[] family, final byte[] qual,
       final AccessControlProtos.Permission.Action... actions) throws Throwable {
-    HTable ht = null;
+    Table ht = null;
     try {
       TableName aclTableName =
           TableName.valueOf(NamespaceDescriptor.SYSTEM_NAMESPACE_NAME_STR, "acl");
@@ -150,7 +152,7 @@ public class AccessControlClient {
   public static RevokeResponse revoke(Configuration conf, final String username,
       final TableName tableName, final byte[] family, final byte[] qualifier,
       final AccessControlProtos.Permission.Action... actions) throws Throwable {
-    HTable ht = null;
+    Table ht = null;
     try {
       TableName aclTableName = TableName.valueOf(NamespaceDescriptor.SYSTEM_NAMESPACE_NAME_STR,
           "acl");
@@ -211,8 +213,8 @@ public class AccessControlClient {
   public static List<UserPermission> getUserPermissions(Configuration conf, String tableRegex)
       throws Throwable {
     List<UserPermission> permList = new ArrayList<UserPermission>();
-    HTable ht = null;
-    HBaseAdmin ha = null;
+    Table ht = null;
+    Admin ha = null;
     try {
       TableName aclTableName =
           TableName.valueOf(NamespaceDescriptor.SYSTEM_NAMESPACE_NAME_STR, "acl");

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityClient.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityClient.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityClient.java
index 8a17994..b1f9c91 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityClient.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityClient.java
@@ -22,6 +22,7 @@ import static org.apache.hadoop.hbase.security.visibility.VisibilityConstants.LA
 import java.io.IOException;
 import java.util.Map;
 
+import org.apache.hadoop.hbase.client.Table;
 import org.apache.hadoop.hbase.util.ByteStringer;
 import org.apache.hadoop.classification.InterfaceAudience;
 import org.apache.hadoop.classification.InterfaceStability;
@@ -72,7 +73,7 @@ public class VisibilityClient {
    */
   public static VisibilityLabelsResponse addLabels(Configuration conf, final String[] labels)
       throws Throwable {
-    HTable ht = null;
+    Table ht = null;
     try {
       ht = new HTable(conf, LABELS_TABLE_NAME.getName());
       Batch.Call<VisibilityLabelsService, VisibilityLabelsResponse> callable = 
@@ -126,7 +127,7 @@ public class VisibilityClient {
    * @throws Throwable
    */
   public static GetAuthsResponse getAuths(Configuration conf, final String user) throws Throwable {
-    HTable ht = null;
+    Table ht = null;
     try {
       ht = new HTable(conf, LABELS_TABLE_NAME.getName());
       Batch.Call<VisibilityLabelsService, GetAuthsResponse> callable = 
@@ -168,7 +169,7 @@ public class VisibilityClient {
 
   private static VisibilityLabelsResponse setOrClearAuths(Configuration conf, final String[] auths,
       final String user, final boolean setOrClear) throws IOException, ServiceException, Throwable {
-    HTable ht = null;
+    Table ht = null;
     try {
       ht = new HTable(conf, LABELS_TABLE_NAME.getName());
       Batch.Call<VisibilityLabelsService, VisibilityLabelsResponse> callable = 

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java b/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java
index 9e878b4..df5e693 100644
--- a/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java
+++ b/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java
@@ -150,7 +150,7 @@ public class TestClientNoCluster extends Configured implements Tool {
     Configuration localConfig = HBaseConfiguration.create(this.conf);
     // This override mocks up our exists/get call to throw a RegionServerStoppedException.
     localConfig.set("hbase.client.connection.impl", RpcTimeoutConnection.class.getName());
-    HTable table = new HTable(localConfig, TableName.META_TABLE_NAME);
+    Table table = new HTable(localConfig, TableName.META_TABLE_NAME);
     Throwable t = null;
     LOG.info("Start");
     try {
@@ -187,7 +187,7 @@ public class TestClientNoCluster extends Configured implements Tool {
     // and it has expired.  Otherwise, if this functionality is broke, all retries will be run --
     // all ten of them -- and we'll get the RetriesExhaustedException exception.
     localConfig.setInt(HConstants.HBASE_CLIENT_META_OPERATION_TIMEOUT, pause - 1);
-    HTable table = new HTable(localConfig, TableName.META_TABLE_NAME);
+    Table table = new HTable(localConfig, TableName.META_TABLE_NAME);
     Throwable t = null;
     try {
       // An exists call turns into a get w/ a flag.
@@ -219,7 +219,7 @@ public class TestClientNoCluster extends Configured implements Tool {
     // Go against meta else we will try to find first region for the table on construction which
     // means we'll have to do a bunch more mocking.  Tests that go against meta only should be
     // good for a bit of testing.
-    HTable table = new HTable(this.conf, TableName.META_TABLE_NAME);
+    Table table = new HTable(this.conf, TableName.META_TABLE_NAME);
     ResultScanner scanner = table.getScanner(HConstants.CATALOG_FAMILY);
     try {
       Result result = null;
@@ -239,7 +239,7 @@ public class TestClientNoCluster extends Configured implements Tool {
     // Go against meta else we will try to find first region for the table on construction which
     // means we'll have to do a bunch more mocking.  Tests that go against meta only should be
     // good for a bit of testing.
-    HTable table = new HTable(this.conf, TableName.META_TABLE_NAME);
+    Table table = new HTable(this.conf, TableName.META_TABLE_NAME);
     ResultScanner scanner = table.getScanner(HConstants.CATALOG_FAMILY);
     try {
       Result result = null;
@@ -700,7 +700,7 @@ public class TestClientNoCluster extends Configured implements Tool {
    * @throws IOException
    */
   static void cycle(int id, final Configuration c, final HConnection sharedConnection) throws IOException {
-    HTableInterface table = sharedConnection.getTable(BIG_USER_TABLE);
+    Table table = sharedConnection.getTable(BIG_USER_TABLE);
     table.setAutoFlushTo(false);
     long namespaceSpan = c.getLong("hbase.test.namespace.span", 1000000);
     long startTime = System.currentTimeMillis();

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromAdmin.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromAdmin.java b/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromAdmin.java
index 1619467..572128a 100644
--- a/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromAdmin.java
+++ b/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromAdmin.java
@@ -99,7 +99,7 @@ public class TestSnapshotFromAdmin {
       builder.build(), builder.build(), builder.build(), builder.setDone(true).build());
 
     // setup the admin and run the test
-    HBaseAdmin admin = new HBaseAdmin(mockConnection);
+    Admin admin = new HBaseAdmin(mockConnection);
     String snapshot = "snapshot";
     TableName table = TableName.valueOf("table");
     // get start time
@@ -122,7 +122,7 @@ public class TestSnapshotFromAdmin {
         .mock(ConnectionManager.HConnectionImplementation.class);
     Configuration conf = HBaseConfiguration.create();
     Mockito.when(mockConnection.getConfiguration()).thenReturn(conf);
-    HBaseAdmin admin = new HBaseAdmin(mockConnection);
+    Admin admin = new HBaseAdmin(mockConnection);
     SnapshotDescription.Builder builder = SnapshotDescription.newBuilder();
     // check that invalid snapshot names fail
     failSnapshotStart(admin, builder.setName(HConstants.SNAPSHOT_DIR_NAME).build());
@@ -152,7 +152,7 @@ public class TestSnapshotFromAdmin {
     admin.snapshot(builder.setName("snapshot").setTable("table").build());
   }
 
-  private void failSnapshotStart(HBaseAdmin admin, SnapshotDescription snapshot) throws IOException {
+  private void failSnapshotStart(Admin admin, SnapshotDescription snapshot) throws IOException {
     try {
       admin.snapshot(snapshot);
       fail("Snapshot should not have succeed with name:" + snapshot.getName());

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestBulkDeleteProtocol.java
----------------------------------------------------------------------
diff --git a/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestBulkDeleteProtocol.java b/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestBulkDeleteProtocol.java
index 8dbb16c..16327ba 100644
--- a/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestBulkDeleteProtocol.java
+++ b/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestBulkDeleteProtocol.java
@@ -37,6 +37,7 @@ import org.apache.hadoop.hbase.client.HTable;
 import org.apache.hadoop.hbase.client.Put;
 import org.apache.hadoop.hbase.client.Result;
 import org.apache.hadoop.hbase.client.Scan;
+import org.apache.hadoop.hbase.client.Table;
 import org.apache.hadoop.hbase.client.coprocessor.Batch;
 import org.apache.hadoop.hbase.coprocessor.CoprocessorHost;
 import org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest;
@@ -78,7 +79,7 @@ public class TestBulkDeleteProtocol {
   // @Ignore @Test
   public void testBulkDeleteEndpoint() throws Throwable {
     byte[] tableName = Bytes.toBytes("testBulkDeleteEndpoint");
-    HTable ht = createTable(tableName);
+    Table ht = createTable(tableName);
     List<Put> puts = new ArrayList<Put>(100);
     for (int j = 0; j < 100; j++) {
       byte[] rowkey = Bytes.toBytes(j);
@@ -102,7 +103,7 @@ public class TestBulkDeleteProtocol {
       throws Throwable {
     byte[] tableName = Bytes
         .toBytes("testBulkDeleteEndpointWhenRowBatchSizeLessThanRowsToDeleteFromARegion");
-    HTable ht = createTable(tableName);
+    Table ht = createTable(tableName);
     List<Put> puts = new ArrayList<Put>(100);
     for (int j = 0; j < 100; j++) {
       byte[] rowkey = Bytes.toBytes(j);
@@ -123,7 +124,7 @@ public class TestBulkDeleteProtocol {
 
   private long invokeBulkDeleteProtocol(byte[] tableName, final Scan scan, final int rowBatchSize,
       final DeleteType deleteType, final Long timeStamp) throws Throwable {
-    HTable ht = new HTable(TEST_UTIL.getConfiguration(), tableName);
+    Table ht = new HTable(TEST_UTIL.getConfiguration(), tableName);
     long noOfDeletedRows = 0L;
     Batch.Call<BulkDeleteService, BulkDeleteResponse> callable =
       new Batch.Call<BulkDeleteService, BulkDeleteResponse>() {
@@ -155,7 +156,7 @@ public class TestBulkDeleteProtocol {
   // @Ignore @Test
   public void testBulkDeleteWithConditionBasedDelete() throws Throwable {
     byte[] tableName = Bytes.toBytes("testBulkDeleteWithConditionBasedDelete");
-    HTable ht = createTable(tableName);
+    Table ht = createTable(tableName);
     List<Put> puts = new ArrayList<Put>(100);
     for (int j = 0; j < 100; j++) {
       byte[] rowkey = Bytes.toBytes(j);
@@ -185,7 +186,7 @@ public class TestBulkDeleteProtocol {
   // @Ignore @Test
   public void testBulkDeleteColumn() throws Throwable {
     byte[] tableName = Bytes.toBytes("testBulkDeleteColumn");
-    HTable ht = createTable(tableName);
+    Table ht = createTable(tableName);
     List<Put> puts = new ArrayList<Put>(100);
     for (int j = 0; j < 100; j++) {
       byte[] rowkey = Bytes.toBytes(j);
@@ -218,7 +219,7 @@ public class TestBulkDeleteProtocol {
     htd.addFamily(new HColumnDescriptor(FAMILY1));
     htd.addFamily(new HColumnDescriptor(FAMILY2));
     TEST_UTIL.getHBaseAdmin().createTable(htd, Bytes.toBytes(0), Bytes.toBytes(120), 5);
-    HTable ht = new HTable(TEST_UTIL.getConfiguration(), tableName);
+    Table ht = new HTable(TEST_UTIL.getConfiguration(), tableName);
     List<Put> puts = new ArrayList<Put>(100);
     for (int j = 0; j < 100; j++) {
       Put put = new Put(Bytes.toBytes(j));
@@ -245,7 +246,7 @@ public class TestBulkDeleteProtocol {
   // @Ignore @Test
   public void testBulkDeleteColumnVersion() throws Throwable {
     byte[] tableName = Bytes.toBytes("testBulkDeleteColumnVersion");
-    HTable ht = createTable(tableName);
+    Table ht = createTable(tableName);
     List<Put> puts = new ArrayList<Put>(100);
     for (int j = 0; j < 100; j++) {
       Put put = new Put(Bytes.toBytes(j));
@@ -293,7 +294,7 @@ public class TestBulkDeleteProtocol {
   // @Ignore @Test
   public void testBulkDeleteColumnVersionBasedOnTS() throws Throwable {
     byte[] tableName = Bytes.toBytes("testBulkDeleteColumnVersionBasedOnTS");
-    HTable ht = createTable(tableName);
+    Table ht = createTable(tableName);
     List<Put> puts = new ArrayList<Put>(100);
     for (int j = 0; j < 100; j++) {
       Put put = new Put(Bytes.toBytes(j));
@@ -340,7 +341,7 @@ public class TestBulkDeleteProtocol {
   // @Ignore @Test
   public void testBulkDeleteWithNumberOfVersions() throws Throwable {
     byte[] tableName = Bytes.toBytes("testBulkDeleteWithNumberOfVersions");
-    HTable ht = createTable(tableName);
+    Table ht = createTable(tableName);
     List<Put> puts = new ArrayList<Put>(100);
     for (int j = 0; j < 100; j++) {
       Put put = new Put(Bytes.toBytes(j));
@@ -422,13 +423,13 @@ public class TestBulkDeleteProtocol {
     ht.close();
   }
 
-  private HTable createTable(byte[] tableName) throws IOException {
+  private Table createTable(byte[] tableName) throws IOException {
     HTableDescriptor htd = new HTableDescriptor(TableName.valueOf(tableName));
     HColumnDescriptor hcd = new HColumnDescriptor(FAMILY1);
     hcd.setMaxVersions(10);// Just setting 10 as I am not testing with more than 10 versions here
     htd.addFamily(hcd);
     TEST_UTIL.getHBaseAdmin().createTable(htd, Bytes.toBytes(0), Bytes.toBytes(120), 5);
-    HTable ht = new HTable(TEST_UTIL.getConfiguration(), tableName);
+    Table ht = new HTable(TEST_UTIL.getConfiguration(), tableName);
     return ht;
   }
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestRowCountEndpoint.java
----------------------------------------------------------------------
diff --git a/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestRowCountEndpoint.java b/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestRowCountEndpoint.java
index fb04b4d..cc16483 100644
--- a/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestRowCountEndpoint.java
+++ b/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestRowCountEndpoint.java
@@ -23,16 +23,13 @@ import org.apache.hadoop.hbase.HBaseTestingUtility;
 import org.apache.hadoop.hbase.MediumTests;
 import org.apache.hadoop.hbase.client.HTable;
 import org.apache.hadoop.hbase.client.Put;
+import org.apache.hadoop.hbase.client.Table;
 import org.apache.hadoop.hbase.client.coprocessor.Batch;
 import org.apache.hadoop.hbase.coprocessor.CoprocessorHost;
 import org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos;
 import org.apache.hadoop.hbase.ipc.BlockingRpcCallback;
 import org.apache.hadoop.hbase.ipc.ServerRpcController;
 import org.apache.hadoop.hbase.util.Bytes;
-import org.junit.AfterClass;
-import org.junit.BeforeClass;
-import org.junit.Test;
-import org.junit.Ignore;
 import org.junit.experimental.categories.Category;
 
 import java.io.IOException;
@@ -72,7 +69,7 @@ public class TestRowCountEndpoint {
 
   // @Ignore @Test
   public void testEndpoint() throws Throwable {
-    HTable table = new HTable(CONF, TEST_TABLE);
+    Table table = new HTable(CONF, TEST_TABLE);
 
     // insert some test rows
     for (int i=0; i<5; i++) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestZooKeeperScanPolicyObserver.java
----------------------------------------------------------------------
diff --git a/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestZooKeeperScanPolicyObserver.java b/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestZooKeeperScanPolicyObserver.java
index 824910a..af51504 100644
--- a/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestZooKeeperScanPolicyObserver.java
+++ b/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestZooKeeperScanPolicyObserver.java
@@ -32,6 +32,7 @@ import org.apache.hadoop.hbase.client.Get;
 import org.apache.hadoop.hbase.client.HTable;
 import org.apache.hadoop.hbase.client.Put;
 import org.apache.hadoop.hbase.client.Result;
+import org.apache.hadoop.hbase.client.Table;
 import org.apache.hadoop.hbase.coprocessor.CoprocessorHost;
 import org.apache.hadoop.hbase.util.Bytes;
 import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
@@ -74,7 +75,7 @@ public class TestZooKeeperScanPolicyObserver {
     .setTimeToLive(1);
     desc.addFamily(hcd);
     TEST_UTIL.getHBaseAdmin().createTable(desc);
-    HTable t = new HTable(new Configuration(TEST_UTIL.getConfiguration()), tableName);
+    Table t = new HTable(new Configuration(TEST_UTIL.getConfiguration()), tableName);
     long now = EnvironmentEdgeManager.currentTime();
 
     ZooKeeperWatcher zkw = new ZooKeeperWatcher(TEST_UTIL.getConfiguration(), "test", null);

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-it/src/test/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java
index 9ae00f9..f1f5f93 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java
@@ -25,6 +25,7 @@ import java.util.List;
 import org.apache.hadoop.classification.InterfaceAudience;
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.hbase.ClusterManager.ServiceType;
+import org.apache.hadoop.hbase.client.Admin;
 import org.apache.hadoop.hbase.client.HBaseAdmin;
 import org.apache.hadoop.hbase.client.HConnection;
 import org.apache.hadoop.hbase.client.HConnectionManager;
@@ -45,7 +46,7 @@ import com.google.common.collect.Sets;
 @InterfaceAudience.Private
 public class DistributedHBaseCluster extends HBaseCluster {
 
-  private HBaseAdmin admin;
+  private Admin admin;
 
   private ClusterManager clusterManager;
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestLazyCfLoading.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestLazyCfLoading.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestLazyCfLoading.java
index a1e306d..750376a 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestLazyCfLoading.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestLazyCfLoading.java
@@ -31,6 +31,7 @@ import org.apache.hadoop.hbase.client.HTable;
 import org.apache.hadoop.hbase.client.Result;
 import org.apache.hadoop.hbase.client.ResultScanner;
 import org.apache.hadoop.hbase.client.Scan;
+import org.apache.hadoop.hbase.client.Table;
 import org.apache.hadoop.hbase.filter.CompareFilter;
 import org.apache.hadoop.hbase.filter.Filter;
 import org.apache.hadoop.hbase.filter.SingleColumnValueFilter;
@@ -222,7 +223,7 @@ public class IntegrationTestLazyCfLoading {
     long maxRuntime = conf.getLong(timeoutKey, DEFAULT_TIMEOUT_MINUTES);
     long serverCount = util.getHBaseClusterInterface().getClusterStatus().getServersSize();
     long keysToWrite = serverCount * KEYS_TO_WRITE_PER_SERVER;
-    HTable table = new HTable(conf, TABLE_NAME);
+    Table table = new HTable(conf, TABLE_NAME);
 
     // Create multi-threaded writer and start it. We write multiple columns/CFs and verify
     // their integrity, therefore multi-put is necessary.

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-it/src/test/java/org/apache/hadoop/hbase/mapreduce/IntegrationTestImportTsv.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/mapreduce/IntegrationTestImportTsv.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/mapreduce/IntegrationTestImportTsv.java
index 766c66f..e99677b 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/mapreduce/IntegrationTestImportTsv.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/mapreduce/IntegrationTestImportTsv.java
@@ -48,6 +48,7 @@ import org.apache.hadoop.hbase.TableName;
 import org.apache.hadoop.hbase.client.HTable;
 import org.apache.hadoop.hbase.client.Result;
 import org.apache.hadoop.hbase.client.Scan;
+import org.apache.hadoop.hbase.client.Table;
 import org.apache.hadoop.hbase.util.Bytes;
 import org.apache.hadoop.io.LongWritable;
 import org.apache.hadoop.io.Text;
@@ -153,7 +154,7 @@ public class IntegrationTestImportTsv implements Configurable, Tool {
     assertEquals("Loading HFiles failed.",
       0, ToolRunner.run(new LoadIncrementalHFiles(new Configuration(getConf())), args));
 
-    HTable table = null;
+    Table table = null;
     Scan scan = new Scan() {{
       setCacheBlocks(false);
       setCaching(1000);

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-it/src/test/java/org/apache/hadoop/hbase/mttr/IntegrationTestMTTR.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/mttr/IntegrationTestMTTR.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/mttr/IntegrationTestMTTR.java
index 65e1026..55cc70e 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/mttr/IntegrationTestMTTR.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/mttr/IntegrationTestMTTR.java
@@ -49,6 +49,7 @@ import org.apache.hadoop.hbase.chaos.actions.RestartActiveMasterAction;
 import org.apache.hadoop.hbase.chaos.actions.RestartRsHoldingMetaAction;
 import org.apache.hadoop.hbase.chaos.actions.RestartRsHoldingTableAction;
 import org.apache.hadoop.hbase.chaos.factories.MonkeyConstants;
+import org.apache.hadoop.hbase.client.Admin;
 import org.apache.hadoop.hbase.client.HBaseAdmin;
 import org.apache.hadoop.hbase.client.HTable;
 import org.apache.hadoop.hbase.client.Put;
@@ -56,6 +57,7 @@ import org.apache.hadoop.hbase.client.Result;
 import org.apache.hadoop.hbase.client.ResultScanner;
 import org.apache.hadoop.hbase.client.RetriesExhaustedException;
 import org.apache.hadoop.hbase.client.Scan;
+import org.apache.hadoop.hbase.client.Table;
 import org.apache.hadoop.hbase.coprocessor.CoprocessorException;
 import org.apache.hadoop.hbase.filter.KeyOnlyFilter;
 import org.apache.hadoop.hbase.ipc.FatalConnectionException;
@@ -461,7 +463,7 @@ public class IntegrationTestMTTR {
    */
   static class PutCallable extends TimingCallable {
 
-    private final HTable table;
+    private final Table table;
 
     public PutCallable(Future<?> f) throws IOException {
       super(f);
@@ -488,7 +490,7 @@ public class IntegrationTestMTTR {
    * supplied future returns.  Returns the max time taken to scan.
    */
   static class ScanCallable extends TimingCallable {
-    private final HTable table;
+    private final Table table;
 
     public ScanCallable(Future<?> f) throws IOException {
       super(f);
@@ -531,7 +533,7 @@ public class IntegrationTestMTTR {
 
     @Override
     protected boolean doAction() throws Exception {
-      HBaseAdmin admin = null;
+      Admin admin = null;
       try {
         admin = new HBaseAdmin(util.getConfiguration());
         ClusterStatus status = admin.getClusterStatus();

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedList.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedList.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedList.java
index 3adef26..c709f0d 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedList.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedList.java
@@ -50,6 +50,7 @@ import org.apache.hadoop.hbase.IntegrationTestingUtility;
 import org.apache.hadoop.hbase.IntegrationTests;
 import org.apache.hadoop.hbase.MasterNotRunningException;
 import org.apache.hadoop.hbase.TableName;
+import org.apache.hadoop.hbase.client.Admin;
 import org.apache.hadoop.hbase.client.Get;
 import org.apache.hadoop.hbase.client.HBaseAdmin;
 import org.apache.hadoop.hbase.client.HConnection;
@@ -60,6 +61,7 @@ import org.apache.hadoop.hbase.client.Result;
 import org.apache.hadoop.hbase.client.ResultScanner;
 import org.apache.hadoop.hbase.client.Scan;
 import org.apache.hadoop.hbase.client.ScannerCallable;
+import org.apache.hadoop.hbase.client.Table;
 import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
 import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil;
 import org.apache.hadoop.hbase.mapreduce.TableMapper;
@@ -448,7 +450,7 @@ public class IntegrationTestBigLinkedList extends IntegrationTestBase {
 
     protected void createSchema() throws IOException {
       Configuration conf = getConf();
-      HBaseAdmin admin = new HBaseAdmin(conf);
+      Admin admin = new HBaseAdmin(conf);
       TableName tableName = getTableName(conf);
       try {
         if (!admin.tableExists(tableName)) {
@@ -873,7 +875,7 @@ public class IntegrationTestBigLinkedList extends IntegrationTestBase {
         System.exit(-1);
       }
 
-      HTable table = new HTable(getConf(), getTableName(getConf()));
+      Table table = new HTable(getConf(), getTableName(getConf()));
 
       Scan scan = new Scan();
       scan.setBatch(10000);
@@ -923,7 +925,7 @@ public class IntegrationTestBigLinkedList extends IntegrationTestBase {
       org.apache.hadoop.hbase.client.Delete delete
         = new org.apache.hadoop.hbase.client.Delete(val);
 
-      HTable table = new HTable(getConf(), getTableName(getConf()));
+      Table table = new HTable(getConf(), getTableName(getConf()));
 
       table.delete(delete);
       table.flushCommits();
@@ -969,7 +971,7 @@ public class IntegrationTestBigLinkedList extends IntegrationTestBase {
       byte[] startKey = isSpecificStart ? Bytes.toBytesBinary(cmd.getOptionValue('s')) : null;
       int logEvery = cmd.hasOption('l') ? Integer.parseInt(cmd.getOptionValue('l')) : 1;
 
-      HTable table = new HTable(getConf(), getTableName(getConf()));
+      Table table = new HTable(getConf(), getTableName(getConf()));
       long numQueries = 0;
       // If isSpecificStart is set, only walk one list from that particular node.
       // Note that in case of circular (or P-shaped) list it will walk forever, as is
@@ -1005,7 +1007,7 @@ public class IntegrationTestBigLinkedList extends IntegrationTestBase {
       return 0;
     }
 
-    private static CINode findStartNode(HTable table, byte[] startKey) throws IOException {
+    private static CINode findStartNode(Table table, byte[] startKey) throws IOException {
       Scan scan = new Scan();
       scan.setStartRow(startKey);
       scan.setBatch(1);
@@ -1028,7 +1030,7 @@ public class IntegrationTestBigLinkedList extends IntegrationTestBase {
       return null;
     }
 
-    private CINode getNode(byte[] row, HTable table, CINode node) throws IOException {
+    private CINode getNode(byte[] row, Table table, CINode node) throws IOException {
       Get get = new Get(row);
       get.addColumn(FAMILY_NAME, COLUMN_PREV);
       Result result = table.get(get);

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedListWithVisibility.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedListWithVisibility.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedListWithVisibility.java
index 9748b31..be7e36f 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedListWithVisibility.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedListWithVisibility.java
@@ -38,6 +38,7 @@ import org.apache.hadoop.hbase.IntegrationTestingUtility;
 import org.apache.hadoop.hbase.IntegrationTests;
 import org.apache.hadoop.hbase.TableName;
 import org.apache.hadoop.hbase.chaos.factories.MonkeyFactory;
+import org.apache.hadoop.hbase.client.Admin;
 import org.apache.hadoop.hbase.client.Delete;
 import org.apache.hadoop.hbase.client.HBaseAdmin;
 import org.apache.hadoop.hbase.client.HConnection;
@@ -46,6 +47,7 @@ import org.apache.hadoop.hbase.client.HTable;
 import org.apache.hadoop.hbase.client.Put;
 import org.apache.hadoop.hbase.client.Result;
 import org.apache.hadoop.hbase.client.Scan;
+import org.apache.hadoop.hbase.client.Table;
 import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
 import org.apache.hadoop.hbase.io.hfile.HFile;
 import org.apache.hadoop.hbase.mapreduce.Import;
@@ -130,7 +132,7 @@ public class IntegrationTestBigLinkedListWithVisibility extends IntegrationTestB
       if(!acl) {
         LOG.info("No ACL available.");
       }
-      HBaseAdmin admin = new HBaseAdmin(getConf());
+      Admin admin = new HBaseAdmin(getConf());
       for (int i = 0; i < DEFAULT_TABLES_COUNT; i++) {
         TableName tableName = IntegrationTestBigLinkedListWithVisibility.getTableName(i);
         createTable(admin, tableName, false, acl);
@@ -140,7 +142,7 @@ public class IntegrationTestBigLinkedListWithVisibility extends IntegrationTestB
       admin.close();
     }
 
-    private void createTable(HBaseAdmin admin, TableName tableName, boolean setVersion, 
+    private void createTable(Admin admin, TableName tableName, boolean setVersion,
         boolean acl) throws IOException {
       if (!admin.tableExists(tableName)) {
         HTableDescriptor htd = new HTableDescriptor(tableName);
@@ -170,8 +172,8 @@ public class IntegrationTestBigLinkedListWithVisibility extends IntegrationTestB
     }
 
     static class VisibilityGeneratorMapper extends GeneratorMapper {
-      HTable[] tables = new HTable[DEFAULT_TABLES_COUNT];
-      HTable commonTable = null;
+      Table[] tables = new Table[DEFAULT_TABLES_COUNT];
+      Table commonTable = null;
 
       @Override
       protected void setup(org.apache.hadoop.mapreduce.Mapper.Context context) throws IOException,

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestLoadAndVerify.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestLoadAndVerify.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestLoadAndVerify.java
index 0da5107..6bab237 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestLoadAndVerify.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestLoadAndVerify.java
@@ -448,7 +448,7 @@ public void cleanUpCluster() throws Exception {
     HTableDescriptor htd = new HTableDescriptor(table);
     htd.addFamily(new HColumnDescriptor(TEST_FAMILY));
 
-    HBaseAdmin admin = new HBaseAdmin(getConf());
+    Admin admin = new HBaseAdmin(getConf());
     if (doLoad) {
       admin.createTable(htd, Bytes.toBytes(0L), Bytes.toBytes(-1L), numPresplits);
       doLoad(getConf(), htd);

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestTimeBoundedRequestsWithRegionReplicas.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestTimeBoundedRequestsWithRegionReplicas.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestTimeBoundedRequestsWithRegionReplicas.java
index 5ca0e36..63bd42f 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestTimeBoundedRequestsWithRegionReplicas.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestTimeBoundedRequestsWithRegionReplicas.java
@@ -39,8 +39,8 @@ import org.apache.hadoop.hbase.chaos.factories.MonkeyFactory;
 import org.apache.hadoop.hbase.client.Admin;
 import org.apache.hadoop.hbase.client.Consistency;
 import org.apache.hadoop.hbase.client.Get;
-import org.apache.hadoop.hbase.client.HTableInterface;
 import org.apache.hadoop.hbase.client.Result;
+import org.apache.hadoop.hbase.client.Table;
 import org.apache.hadoop.hbase.regionserver.StorefileRefresherChore;
 import org.apache.hadoop.hbase.util.LoadTestTool;
 import org.apache.hadoop.hbase.util.MultiThreadedReader;
@@ -326,7 +326,7 @@ public class IntegrationTestTimeBoundedRequestsWithRegionReplicas extends Integr
 
       @Override
       protected void verifyResultsAndUpdateMetrics(boolean verify, Get[] gets, long elapsedNano,
-          Result[] results, HTableInterface table, boolean isNullExpected)
+          Result[] results, Table table, boolean isNullExpected)
           throws IOException {
         super.verifyResultsAndUpdateMetrics(verify, gets, elapsedNano, results, table, isNullExpected);
         for (Result r : results) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestWithCellVisibilityLoadAndVerify.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestWithCellVisibilityLoadAndVerify.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestWithCellVisibilityLoadAndVerify.java
index f4e4250..fc3ae3c 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestWithCellVisibilityLoadAndVerify.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestWithCellVisibilityLoadAndVerify.java
@@ -33,6 +33,7 @@ import org.apache.hadoop.hbase.HConstants;
 import org.apache.hadoop.hbase.HTableDescriptor;
 import org.apache.hadoop.hbase.IntegrationTestingUtility;
 import org.apache.hadoop.hbase.IntegrationTests;
+import org.apache.hadoop.hbase.client.Admin;
 import org.apache.hadoop.hbase.client.HBaseAdmin;
 import org.apache.hadoop.hbase.client.Put;
 import org.apache.hadoop.hbase.client.Result;
@@ -369,7 +370,7 @@ public class IntegrationTestWithCellVisibilityLoadAndVerify extends IntegrationT
     HTableDescriptor htd = new HTableDescriptor(getTablename());
     htd.addFamily(new HColumnDescriptor(TEST_FAMILY));
 
-    HBaseAdmin admin = new HBaseAdmin(getConf());
+    Admin admin = new HBaseAdmin(getConf());
     try {
       admin.createTable(htd, Bytes.toBytes(0L), Bytes.toBytes(-1L), numPresplits);
     } finally {

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-it/src/test/java/org/apache/hadoop/hbase/trace/IntegrationTestSendTraceRequests.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/trace/IntegrationTestSendTraceRequests.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/trace/IntegrationTestSendTraceRequests.java
index 7961eb9..2ec5838 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/trace/IntegrationTestSendTraceRequests.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/trace/IntegrationTestSendTraceRequests.java
@@ -31,6 +31,7 @@ import org.apache.hadoop.hbase.client.Put;
 import org.apache.hadoop.hbase.client.Result;
 import org.apache.hadoop.hbase.client.ResultScanner;
 import org.apache.hadoop.hbase.client.Scan;
+import org.apache.hadoop.hbase.client.Table;
 import org.apache.hadoop.hbase.util.AbstractHBaseTool;
 import org.apache.hadoop.hbase.util.Bytes;
 import org.apache.hadoop.util.ToolRunner;
@@ -123,7 +124,7 @@ public class IntegrationTestSendTraceRequests extends AbstractHBaseTool {
             ResultScanner rs = null;
             try {
               innerScope = Trace.startSpan("Scan", Sampler.ALWAYS);
-              HTable ht = new HTable(util.getConfiguration(), tableName);
+              Table ht = new HTable(util.getConfiguration(), tableName);
               Scan s = new Scan();
               s.setStartRow(Bytes.toBytes(rowKeyQueue.take()));
               s.setBatch(7);
@@ -171,7 +172,7 @@ public class IntegrationTestSendTraceRequests extends AbstractHBaseTool {
         public void run() {
 
 
-          HTable ht = null;
+          Table ht = null;
           try {
             ht = new HTable(util.getConfiguration(), tableName);
           } catch (IOException e) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/LocalHBaseCluster.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/LocalHBaseCluster.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/LocalHBaseCluster.java
index a30819d..d4a87d3 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/LocalHBaseCluster.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/LocalHBaseCluster.java
@@ -29,6 +29,7 @@ import org.apache.commons.logging.LogFactory;
 import org.apache.hadoop.classification.InterfaceAudience;
 import org.apache.hadoop.classification.InterfaceStability;
 import org.apache.hadoop.conf.Configuration;
+import org.apache.hadoop.hbase.client.Admin;
 import org.apache.hadoop.hbase.client.HBaseAdmin;
 import org.apache.hadoop.hbase.regionserver.HRegionServer;
 import org.apache.hadoop.hbase.security.User;
@@ -462,7 +463,7 @@ public class LocalHBaseCluster {
     Configuration conf = HBaseConfiguration.create();
     LocalHBaseCluster cluster = new LocalHBaseCluster(conf);
     cluster.startup();
-    HBaseAdmin admin = new HBaseAdmin(conf);
+    Admin admin = new HBaseAdmin(conf);
     HTableDescriptor htd =
       new HTableDescriptor(TableName.valueOf(cluster.getClass().getName()));
     admin.createTable(htd);

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.java
index ab76cd4..155990e 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.java
@@ -47,6 +47,7 @@ import org.apache.hadoop.hbase.TableName;
 import org.apache.hadoop.hbase.client.HTable;
 import org.apache.hadoop.hbase.client.HTableInterface;
 import org.apache.hadoop.hbase.client.HTableWrapper;
+import org.apache.hadoop.hbase.client.Table;
 import org.apache.hadoop.hbase.util.Bytes;
 import org.apache.hadoop.hbase.util.CoprocessorClassLoader;
 import org.apache.hadoop.hbase.util.SortedCopyOnWriteSet;

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/HRegionPartitioner.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/HRegionPartitioner.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/HRegionPartitioner.java
index b6419b8..3af2351 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/HRegionPartitioner.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/HRegionPartitioner.java
@@ -26,6 +26,7 @@ import org.apache.hadoop.classification.InterfaceAudience;
 import org.apache.hadoop.classification.InterfaceStability;
 import org.apache.hadoop.hbase.HBaseConfiguration;
 import org.apache.hadoop.hbase.client.HTable;
+import org.apache.hadoop.hbase.client.RegionLocator;
 import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
 import org.apache.hadoop.hbase.util.Bytes;
 import org.apache.hadoop.mapred.JobConf;
@@ -46,7 +47,7 @@ import org.apache.hadoop.mapred.Partitioner;
 public class HRegionPartitioner<K2,V2>
 implements Partitioner<ImmutableBytesWritable, V2> {
   private static final Log LOG = LogFactory.getLog(HRegionPartitioner.class);
-  private HTable table;
+  private RegionLocator table;
   private byte[][] startKeys;
 
   public void configure(JobConf job) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableOutputFormat.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableOutputFormat.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableOutputFormat.java
index 327e404..e9d3932 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableOutputFormat.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableOutputFormat.java
@@ -28,6 +28,7 @@ import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.hbase.HBaseConfiguration;
 import org.apache.hadoop.hbase.client.HTable;
 import org.apache.hadoop.hbase.client.Put;
+import org.apache.hadoop.hbase.client.Table;
 import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
 import org.apache.hadoop.fs.FileAlreadyExistsException;
 import org.apache.hadoop.mapred.InvalidJobConfException;
@@ -56,14 +57,14 @@ FileOutputFormat<ImmutableBytesWritable, Put> {
    */
   protected static class TableRecordWriter
     implements RecordWriter<ImmutableBytesWritable, Put> {
-    private HTable m_table;
+    private Table m_table;
 
     /**
      * Instantiate a TableRecordWriter with the HBase HClient for writing.
      *
      * @param table
      */
-    public TableRecordWriter(HTable table) {
+    public TableRecordWriter(Table table) {
       m_table = table;
     }
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableRecordReader.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableRecordReader.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableRecordReader.java
index 7713180..a9496a1 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableRecordReader.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableRecordReader.java
@@ -24,6 +24,7 @@ import org.apache.hadoop.classification.InterfaceAudience;
 import org.apache.hadoop.classification.InterfaceStability;
 import org.apache.hadoop.hbase.client.HTable;
 import org.apache.hadoop.hbase.client.Result;
+import org.apache.hadoop.hbase.client.Table;
 import org.apache.hadoop.hbase.filter.Filter;
 import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
 import org.apache.hadoop.mapred.RecordReader;
@@ -62,7 +63,7 @@ implements RecordReader<ImmutableBytesWritable, Result> {
   /**
    * @param htable the {@link HTable} to scan.
    */
-  public void setHTable(HTable htable) {
+  public void setHTable(Table htable) {
     this.recordReaderImpl.setHTable(htable);
   }
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableRecordReaderImpl.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableRecordReaderImpl.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableRecordReaderImpl.java
index f4043f4..7517c1f 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableRecordReaderImpl.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableRecordReaderImpl.java
@@ -30,6 +30,7 @@ import org.apache.hadoop.hbase.client.Result;
 import org.apache.hadoop.hbase.client.ResultScanner;
 import org.apache.hadoop.hbase.client.Scan;
 import org.apache.hadoop.hbase.client.ScannerCallable;
+import org.apache.hadoop.hbase.client.Table;
 import org.apache.hadoop.hbase.filter.Filter;
 import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
 import org.apache.hadoop.hbase.mapreduce.TableInputFormat;
@@ -52,7 +53,7 @@ public class TableRecordReaderImpl {
   private byte [] lastSuccessfulRow;
   private Filter trrRowFilter;
   private ResultScanner scanner;
-  private HTable htable;
+  private Table htable;
   private byte [][] trrInputColumns;
   private long timestamp;
   private int rowcount;
@@ -116,7 +117,7 @@ public class TableRecordReaderImpl {
   /**
    * @param htable the {@link HTable} to scan.
    */
-  public void setHTable(HTable htable) {
+  public void setHTable(Table htable) {
     Configuration conf = htable.getConfiguration();
     logScannerActivity = conf.getBoolean(
       ScannerCallable.LOG_SCANNER_ACTIVITY, false);

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/DefaultVisibilityExpressionResolver.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/DefaultVisibilityExpressionResolver.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/DefaultVisibilityExpressionResolver.java
index 62bfba4..d09601d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/DefaultVisibilityExpressionResolver.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/DefaultVisibilityExpressionResolver.java
@@ -36,6 +36,7 @@ import org.apache.hadoop.hbase.client.HTable;
 import org.apache.hadoop.hbase.client.Result;
 import org.apache.hadoop.hbase.client.ResultScanner;
 import org.apache.hadoop.hbase.client.Scan;
+import org.apache.hadoop.hbase.client.Table;
 import org.apache.hadoop.hbase.security.visibility.Authorizations;
 import org.apache.hadoop.hbase.security.visibility.VisibilityLabelOrdinalProvider;
 import org.apache.hadoop.hbase.security.visibility.VisibilityUtils;
@@ -66,7 +67,7 @@ public class DefaultVisibilityExpressionResolver implements VisibilityExpression
   public void init() {
     // Reading all the labels and ordinal.
     // This scan should be done by user with global_admin previliges.. Ensure that it works
-    HTable labelsTable = null;
+    Table labelsTable = null;
     try {
       labelsTable = new HTable(conf, LABELS_TABLE_NAME);
     } catch (TableNotFoundException e) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat.java
index df063a4..24ca013 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat.java
@@ -29,6 +29,7 @@ import org.apache.hadoop.classification.InterfaceStability;
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.hbase.KeyValue;
 import org.apache.hadoop.hbase.client.HTable;
+import org.apache.hadoop.hbase.client.Table;
 import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
 import org.apache.hadoop.hbase.io.compress.Compression.Algorithm;
 import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding;
@@ -160,7 +161,7 @@ public class HFileOutputFormat extends FileOutputFormat<ImmutableBytesWritable,
   @edu.umd.cs.findbugs.annotations.SuppressWarnings(
       value="RCN_REDUNDANT_NULLCHECK_OF_NONNULL_VALUE")
   @VisibleForTesting
-  static void configureCompression(HTable table, Configuration conf) throws IOException {
+  static void configureCompression(Table table, Configuration conf) throws IOException {
     HFileOutputFormat2.configureCompression(table, conf);
   }
 
@@ -174,7 +175,7 @@ public class HFileOutputFormat extends FileOutputFormat<ImmutableBytesWritable,
    *           on failure to read column family descriptors
    */
   @VisibleForTesting
-  static void configureBlockSize(HTable table, Configuration conf) throws IOException {
+  static void configureBlockSize(Table table, Configuration conf) throws IOException {
     HFileOutputFormat2.configureBlockSize(table, conf);
   }
 
@@ -188,7 +189,7 @@ public class HFileOutputFormat extends FileOutputFormat<ImmutableBytesWritable,
    *           on failure to read column family descriptors
    */
   @VisibleForTesting
-  static void configureBloomType(HTable table, Configuration conf) throws IOException {
+  static void configureBloomType(Table table, Configuration conf) throws IOException {
     HFileOutputFormat2.configureBloomType(table, conf);
   }
 
@@ -202,7 +203,7 @@ public class HFileOutputFormat extends FileOutputFormat<ImmutableBytesWritable,
    *           on failure to read column family descriptors
    */
   @VisibleForTesting
-  static void configureDataBlockEncoding(HTable table,
+  static void configureDataBlockEncoding(Table table,
       Configuration conf) throws IOException {
     HFileOutputFormat2.configureDataBlockEncoding(table, conf);
   }

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java
index e011308..eef81d3 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java
@@ -45,6 +45,8 @@ import org.apache.hadoop.hbase.KeyValue;
 import org.apache.hadoop.hbase.KeyValueUtil;
 import org.apache.hadoop.hbase.client.HTable;
 import org.apache.hadoop.hbase.client.Put;
+import org.apache.hadoop.hbase.client.RegionLocator;
+import org.apache.hadoop.hbase.client.Table;
 import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
 import org.apache.hadoop.hbase.io.compress.Compression;
 import org.apache.hadoop.hbase.io.compress.Compression.Algorithm;
@@ -284,7 +286,7 @@ public class HFileOutputFormat2
    * Return the start keys of all of the regions in this table,
    * as a list of ImmutableBytesWritable.
    */
-  private static List<ImmutableBytesWritable> getRegionStartKeys(HTable table)
+  private static List<ImmutableBytesWritable> getRegionStartKeys(RegionLocator table)
   throws IOException {
     byte[][] byteKeys = table.getStartKeys();
     ArrayList<ImmutableBytesWritable> ret =
@@ -544,7 +546,7 @@ public class HFileOutputFormat2
       value="RCN_REDUNDANT_NULLCHECK_OF_NONNULL_VALUE")
   @VisibleForTesting
   static void configureCompression(
-      HTable table, Configuration conf) throws IOException {
+      Table table, Configuration conf) throws IOException {
     StringBuilder compressionConfigValue = new StringBuilder();
     HTableDescriptor tableDescriptor = table.getTableDescriptor();
     if(tableDescriptor == null){
@@ -578,7 +580,7 @@ public class HFileOutputFormat2
    */
   @VisibleForTesting
   static void configureBlockSize(
-      HTable table, Configuration conf) throws IOException {
+      Table table, Configuration conf) throws IOException {
     StringBuilder blockSizeConfigValue = new StringBuilder();
     HTableDescriptor tableDescriptor = table.getTableDescriptor();
     if (tableDescriptor == null) {
@@ -612,7 +614,7 @@ public class HFileOutputFormat2
    */
   @VisibleForTesting
   static void configureBloomType(
-      HTable table, Configuration conf) throws IOException {
+      Table table, Configuration conf) throws IOException {
     HTableDescriptor tableDescriptor = table.getTableDescriptor();
     if (tableDescriptor == null) {
       // could happen with mock table instance
@@ -647,7 +649,7 @@ public class HFileOutputFormat2
    *           on failure to read column family descriptors
    */
   @VisibleForTesting
-  static void configureDataBlockEncoding(HTable table,
+  static void configureDataBlockEncoding(Table table,
       Configuration conf) throws IOException {
     HTableDescriptor tableDescriptor = table.getTableDescriptor();
     if (tableDescriptor == null) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HRegionPartitioner.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HRegionPartitioner.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HRegionPartitioner.java
index 02727cc..e3e9dd0 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HRegionPartitioner.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HRegionPartitioner.java
@@ -28,6 +28,7 @@ import org.apache.hadoop.conf.Configurable;
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.hbase.HBaseConfiguration;
 import org.apache.hadoop.hbase.client.HTable;
+import org.apache.hadoop.hbase.client.RegionLocator;
 import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
 import org.apache.hadoop.hbase.util.Bytes;
 import org.apache.hadoop.mapreduce.Partitioner;
@@ -53,7 +54,7 @@ implements Configurable {
 
   private static final Log LOG = LogFactory.getLog(HRegionPartitioner.class);
   private Configuration conf = null;
-  private HTable table;
+  private RegionLocator table;
   private byte[][] startKeys;
 
   /**