You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@phoenix.apache.org by an...@apache.org on 2017/11/15 05:59:12 UTC

phoenix git commit: PHOENIX-4376 Fix all compilation error for package org.apache.phoenix.schema.stats and org.apache.phoenix.coprocessor

Repository: phoenix
Updated Branches:
  refs/heads/5.x-HBase-2.0 10a0778eb -> e1238aa3c


PHOENIX-4376 Fix all compilation error for package org.apache.phoenix.schema.stats and org.apache.phoenix.coprocessor


Project: http://git-wip-us.apache.org/repos/asf/phoenix/repo
Commit: http://git-wip-us.apache.org/repos/asf/phoenix/commit/e1238aa3
Tree: http://git-wip-us.apache.org/repos/asf/phoenix/tree/e1238aa3
Diff: http://git-wip-us.apache.org/repos/asf/phoenix/diff/e1238aa3

Branch: refs/heads/5.x-HBase-2.0
Commit: e1238aa3cb8cdf962b70ca8f1ae197285c558891
Parents: 10a0778
Author: Ankit Singhal <an...@gmail.com>
Authored: Wed Nov 15 11:28:57 2017 +0530
Committer: Ankit Singhal <an...@gmail.com>
Committed: Wed Nov 15 11:28:57 2017 +0530

----------------------------------------------------------------------
 .../hadoop/hbase/regionserver/ScanInfoUtil.java | 35 -------------------
 .../coprocessor/BaseScannerRegionObserver.java  | 26 ++------------
 .../coprocessor/DelegateRegionObserver.java     | 32 ++++++++++-------
 .../coprocessor/DelegateRegionScanner.java      |  5 ---
 .../coprocessor/HashJoinRegionScanner.java      |  4 +--
 .../coprocessor/MetaDataEndpointImpl.java       |  4 +--
 .../UngroupedAggregateRegionObserver.java       | 36 +++++++++++---------
 .../stats/DefaultStatisticsCollector.java       |  6 ++--
 .../stats/StatisticsCollectionRunTracker.java   | 20 +++++------
 .../phoenix/schema/stats/StatisticsUtil.java    |  4 +--
 10 files changed, 60 insertions(+), 112 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/phoenix/blob/e1238aa3/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/ScanInfoUtil.java
----------------------------------------------------------------------
diff --git a/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/ScanInfoUtil.java b/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/ScanInfoUtil.java
deleted file mode 100644
index 9885c78..0000000
--- a/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/ScanInfoUtil.java
+++ /dev/null
@@ -1,35 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.hadoop.hbase.regionserver;
-
-import org.apache.hadoop.hbase.KeepDeletedCells;
-
-public class ScanInfoUtil {
-    private ScanInfoUtil() {
-    }
-    
-    public static boolean isKeepDeletedCells(ScanInfo scanInfo) {
-        return scanInfo.getKeepDeletedCells() != KeepDeletedCells.FALSE;
-    }
-    
-    public static ScanInfo cloneScanInfoWithKeepDeletedCells(ScanInfo scanInfo) {
-        return new ScanInfo(scanInfo.getConfiguration(), scanInfo.getFamily(), Math.max(scanInfo.getMinVersions(), 1),
-                    scanInfo.getMaxVersions(), scanInfo.getTtl(), KeepDeletedCells.TRUE,
-                    scanInfo.getTimeToPurgeDeletes(), scanInfo.getComparator());
-    }
-}

http://git-wip-us.apache.org/repos/asf/phoenix/blob/e1238aa3/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/BaseScannerRegionObserver.java
----------------------------------------------------------------------
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/BaseScannerRegionObserver.java b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/BaseScannerRegionObserver.java
index 7c6df8f..ee34a8a 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/BaseScannerRegionObserver.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/BaseScannerRegionObserver.java
@@ -19,11 +19,9 @@ package org.apache.phoenix.coprocessor;
 
 import java.io.IOException;
 import java.util.List;
-import java.util.NavigableSet;
 
 import org.apache.hadoop.hbase.Cell;
 import org.apache.hadoop.hbase.DoNotRetryIOException;
-import org.apache.hadoop.hbase.HConstants;
 import org.apache.hadoop.hbase.NotServingRegionException;
 import org.apache.hadoop.hbase.client.Scan;
 import org.apache.hadoop.hbase.coprocessor.ObserverContext;
@@ -31,15 +29,10 @@ import org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment;
 import org.apache.hadoop.hbase.coprocessor.RegionObserver;
 import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
 import org.apache.hadoop.hbase.io.TimeRange;
-import org.apache.hadoop.hbase.regionserver.KeyValueScanner;
 import org.apache.hadoop.hbase.regionserver.Region;
 import org.apache.hadoop.hbase.regionserver.RegionScanner;
-import org.apache.hadoop.hbase.regionserver.ScanInfo;
-import org.apache.hadoop.hbase.regionserver.ScanInfoUtil;
 import org.apache.hadoop.hbase.regionserver.ScannerContext;
 import org.apache.hadoop.hbase.regionserver.ScannerContextUtil;
-import org.apache.hadoop.hbase.regionserver.Store;
-import org.apache.hadoop.hbase.regionserver.StoreScanner;
 import org.apache.hadoop.hbase.util.Bytes;
 import org.apache.htrace.Span;
 import org.apache.htrace.Trace;
@@ -54,7 +47,6 @@ import org.apache.phoenix.schema.types.PUnsignedTinyint;
 import org.apache.phoenix.util.EncodedColumnsUtil;
 import org.apache.phoenix.util.ScanUtil;
 import org.apache.phoenix.util.ServerUtil;
-import org.apache.phoenix.util.TransactionUtil;
 
 
 abstract public class BaseScannerRegionObserver implements RegionObserver {
@@ -187,8 +179,8 @@ abstract public class BaseScannerRegionObserver implements RegionObserver {
     }
 
     @Override
-    public RegionScanner preScannerOpen(final ObserverContext<RegionCoprocessorEnvironment> c,
-        final Scan scan, final RegionScanner s) throws IOException {
+    public void preScannerOpen(org.apache.hadoop.hbase.coprocessor.ObserverContext<RegionCoprocessorEnvironment> c,
+            Scan scan) throws IOException {
         byte[] txnScn = scan.getAttribute(TX_SCN);
         if (txnScn!=null) {
             TimeRange timeRange = scan.getTimeRange();
@@ -208,7 +200,6 @@ abstract public class BaseScannerRegionObserver implements RegionObserver {
         }
         this.encodingScheme = EncodedColumnsUtil.getQualifierEncodingScheme(scan);
         this.useNewValueColumnQualifier = EncodedColumnsUtil.useNewValueColumnQualifier(scan);
-        return s;
     }
 
     private class RegionScannerHolder extends DelegateRegionScanner {
@@ -352,17 +343,4 @@ abstract public class BaseScannerRegionObserver implements RegionObserver {
                 dataRegion, indexMaintainer, null, viewConstants, null, null, projector, ptr, useQualiferAsListIndex);
     }
 
-    @Override
-    public KeyValueScanner preStoreScannerOpen(final ObserverContext<RegionCoprocessorEnvironment> c,
-        final Store store, final Scan scan, final NavigableSet<byte[]> targetCols,
-        final KeyValueScanner s) throws IOException {
-
-      if (scan.isRaw() || ScanInfoUtil.isKeepDeletedCells(store.getScanInfo()) || scan.getTimeRange().getMax() == HConstants.LATEST_TIMESTAMP || TransactionUtil.isTransactionalTimestamp(scan.getTimeRange().getMax())) {
-        return s;
-      }
-
-      ScanInfo scanInfo = ScanInfoUtil.cloneScanInfoWithKeepDeletedCells(store.getScanInfo());
-      return new StoreScanner(store, scanInfo, scan, targetCols,
-          c.getEnvironment().getRegion().getReadpoint(scan.getIsolationLevel()));
-    }
 }

http://git-wip-us.apache.org/repos/asf/phoenix/blob/e1238aa3/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/DelegateRegionObserver.java
----------------------------------------------------------------------
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/DelegateRegionObserver.java b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/DelegateRegionObserver.java
index 34eee78..8fcd68d 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/DelegateRegionObserver.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/DelegateRegionObserver.java
@@ -74,25 +74,31 @@ public class DelegateRegionObserver implements RegionObserver {
 
 
     @Override
-    public void preFlush(ObserverContext<RegionCoprocessorEnvironment> c) throws IOException {
-        delegate.preFlush(c);
+    public void preFlush(org.apache.hadoop.hbase.coprocessor.ObserverContext<RegionCoprocessorEnvironment> c,
+            org.apache.hadoop.hbase.regionserver.FlushLifeCycleTracker tracker) throws IOException {
+        delegate.preFlush(c, tracker);
+        ;
     }
 
     @Override
-    public InternalScanner preFlush(ObserverContext<RegionCoprocessorEnvironment> c, Store store,
-            InternalScanner scanner) throws IOException {
-        return delegate.preFlush(c, store, scanner);
+    public InternalScanner preFlush(org.apache.hadoop.hbase.coprocessor.ObserverContext<RegionCoprocessorEnvironment> c,
+            Store store, InternalScanner scanner, org.apache.hadoop.hbase.regionserver.FlushLifeCycleTracker tracker)
+            throws IOException {
+        return delegate.preFlush(c, store, scanner, tracker);
     }
 
     @Override
-    public void postFlush(ObserverContext<RegionCoprocessorEnvironment> c) throws IOException {
-        delegate.postFlush(c);
+    public void postFlush(org.apache.hadoop.hbase.coprocessor.ObserverContext<RegionCoprocessorEnvironment> c,
+            org.apache.hadoop.hbase.regionserver.FlushLifeCycleTracker tracker) throws IOException {
+        delegate.postFlush(c, tracker);
     }
+    
 
     @Override
-    public void postFlush(ObserverContext<RegionCoprocessorEnvironment> c, Store store,
-            StoreFile resultFile) throws IOException {
-        delegate.postFlush(c, store, resultFile);
+    public void postFlush(org.apache.hadoop.hbase.coprocessor.ObserverContext<RegionCoprocessorEnvironment> c,
+            Store store, StoreFile resultFile, org.apache.hadoop.hbase.regionserver.FlushLifeCycleTracker tracker)
+            throws IOException {
+        delegate.postFlush(c, store, resultFile, tracker);
     }
 
     
@@ -232,9 +238,9 @@ public class DelegateRegionObserver implements RegionObserver {
     }
 
     @Override
-    public RegionScanner preScannerOpen(ObserverContext<RegionCoprocessorEnvironment> c, Scan scan,
-            RegionScanner s) throws IOException {
-        return delegate.preScannerOpen(c, scan, s);
+    public void preScannerOpen(org.apache.hadoop.hbase.coprocessor.ObserverContext<RegionCoprocessorEnvironment> c,
+            Scan scan) throws IOException {
+        delegate.preScannerOpen(c, scan);
     }
 
     @Override

http://git-wip-us.apache.org/repos/asf/phoenix/blob/e1238aa3/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/DelegateRegionScanner.java
----------------------------------------------------------------------
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/DelegateRegionScanner.java b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/DelegateRegionScanner.java
index 21a8eef..1955708 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/DelegateRegionScanner.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/DelegateRegionScanner.java
@@ -82,11 +82,6 @@ public class DelegateRegionScanner implements RegionScanner {
         return delegate.getBatch();
     }
 
-    @Override
-    public void shipped() throws IOException {
-        delegate.shipped();
-        
-    }
 
     @Override
     public RegionInfo getRegionInfo() {

http://git-wip-us.apache.org/repos/asf/phoenix/blob/e1238aa3/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/HashJoinRegionScanner.java
----------------------------------------------------------------------
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/HashJoinRegionScanner.java b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/HashJoinRegionScanner.java
index 59f844d..c24bb74 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/HashJoinRegionScanner.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/HashJoinRegionScanner.java
@@ -25,7 +25,7 @@ import java.util.Queue;
 
 import org.apache.hadoop.hbase.Cell;
 import org.apache.hadoop.hbase.DoNotRetryIOException;
-import org.apache.hadoop.hbase.HRegionInfo;
+import org.apache.hadoop.hbase.client.RegionInfo;
 import org.apache.hadoop.hbase.client.Result;
 import org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment;
 import org.apache.hadoop.hbase.regionserver.RegionScanner;
@@ -248,7 +248,7 @@ public class HashJoinRegionScanner implements RegionScanner {
     }
 
     @Override
-    public HRegionInfo getRegionInfo() {
+    public RegionInfo getRegionInfo() {
         return scanner.getRegionInfo();
     }
 

http://git-wip-us.apache.org/repos/asf/phoenix/blob/e1238aa3/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
----------------------------------------------------------------------
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
index a87e961..63719cf 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
@@ -1728,7 +1728,7 @@ public class MetaDataEndpointImpl extends MetaDataProtocol implements Coprocesso
         // Table hTable = env.getTable(systemCatalogTableName);
         // These deprecated calls work around the issue
         try (Table hTable = ServerUtil.getHTableForCoprocessorScan(env,
-            region.getTableDesc().getTableName())) {
+            region.getTableDescriptor().getTableName())) {
             boolean allViewsInCurrentRegion = true;
             int numOfChildViews = 0;
             List<ViewInfo> viewInfoList = Lists.newArrayList();
@@ -1777,7 +1777,7 @@ public class MetaDataEndpointImpl extends MetaDataProtocol implements Coprocesso
         // Table hTable = env.getTable(systemCatalogTableName);
         // These deprecated calls work around the issue
         try (Table hTable = ServerUtil.getHTableForCoprocessorScan(env,
-            region.getTableDesc().getTableName())) {
+            region.getTableDescriptor().getTableName())) {
             boolean allViewsInCurrentRegion = true;
             int numOfChildViews = 0;
             List<ViewInfo> viewInfoList = Lists.newArrayList();

http://git-wip-us.apache.org/repos/asf/phoenix/blob/e1238aa3/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/UngroupedAggregateRegionObserver.java
----------------------------------------------------------------------
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/UngroupedAggregateRegionObserver.java b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/UngroupedAggregateRegionObserver.java
index 82bfc07..332ecf2 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/UngroupedAggregateRegionObserver.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/UngroupedAggregateRegionObserver.java
@@ -47,20 +47,21 @@ import org.apache.hadoop.hbase.Cell;
 import org.apache.hadoop.hbase.CellUtil;
 import org.apache.hadoop.hbase.CoprocessorEnvironment;
 import org.apache.hadoop.hbase.DoNotRetryIOException;
-import org.apache.hadoop.hbase.HColumnDescriptor;
 import org.apache.hadoop.hbase.HConstants;
-import org.apache.hadoop.hbase.HRegionInfo;
 import org.apache.hadoop.hbase.HTableDescriptor;
 import org.apache.hadoop.hbase.KeyValue;
 import org.apache.hadoop.hbase.NamespaceDescriptor;
 import org.apache.hadoop.hbase.TableName;
+import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
 import org.apache.hadoop.hbase.client.ConnectionFactory;
 import org.apache.hadoop.hbase.client.Delete;
 import org.apache.hadoop.hbase.client.Durability;
 import org.apache.hadoop.hbase.client.Mutation;
 import org.apache.hadoop.hbase.client.Put;
+import org.apache.hadoop.hbase.client.RegionInfo;
 import org.apache.hadoop.hbase.client.Scan;
 import org.apache.hadoop.hbase.client.Table;
+import org.apache.hadoop.hbase.client.TableDescriptor;
 import org.apache.hadoop.hbase.coprocessor.ObserverContext;
 import org.apache.hadoop.hbase.coprocessor.RegionCoprocessor;
 import org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment;
@@ -304,18 +305,17 @@ public class UngroupedAggregateRegionObserver extends BaseScannerRegionObserver
     }
 
     @Override
-    public RegionScanner preScannerOpen(ObserverContext<RegionCoprocessorEnvironment> e, Scan scan, RegionScanner s)
-            throws IOException {
-        s = super.preScannerOpen(e, scan, s);
+    public void preScannerOpen(org.apache.hadoop.hbase.coprocessor.ObserverContext<RegionCoprocessorEnvironment> c,
+            Scan scan) throws IOException {
+    
         if (ScanUtil.isAnalyzeTable(scan)) {
             // We are setting the start row and stop row such that it covers the entire region. As part
             // of Phonenix-1263 we are storing the guideposts against the physical table rather than
             // individual tenant specific tables.
-            scan.setStartRow(HConstants.EMPTY_START_ROW);
-            scan.setStopRow(HConstants.EMPTY_END_ROW);
+            scan.withStartRow(HConstants.EMPTY_START_ROW);
+            scan.withStopRow(HConstants.EMPTY_END_ROW);
             scan.setFilter(null);
         }
-        return s;
     }
 
    public static class MutationList extends ArrayList<Mutation> {
@@ -831,7 +831,7 @@ public class UngroupedAggregateRegionObserver extends BaseScannerRegionObserver
 
     private void checkForLocalIndexColumnFamilies(Region region,
             List<IndexMaintainer> indexMaintainers) throws IOException {
-        HTableDescriptor tableDesc = region.getTableDesc();
+        TableDescriptor tableDesc = region.getTableDescriptor();
         String schemaName =
                 tableDesc.getTableName().getNamespaceAsString()
                         .equals(NamespaceDescriptor.DEFAULT_NAMESPACE_NAME_STR) ? SchemaUtil
@@ -843,7 +843,7 @@ public class UngroupedAggregateRegionObserver extends BaseScannerRegionObserver
             if(coveredColumns.isEmpty()) {
                 byte[] localIndexCf = indexMaintainer.getEmptyKeyValueFamily().get();
                 // When covered columns empty we store index data in default column family so check for it.
-                if (tableDesc.getFamily(localIndexCf) == null) {
+                if (tableDesc.getColumnFamily(localIndexCf) == null) {
                     ServerUtil.throwIOException("Column Family Not Found",
                         new ColumnFamilyNotFoundException(schemaName, tableName, Bytes
                                 .toString(localIndexCf)));
@@ -851,7 +851,7 @@ public class UngroupedAggregateRegionObserver extends BaseScannerRegionObserver
             }
             for (ColumnReference reference : coveredColumns) {
                 byte[] cf = IndexUtil.getLocalIndexColumnFamily(reference.getFamily());
-                HColumnDescriptor family = region.getTableDesc().getFamily(cf);
+                ColumnFamilyDescriptor family = region.getTableDescriptor().getColumnFamily(cf);
                 if (family == null) {
                     ServerUtil.throwIOException("Column Family Not Found",
                         new ColumnFamilyNotFoundException(schemaName, tableName, Bytes.toString(cf)));
@@ -882,7 +882,7 @@ public class UngroupedAggregateRegionObserver extends BaseScannerRegionObserver
         //if we're writing to the same table, but the PK can change, that means that some
         //mutations might be in our current region, and others in a different one.
         if (areMutationsInSameTable && isPKChanging) {
-            HRegionInfo regionInfo = region.getRegionInfo();
+            RegionInfo regionInfo = region.getRegionInfo();
             for (Mutation mutation : mutations){
                 if (regionInfo.containsRow(mutation.getRow())){
                     localRegionMutations.add(mutation);
@@ -898,8 +898,8 @@ public class UngroupedAggregateRegionObserver extends BaseScannerRegionObserver
     }
 
     private boolean areMutationsInSameTable(Table targetHTable, Region region) {
-        return (targetHTable == null || Bytes.compareTo(targetHTable.getName(),
-                region.getTableDesc().getTableName().getName()) == 0);
+        return (targetHTable == null || Bytes.compareTo(targetHTable.getName().getName(),
+                region.getTableDescriptor().getTableName().getName()) == 0);
     }
 
     @Override
@@ -1101,7 +1101,7 @@ public class UngroupedAggregateRegionObserver extends BaseScannerRegionObserver
 
         RegionScanner scanner = new BaseRegionScanner(innerScanner) {
             @Override
-            public HRegionInfo getRegionInfo() {
+            public RegionInfo getRegionInfo() {
                 return region.getRegionInfo();
             }
 
@@ -1159,7 +1159,7 @@ public class UngroupedAggregateRegionObserver extends BaseScannerRegionObserver
                     SINGLE_COLUMN, AGG_TIMESTAMP, rowCountBytes, 0, rowCountBytes.length);
         RegionScanner scanner = new BaseRegionScanner(innerScanner) {
             @Override
-            public HRegionInfo getRegionInfo() {
+            public RegionInfo getRegionInfo() {
                 return region.getRegionInfo();
             }
 
@@ -1325,6 +1325,10 @@ public class UngroupedAggregateRegionObserver extends BaseScannerRegionObserver
         }
     }
 
+    
+    /*
+     * TODO: use waitForFlushes PHOENIX-4352
+     */
     @Override
     public void preSplit(ObserverContext<RegionCoprocessorEnvironment> c, byte[] splitRow)
             throws IOException {

http://git-wip-us.apache.org/repos/asf/phoenix/blob/e1238aa3/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/DefaultStatisticsCollector.java
----------------------------------------------------------------------
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/DefaultStatisticsCollector.java b/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/DefaultStatisticsCollector.java
index c14759f..42f8b91 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/DefaultStatisticsCollector.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/DefaultStatisticsCollector.java
@@ -135,7 +135,7 @@ class DefaultStatisticsCollector implements StatisticsCollector {
                 guidepostWidth = PLong.INSTANCE.getCodec().decodeInt(guidePostWidthBytes, 0, SortOrder.getDefault());
             }
             this.guidePostDepth = StatisticsUtil.getGuidePostDepth(guidepostPerRegion, guidepostWidth,
-                env.getRegion().getTableDesc());
+                env.getRegion().getTableDescriptor());
         } else {
             long guidepostWidth = -1;
             Table htable = null;
@@ -203,7 +203,7 @@ class DefaultStatisticsCollector implements StatisticsCollector {
                     config.getLong(
                         QueryServices.STATS_GUIDEPOST_WIDTH_BYTES_ATTRIB,
                         QueryServicesOptions.DEFAULT_STATS_GUIDEPOST_WIDTH_BYTES),
-                    env.getRegion().getTableDesc());
+                    env.getRegion().getTableDescriptor());
             }
         }
     }
@@ -353,7 +353,7 @@ class DefaultStatisticsCollector implements StatisticsCollector {
         if (logger.isDebugEnabled()) {
             logger.debug("Compaction scanner created for stats");
         }
-        ImmutableBytesPtr cfKey = new ImmutableBytesPtr(store.getFamily().getName());
+        ImmutableBytesPtr cfKey = new ImmutableBytesPtr(store.getColumnFamilyDescriptor().getName());
         // Potentially perform a cross region server get in order to use the correct guide posts
         // width for the table being compacted.
         init();

http://git-wip-us.apache.org/repos/asf/phoenix/blob/e1238aa3/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/StatisticsCollectionRunTracker.java
----------------------------------------------------------------------
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/StatisticsCollectionRunTracker.java b/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/StatisticsCollectionRunTracker.java
index 560fc0a..1a928db 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/StatisticsCollectionRunTracker.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/StatisticsCollectionRunTracker.java
@@ -26,7 +26,7 @@ import java.util.concurrent.Executors;
 import java.util.concurrent.Future;
 
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.hbase.HRegionInfo;
+import org.apache.hadoop.hbase.client.RegionInfo;
 import org.apache.phoenix.query.QueryServices;
 import org.apache.phoenix.query.QueryServicesOptions;
 
@@ -38,10 +38,10 @@ import com.google.common.util.concurrent.ThreadFactoryBuilder;
  */
 public class StatisticsCollectionRunTracker {
     private static volatile StatisticsCollectionRunTracker INSTANCE;
-    private final Set<HRegionInfo> updateStatsRegions = Collections
-            .newSetFromMap(new ConcurrentHashMap<HRegionInfo, Boolean>());
-    private final Set<HRegionInfo> compactingRegions = Collections
-            .newSetFromMap(new ConcurrentHashMap<HRegionInfo, Boolean>());
+    private final Set<RegionInfo> updateStatsRegions = Collections
+            .newSetFromMap(new ConcurrentHashMap<RegionInfo, Boolean>());
+    private final Set<RegionInfo> compactingRegions = Collections
+            .newSetFromMap(new ConcurrentHashMap<RegionInfo, Boolean>());
     private final ExecutorService executor;
     
     // Constants added for testing purposes
@@ -77,7 +77,7 @@ public class StatisticsCollectionRunTracker {
      * @return true if the region wasn't already marked for stats collection via compaction, false
      *         otherwise.
      */
-    public boolean addCompactingRegion(HRegionInfo regionInfo) {
+    public boolean addCompactingRegion(RegionInfo regionInfo) {
         return compactingRegions.add(regionInfo);
     }
 
@@ -86,7 +86,7 @@ public class StatisticsCollectionRunTracker {
      *            major compaction.
      * @return true if the region was marked for stats collection via compaction, false otherwise.
      */
-    public boolean removeCompactingRegion(HRegionInfo regionInfo) {
+    public boolean removeCompactingRegion(RegionInfo regionInfo) {
         return compactingRegions.remove(regionInfo);
     }
 
@@ -95,7 +95,7 @@ public class StatisticsCollectionRunTracker {
      * @return true if stats are being collected for the region via major compaction, false
      *         otherwise.
      */
-    public boolean areStatsBeingCollectedOnCompaction(HRegionInfo regionInfo) {
+    public boolean areStatsBeingCollectedOnCompaction(RegionInfo regionInfo) {
         return compactingRegions.contains(regionInfo);
     }
 
@@ -103,7 +103,7 @@ public class StatisticsCollectionRunTracker {
      * @param regionInfo for the region to run UPDATE STATISTICS command on.
      * @return true if UPDATE STATISTICS wasn't already running on the region, false otherwise.
      */
-    public boolean addUpdateStatsCommandRegion(HRegionInfo regionInfo) {
+    public boolean addUpdateStatsCommandRegion(RegionInfo regionInfo) {
         return updateStatsRegions.add(regionInfo);
     }
 
@@ -111,7 +111,7 @@ public class StatisticsCollectionRunTracker {
      * @param regionInfo for the region to mark as not running UPDATE STATISTICS command on.
      * @return true if UPDATE STATISTICS was running on the region, false otherwise.
      */
-    public boolean removeUpdateStatsCommandRegion(HRegionInfo regionInfo) {
+    public boolean removeUpdateStatsCommandRegion(RegionInfo regionInfo) {
         return updateStatsRegions.remove(regionInfo);
     }
 

http://git-wip-us.apache.org/repos/asf/phoenix/blob/e1238aa3/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/StatisticsUtil.java
----------------------------------------------------------------------
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/StatisticsUtil.java b/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/StatisticsUtil.java
index 71b01ae..7f1b4c2 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/StatisticsUtil.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/StatisticsUtil.java
@@ -24,12 +24,12 @@ import java.util.Set;
 import org.apache.hadoop.hbase.Cell;
 import org.apache.hadoop.hbase.CellScanner;
 import org.apache.hadoop.hbase.HConstants;
-import org.apache.hadoop.hbase.HTableDescriptor;
 import org.apache.hadoop.hbase.TableName;
 import org.apache.hadoop.hbase.client.Result;
 import org.apache.hadoop.hbase.client.ResultScanner;
 import org.apache.hadoop.hbase.client.Scan;
 import org.apache.hadoop.hbase.client.Table;
+import org.apache.hadoop.hbase.client.TableDescriptor;
 import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
 import org.apache.hadoop.hbase.util.Bytes;
 import org.apache.phoenix.coprocessor.MetaDataProtocol;
@@ -190,7 +190,7 @@ public class StatisticsUtil {
         return current == null ? GuidePostsInfo.NO_GUIDEPOST : guidePostsInfoBuilder.isEmpty() ? emptyGuidePost : guidePostsInfoBuilder.build();
     }
 
-    public static long getGuidePostDepth(int guidepostPerRegion, long guidepostWidth, HTableDescriptor tableDesc) {
+    public static long getGuidePostDepth(int guidepostPerRegion, long guidepostWidth, TableDescriptor tableDesc) {
         if (guidepostPerRegion > 0) {
             long maxFileSize = HConstants.DEFAULT_MAX_FILE_SIZE;
             if (tableDesc != null) {