You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by mb...@apache.org on 2013/02/13 01:08:48 UTC
svn commit: r1445434 - in /hbase/trunk:
hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/
hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/
hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ hbase-ser...
Author: mbertozzi
Date: Wed Feb 13 00:08:47 2013
New Revision: 1445434
URL: http://svn.apache.org/r1445434
Log:
HBASE-7802 Use the Store interface instead of HStore (coprocessor, compaction, ...)
Added:
hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ScanInfo.java
Modified:
hbase/trunk/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.java
hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRegionObserver.java
hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionObserver.java
hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java
hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ScanQueryMatcher.java
hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java
hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreScanner.java
hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/DefaultCompactor.java
hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java
hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/SimpleRegionObserver.java
hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorInterface.java
hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverInterface.java
hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/NoOpScanPolicyObserver.java
hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompaction.java
hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMemStore.java
hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestQueryMatcher.java
hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStore.java
hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreScanner.java
hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestCoprocessorScanPolicy.java
Modified: hbase/trunk/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.java?rev=1445434&r1=1445433&r2=1445434&view=diff
==============================================================================
--- hbase/trunk/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.java (original)
+++ hbase/trunk/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.java Wed Feb 13 00:08:47 2013
@@ -32,8 +32,8 @@ import org.apache.hadoop.hbase.coprocess
import org.apache.hadoop.hbase.coprocessor.ObserverContext;
import org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment;
import org.apache.hadoop.hbase.coprocessor.RegionObserver;
-import org.apache.hadoop.hbase.regionserver.HStore;
-import org.apache.hadoop.hbase.regionserver.HStore.ScanInfo;
+import org.apache.hadoop.hbase.regionserver.Store;
+import org.apache.hadoop.hbase.regionserver.ScanInfo;
import org.apache.hadoop.hbase.regionserver.InternalScanner;
import org.apache.hadoop.hbase.regionserver.KeyValueScanner;
import org.apache.hadoop.hbase.regionserver.ScanType;
@@ -175,7 +175,7 @@ public class ZooKeeperScanPolicyObserver
// nothing to do here
}
- protected ScanInfo getScanInfo(HStore store, RegionCoprocessorEnvironment e) {
+ protected ScanInfo getScanInfo(Store store, RegionCoprocessorEnvironment e) {
byte[] data = ((ZKWatcher)e.getSharedData().get(zkkey)).getData();
if (data == null) {
return null;
@@ -184,15 +184,16 @@ public class ZooKeeperScanPolicyObserver
if (oldSI.getTtl() == Long.MAX_VALUE) {
return null;
}
- long ttl = Math.max(EnvironmentEdgeManager.currentTimeMillis() - Bytes.toLong(data), oldSI.getTtl());
+ long ttl = Math.max(EnvironmentEdgeManager.currentTimeMillis() -
+ Bytes.toLong(data), oldSI.getTtl());
return new ScanInfo(store.getFamily(), ttl,
oldSI.getTimeToPurgeDeletes(), oldSI.getComparator());
}
@Override
public InternalScanner preFlushScannerOpen(final ObserverContext<RegionCoprocessorEnvironment> c,
- HStore store, KeyValueScanner memstoreScanner, InternalScanner s) throws IOException {
- HStore.ScanInfo scanInfo = getScanInfo(store, c.getEnvironment());
+ Store store, KeyValueScanner memstoreScanner, InternalScanner s) throws IOException {
+ ScanInfo scanInfo = getScanInfo(store, c.getEnvironment());
if (scanInfo == null) {
// take default action
return null;
@@ -200,30 +201,30 @@ public class ZooKeeperScanPolicyObserver
Scan scan = new Scan();
scan.setMaxVersions(scanInfo.getMaxVersions());
return new StoreScanner(store, scanInfo, scan, Collections.singletonList(memstoreScanner),
- ScanType.MINOR_COMPACT, store.getHRegion().getSmallestReadPoint(),
- HConstants.OLDEST_TIMESTAMP);
+ ScanType.MINOR_COMPACT, store.getSmallestReadPoint(), HConstants.OLDEST_TIMESTAMP);
}
@Override
- public InternalScanner preCompactScannerOpen(final ObserverContext<RegionCoprocessorEnvironment> c,
- HStore store, List<? extends KeyValueScanner> scanners, ScanType scanType, long earliestPutTs,
+ public InternalScanner preCompactScannerOpen(
+ final ObserverContext<RegionCoprocessorEnvironment> c,
+ Store store, List<? extends KeyValueScanner> scanners, ScanType scanType, long earliestPutTs,
InternalScanner s) throws IOException {
- HStore.ScanInfo scanInfo = getScanInfo(store, c.getEnvironment());
+ ScanInfo scanInfo = getScanInfo(store, c.getEnvironment());
if (scanInfo == null) {
// take default action
return null;
}
Scan scan = new Scan();
scan.setMaxVersions(scanInfo.getMaxVersions());
- return new StoreScanner(store, scanInfo, scan, scanners, scanType, store.getHRegion()
- .getSmallestReadPoint(), earliestPutTs);
+ return new StoreScanner(store, scanInfo, scan, scanners, scanType,
+ store.getSmallestReadPoint(), earliestPutTs);
}
@Override
public KeyValueScanner preStoreScannerOpen(final ObserverContext<RegionCoprocessorEnvironment> c,
- final HStore store, final Scan scan, final NavigableSet<byte[]> targetCols,
+ final Store store, final Scan scan, final NavigableSet<byte[]> targetCols,
final KeyValueScanner s) throws IOException {
- HStore.ScanInfo scanInfo = getScanInfo(store, c.getEnvironment());
+ ScanInfo scanInfo = getScanInfo(store, c.getEnvironment());
if (scanInfo == null) {
// take default action
return null;
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRegionObserver.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRegionObserver.java?rev=1445434&r1=1445433&r2=1445434&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRegionObserver.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRegionObserver.java Wed Feb 13 00:08:47 2013
@@ -39,7 +39,7 @@ import org.apache.hadoop.hbase.regionser
import org.apache.hadoop.hbase.regionserver.KeyValueScanner;
import org.apache.hadoop.hbase.regionserver.RegionScanner;
import org.apache.hadoop.hbase.regionserver.ScanType;
-import org.apache.hadoop.hbase.regionserver.HStore;
+import org.apache.hadoop.hbase.regionserver.Store;
import org.apache.hadoop.hbase.regionserver.StoreFile;
import org.apache.hadoop.hbase.regionserver.wal.HLogKey;
import org.apache.hadoop.hbase.regionserver.wal.WALEdit;
@@ -77,7 +77,7 @@ public abstract class BaseRegionObserver
@Override
public InternalScanner preFlushScannerOpen(final ObserverContext<RegionCoprocessorEnvironment> c,
- final HStore store, final KeyValueScanner memstoreScanner, final InternalScanner s)
+ final Store store, final KeyValueScanner memstoreScanner, final InternalScanner s)
throws IOException {
return null;
}
@@ -91,13 +91,13 @@ public abstract class BaseRegionObserver
}
@Override
- public InternalScanner preFlush(ObserverContext<RegionCoprocessorEnvironment> e, HStore store,
+ public InternalScanner preFlush(ObserverContext<RegionCoprocessorEnvironment> e, Store store,
InternalScanner scanner) throws IOException {
return scanner;
}
@Override
- public void postFlush(ObserverContext<RegionCoprocessorEnvironment> e, HStore store,
+ public void postFlush(ObserverContext<RegionCoprocessorEnvironment> e, Store store,
StoreFile resultFile) throws IOException {
}
@@ -132,28 +132,28 @@ public abstract class BaseRegionObserver
@Override
public void preCompactSelection(final ObserverContext<RegionCoprocessorEnvironment> c,
- final HStore store, final List<StoreFile> candidates) throws IOException { }
+ final Store store, final List<StoreFile> candidates) throws IOException { }
@Override
public void postCompactSelection(final ObserverContext<RegionCoprocessorEnvironment> c,
- final HStore store, final ImmutableList<StoreFile> selected) { }
+ final Store store, final ImmutableList<StoreFile> selected) { }
@Override
public InternalScanner preCompact(ObserverContext<RegionCoprocessorEnvironment> e,
- final HStore store, final InternalScanner scanner, final ScanType scanType)
+ final Store store, final InternalScanner scanner, final ScanType scanType)
throws IOException {
return scanner;
}
@Override
public InternalScanner preCompactScannerOpen(final ObserverContext<RegionCoprocessorEnvironment> c,
- final HStore store, List<? extends KeyValueScanner> scanners, final ScanType scanType,
+ final Store store, List<? extends KeyValueScanner> scanners, final ScanType scanType,
final long earliestPutTs, final InternalScanner s) throws IOException {
return null;
}
@Override
- public void postCompact(ObserverContext<RegionCoprocessorEnvironment> e, final HStore store,
+ public void postCompact(ObserverContext<RegionCoprocessorEnvironment> e, final Store store,
final StoreFile resultFile) throws IOException {
}
@@ -290,7 +290,7 @@ public abstract class BaseRegionObserver
@Override
public KeyValueScanner preStoreScannerOpen(final ObserverContext<RegionCoprocessorEnvironment> c,
- final HStore store, final Scan scan, final NavigableSet<byte[]> targetCols,
+ final Store store, final Scan scan, final NavigableSet<byte[]> targetCols,
final KeyValueScanner s) throws IOException {
return null;
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionObserver.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionObserver.java?rev=1445434&r1=1445433&r2=1445434&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionObserver.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionObserver.java Wed Feb 13 00:08:47 2013
@@ -38,7 +38,7 @@ import org.apache.hadoop.hbase.regionser
import org.apache.hadoop.hbase.regionserver.KeyValueScanner;
import org.apache.hadoop.hbase.regionserver.RegionScanner;
import org.apache.hadoop.hbase.regionserver.ScanType;
-import org.apache.hadoop.hbase.regionserver.HStore;
+import org.apache.hadoop.hbase.regionserver.Store;
import org.apache.hadoop.hbase.regionserver.StoreFile;
import org.apache.hadoop.hbase.regionserver.StoreFileScanner;
import org.apache.hadoop.hbase.regionserver.wal.HLogKey;
@@ -84,14 +84,14 @@ public interface RegionObserver extends
* @throws IOException if an error occurred on the coprocessor
*/
InternalScanner preFlushScannerOpen(final ObserverContext<RegionCoprocessorEnvironment> c,
- final HStore store, final KeyValueScanner memstoreScanner, final InternalScanner s)
+ final Store store, final KeyValueScanner memstoreScanner, final InternalScanner s)
throws IOException;
/**
* Called before the memstore is flushed to disk.
* @param c the environment provided by the region server
* @throws IOException if an error occurred on the coprocessor
- * @deprecated use {@link #preFlush(ObserverContext, HStore, InternalScanner)} instead
+ * @deprecated use {@link #preFlush(ObserverContext, Store, InternalScanner)} instead
*/
void preFlush(final ObserverContext<RegionCoprocessorEnvironment> c) throws IOException;
@@ -104,14 +104,14 @@ public interface RegionObserver extends
* unless the implementation is writing new store files on its own.
* @throws IOException if an error occurred on the coprocessor
*/
- InternalScanner preFlush(final ObserverContext<RegionCoprocessorEnvironment> c, final HStore store,
+ InternalScanner preFlush(final ObserverContext<RegionCoprocessorEnvironment> c, final Store store,
final InternalScanner scanner) throws IOException;
/**
* Called after the memstore is flushed to disk.
* @param c the environment provided by the region server
* @throws IOException if an error occurred on the coprocessor
- * @deprecated use {@link #preFlush(ObserverContext, HStore, InternalScanner)} instead.
+ * @deprecated use {@link #preFlush(ObserverContext, Store, InternalScanner)} instead.
*/
void postFlush(final ObserverContext<RegionCoprocessorEnvironment> c) throws IOException;
@@ -122,7 +122,7 @@ public interface RegionObserver extends
* @param resultFile the new store file written out during compaction
* @throws IOException if an error occurred on the coprocessor
*/
- void postFlush(final ObserverContext<RegionCoprocessorEnvironment> c, final HStore store,
+ void postFlush(final ObserverContext<RegionCoprocessorEnvironment> c, final Store store,
final StoreFile resultFile) throws IOException;
/**
@@ -135,7 +135,7 @@ public interface RegionObserver extends
* @throws IOException if an error occurred on the coprocessor
*/
void preCompactSelection(final ObserverContext<RegionCoprocessorEnvironment> c,
- final HStore store, final List<StoreFile> candidates) throws IOException;
+ final Store store, final List<StoreFile> candidates) throws IOException;
/**
* Called after the {@link StoreFile}s to compact have been selected from the
@@ -145,7 +145,7 @@ public interface RegionObserver extends
* @param selected the store files selected to compact
*/
void postCompactSelection(final ObserverContext<RegionCoprocessorEnvironment> c,
- final HStore store, final ImmutableList<StoreFile> selected);
+ final Store store, final ImmutableList<StoreFile> selected);
/**
* Called prior to writing the {@link StoreFile}s selected for compaction into
@@ -173,7 +173,7 @@ public interface RegionObserver extends
* @throws IOException if an error occurred on the coprocessor
*/
InternalScanner preCompact(final ObserverContext<RegionCoprocessorEnvironment> c,
- final HStore store, final InternalScanner scanner,
+ final Store store, final InternalScanner scanner,
final ScanType scanType) throws IOException;
/**
@@ -196,7 +196,7 @@ public interface RegionObserver extends
* @throws IOException if an error occurred on the coprocessor
*/
InternalScanner preCompactScannerOpen(final ObserverContext<RegionCoprocessorEnvironment> c,
- final HStore store, List<? extends KeyValueScanner> scanners, final ScanType scanType,
+ final Store store, List<? extends KeyValueScanner> scanners, final ScanType scanType,
final long earliestPutTs, final InternalScanner s) throws IOException;
/**
@@ -207,7 +207,7 @@ public interface RegionObserver extends
* @param resultFile the new store file written out during compaction
* @throws IOException if an error occurred on the coprocessor
*/
- void postCompact(final ObserverContext<RegionCoprocessorEnvironment> c, final HStore store,
+ void postCompact(final ObserverContext<RegionCoprocessorEnvironment> c, final Store store,
StoreFile resultFile) throws IOException;
/**
@@ -215,17 +215,19 @@ public interface RegionObserver extends
* @param c the environment provided by the region server
* (e.getRegion() returns the parent region)
* @throws IOException if an error occurred on the coprocessor
- * @deprecated Use preSplit(final ObserverContext<RegionCoprocessorEnvironment> c, byte[] splitRow)
+ * @deprecated Use preSplit(
+ * final ObserverContext<RegionCoprocessorEnvironment> c, byte[] splitRow)
*/
void preSplit(final ObserverContext<RegionCoprocessorEnvironment> c) throws IOException;
-
+
/**
* Called before the region is split.
* @param c the environment provided by the region server
* (e.getRegion() returns the parent region)
* @throws IOException if an error occurred on the coprocessor
*/
- void preSplit(final ObserverContext<RegionCoprocessorEnvironment> c, byte[] splitRow) throws IOException;
+ void preSplit(final ObserverContext<RegionCoprocessorEnvironment> c, byte[] splitRow)
+ throws IOException;
/**
* Called after the region is split.
@@ -238,28 +240,30 @@ public interface RegionObserver extends
*/
void postSplit(final ObserverContext<RegionCoprocessorEnvironment> c, final HRegion l,
final HRegion r) throws IOException;
-
+
/**
* This will be called before the roll back of the split region is completed
* @param ctx
* @throws IOException
*/
void preRollBackSplit(final ObserverContext<RegionCoprocessorEnvironment> ctx) throws IOException;
-
+
/**
* This will be called after the roll back of the split region is completed
* @param ctx
* @throws IOException
*/
- void postRollBackSplit(final ObserverContext<RegionCoprocessorEnvironment> ctx) throws IOException;
-
+ void postRollBackSplit(final ObserverContext<RegionCoprocessorEnvironment> ctx)
+ throws IOException;
+
/**
* Called after any split request is processed. This will be called irrespective of success or
* failure of the split.
* @param ctx
* @throws IOException
*/
- void postCompleteSplit(final ObserverContext<RegionCoprocessorEnvironment> ctx) throws IOException;
+ void postCompleteSplit(final ObserverContext<RegionCoprocessorEnvironment> ctx)
+ throws IOException;
/**
* Called before the region is reported as closed to the master.
* @param c the environment provided by the region server
@@ -661,8 +665,9 @@ public interface RegionObserver extends
* Called before a store opens a new scanner.
* This hook is called when a "user" scanner is opened.
* <p>
- * See {@link #preFlushScannerOpen(ObserverContext, HStore, KeyValueScanner, InternalScanner)}
- * and {@link #preCompactScannerOpen(ObserverContext, HStore, List, ScanType, long, InternalScanner)}
+ * See {@link #preFlushScannerOpen(ObserverContext, Store, KeyValueScanner, InternalScanner)}
+ * and {@link #preCompactScannerOpen(ObserverContext,
+ * Store, List, ScanType, long, InternalScanner)}
* to override scanners created for flushes or compactions, resp.
* <p>
* Call CoprocessorEnvironment#complete to skip any subsequent chained
@@ -678,7 +683,7 @@ public interface RegionObserver extends
* @throws IOException if an error occurred on the coprocessor
*/
KeyValueScanner preStoreScannerOpen(final ObserverContext<RegionCoprocessorEnvironment> c,
- final HStore store, final Scan scan, final NavigableSet<byte[]> targetCols,
+ final Store store, final Scan scan, final NavigableSet<byte[]> targetCols,
final KeyValueScanner s) throws IOException;
/**
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java?rev=1445434&r1=1445433&r2=1445434&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java Wed Feb 13 00:08:47 2013
@@ -150,7 +150,7 @@ public class HStore implements Store {
private int bytesPerChecksum;
// Comparing KeyValues
- final KeyValue.KVComparator comparator;
+ private final KeyValue.KVComparator comparator;
private Compactor compactor;
@@ -287,6 +287,7 @@ public class HStore implements Store {
return homedir;
}
+ @Override
public FileSystem getFileSystem() {
return this.fs;
}
@@ -803,8 +804,8 @@ public class HStore implements Store {
// treat this as a minor compaction.
InternalScanner scanner = null;
KeyValueScanner memstoreScanner = new CollectionBackedScanner(set, this.comparator);
- if (getHRegion().getCoprocessorHost() != null) {
- scanner = getHRegion().getCoprocessorHost()
+ if (this.region.getCoprocessorHost() != null) {
+ scanner = this.region.getCoprocessorHost()
.preFlushScannerOpen(this, memstoreScanner);
}
if (scanner == null) {
@@ -814,9 +815,9 @@ public class HStore implements Store {
Collections.singletonList(memstoreScanner), ScanType.MINOR_COMPACT,
this.region.getSmallestReadPoint(), HConstants.OLDEST_TIMESTAMP);
}
- if (getHRegion().getCoprocessorHost() != null) {
+ if (this.region.getCoprocessorHost() != null) {
InternalScanner cpScanner =
- getHRegion().getCoprocessorHost().preFlush(this, scanner);
+ this.region.getCoprocessorHost().preFlush(this, scanner);
// NULL scanner returned from coprocessor hooks means skip normal processing
if (cpScanner == null) {
return null;
@@ -1001,7 +1002,8 @@ public class HStore implements Store {
* the line).
* @return all scanners for this store
*/
- protected List<KeyValueScanner> getScanners(boolean cacheBlocks,
+ @Override
+ public List<KeyValueScanner> getScanners(boolean cacheBlocks,
boolean isGet, boolean isCompaction, ScanQueryMatcher matcher, byte[] startRow,
byte[] stopRow) throws IOException {
Collection<StoreFile> storeFilesToScan;
@@ -1029,17 +1031,13 @@ public class HStore implements Store {
return scanners;
}
- /*
- * @param o Observer who wants to know about changes in set of Readers
- */
- void addChangedReaderObserver(ChangedReadersObserver o) {
+ @Override
+ public void addChangedReaderObserver(ChangedReadersObserver o) {
this.changedReaderObservers.add(o);
}
- /*
- * @param o Observer no longer interested in changes in set of Readers.
- */
- void deleteChangedReaderObserver(ChangedReadersObserver o) {
+ @Override
+ public void deleteChangedReaderObserver(ChangedReadersObserver o) {
// We don't check if observer present; it may not be (legitimately)
this.changedReaderObservers.remove(o);
}
@@ -1274,7 +1272,7 @@ public class HStore implements Store {
this.forceMajor = false;
}
- LOG.debug(getHRegionInfo().getEncodedName() + " - " +
+ LOG.debug(getRegionInfo().getEncodedName() + " - " +
getColumnFamilyName() + ": Initiating " +
(isMajor ? "major" : "minor") + " compaction");
@@ -1645,8 +1643,8 @@ public class HStore implements Store {
lock.readLock().lock();
try {
KeyValueScanner scanner = null;
- if (getHRegion().getCoprocessorHost() != null) {
- scanner = getHRegion().getCoprocessorHost().preStoreScannerOpen(this, scan, targetCols);
+ if (this.region.getCoprocessorHost() != null) {
+ scanner = this.region.getCoprocessorHost().preStoreScannerOpen(this, scan, targetCols);
}
if (scanner == null) {
scanner = new StoreScanner(this, getScanInfo(), scan, targetCols);
@@ -1743,15 +1741,30 @@ public class HStore implements Store {
return compactionPolicy.throttleCompaction(compactionSize);
}
- @Override
public HRegion getHRegion() {
return this.region;
}
- HRegionInfo getHRegionInfo() {
+ @Override
+ public RegionCoprocessorHost getCoprocessorHost() {
+ return this.region.getCoprocessorHost();
+ }
+
+ @Override
+ public HRegionInfo getRegionInfo() {
return this.region.getRegionInfo();
}
+ @Override
+ public boolean areWritesEnabled() {
+ return this.region.areWritesEnabled();
+ }
+
+ @Override
+ public long getSmallestReadPoint() {
+ return this.region.getSmallestReadPoint();
+ }
+
/**
* Used in tests. TODO: Remove
*
@@ -1832,7 +1845,7 @@ public class HStore implements Store {
}
storeFile = HStore.this.commitFile(storeFilePath, cacheFlushId,
snapshotTimeRangeTracker, flushedSize, status);
- if (HStore.this.getHRegion().getCoprocessorHost() != null) {
+ if (HStore.this.region.getCoprocessorHost() != null) {
HStore.this.getHRegion()
.getCoprocessorHost()
.postFlush(HStore.this, storeFile);
@@ -1873,6 +1886,7 @@ public class HStore implements Store {
return comparator;
}
+ @Override
public ScanInfo getScanInfo() {
return scanInfo;
}
@@ -1884,84 +1898,4 @@ public class HStore implements Store {
void setScanInfo(ScanInfo scanInfo) {
this.scanInfo = scanInfo;
}
-
- /**
- * Immutable information for scans over a store.
- */
- public static class ScanInfo {
- private byte[] family;
- private int minVersions;
- private int maxVersions;
- private long ttl;
- private boolean keepDeletedCells;
- private long timeToPurgeDeletes;
- private KVComparator comparator;
-
- public static final long FIXED_OVERHEAD = ClassSize.align(ClassSize.OBJECT
- + (2 * ClassSize.REFERENCE) + (2 * Bytes.SIZEOF_INT)
- + Bytes.SIZEOF_LONG + Bytes.SIZEOF_BOOLEAN);
-
- /**
- * @param family {@link HColumnDescriptor} describing the column family
- * @param ttl Store's TTL (in ms)
- * @param timeToPurgeDeletes duration in ms after which a delete marker can
- * be purged during a major compaction.
- * @param comparator The store's comparator
- */
- public ScanInfo(HColumnDescriptor family, long ttl, long timeToPurgeDeletes, KVComparator comparator) {
- this(family.getName(), family.getMinVersions(), family.getMaxVersions(), ttl, family
- .getKeepDeletedCells(), timeToPurgeDeletes, comparator);
- }
- /**
- * @param family Name of this store's column family
- * @param minVersions Store's MIN_VERSIONS setting
- * @param maxVersions Store's VERSIONS setting
- * @param ttl Store's TTL (in ms)
- * @param timeToPurgeDeletes duration in ms after which a delete marker can
- * be purged during a major compaction.
- * @param keepDeletedCells Store's keepDeletedCells setting
- * @param comparator The store's comparator
- */
- public ScanInfo(byte[] family, int minVersions, int maxVersions, long ttl,
- boolean keepDeletedCells, long timeToPurgeDeletes,
- KVComparator comparator) {
-
- this.family = family;
- this.minVersions = minVersions;
- this.maxVersions = maxVersions;
- this.ttl = ttl;
- this.keepDeletedCells = keepDeletedCells;
- this.timeToPurgeDeletes = timeToPurgeDeletes;
- this.comparator = comparator;
- }
-
- public byte[] getFamily() {
- return family;
- }
-
- public int getMinVersions() {
- return minVersions;
- }
-
- public int getMaxVersions() {
- return maxVersions;
- }
-
- public long getTtl() {
- return ttl;
- }
-
- public boolean getKeepDeletedCells() {
- return keepDeletedCells;
- }
-
- public long getTimeToPurgeDeletes() {
- return timeToPurgeDeletes;
- }
-
- public KVComparator getComparator() {
- return comparator;
- }
- }
-
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java?rev=1445434&r1=1445433&r2=1445434&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java Wed Feb 13 00:08:47 2013
@@ -246,7 +246,8 @@ public class RegionCoprocessorHost
* {@link org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost#postOpen()} are such hooks.
*
* See also
- * {@link org.apache.hadoop.hbase.master.MasterCoprocessorHost#handleCoprocessorThrowable(CoprocessorEnvironment, Throwable)}
+ * {@link org.apache.hadoop.hbase.master.MasterCoprocessorHost#handleCoprocessorThrowable(
+ * CoprocessorEnvironment, Throwable)}
* @param env The coprocessor that threw the exception.
* @param e The exception that was thrown.
*/
@@ -256,8 +257,9 @@ public class RegionCoprocessorHost
handleCoprocessorThrowable(env,e);
} catch (IOException ioe) {
// We cannot throw exceptions from the caller hook, so ignore.
- LOG.warn("handleCoprocessorThrowable() threw an IOException while attempting to handle Throwable " + e
- + ". Ignoring.",e);
+ LOG.warn(
+ "handleCoprocessorThrowable() threw an IOException while attempting to handle Throwable " +
+ e + ". Ignoring.",e);
}
}
@@ -343,9 +345,10 @@ public class RegionCoprocessorHost
/**
* See
- * {@link RegionObserver#preCompactScannerOpen(ObserverContext, HStore, List, ScanType, long, InternalScanner)}
+ * {@link RegionObserver#preCompactScannerOpen(ObserverContext,
+ * Store, List, ScanType, long, InternalScanner)}
*/
- public InternalScanner preCompactScannerOpen(HStore store, List<StoreFileScanner> scanners,
+ public InternalScanner preCompactScannerOpen(Store store, List<StoreFileScanner> scanners,
ScanType scanType, long earliestPutTs) throws IOException {
ObserverContext<RegionCoprocessorEnvironment> ctx = null;
InternalScanner s = null;
@@ -374,7 +377,7 @@ public class RegionCoprocessorHost
* @return If {@code true}, skip the normal selection process and use the current list
* @throws IOException
*/
- public boolean preCompactSelection(HStore store, List<StoreFile> candidates) throws IOException {
+ public boolean preCompactSelection(Store store, List<StoreFile> candidates) throws IOException {
ObserverContext<RegionCoprocessorEnvironment> ctx = null;
boolean bypass = false;
for (RegionEnvironment env: coprocessors) {
@@ -402,7 +405,7 @@ public class RegionCoprocessorHost
* @param store The store where compaction is being requested
* @param selected The store files selected to compact
*/
- public void postCompactSelection(HStore store,
+ public void postCompactSelection(Store store,
ImmutableList<StoreFile> selected) {
ObserverContext<RegionCoprocessorEnvironment> ctx = null;
for (RegionEnvironment env: coprocessors) {
@@ -428,7 +431,7 @@ public class RegionCoprocessorHost
* @param scanType type of Scan
* @throws IOException
*/
- public InternalScanner preCompact(HStore store, InternalScanner scanner,
+ public InternalScanner preCompact(Store store, InternalScanner scanner,
ScanType scanType) throws IOException {
ObserverContext<RegionCoprocessorEnvironment> ctx = null;
boolean bypass = false;
@@ -456,7 +459,7 @@ public class RegionCoprocessorHost
* @param resultFile the new store file written during compaction
* @throws IOException
*/
- public void postCompact(HStore store, StoreFile resultFile) throws IOException {
+ public void postCompact(Store store, StoreFile resultFile) throws IOException {
ObserverContext<RegionCoprocessorEnvironment> ctx = null;
for (RegionEnvironment env: coprocessors) {
if (env.getInstance() instanceof RegionObserver) {
@@ -477,7 +480,7 @@ public class RegionCoprocessorHost
* Invoked before a memstore flush
* @throws IOException
*/
- public InternalScanner preFlush(HStore store, InternalScanner scanner) throws IOException {
+ public InternalScanner preFlush(Store store, InternalScanner scanner) throws IOException {
ObserverContext<RegionCoprocessorEnvironment> ctx = null;
boolean bypass = false;
for (RegionEnvironment env: coprocessors) {
@@ -521,16 +524,19 @@ public class RegionCoprocessorHost
/**
* See
- * {@link RegionObserver#preFlushScannerOpen(ObserverContext, HStore, KeyValueScanner, InternalScanner)}
+ * {@link RegionObserver#preFlushScannerOpen(ObserverContext,
+ * Store, KeyValueScanner, InternalScanner)}
*/
- public InternalScanner preFlushScannerOpen(HStore store, KeyValueScanner memstoreScanner) throws IOException {
+ public InternalScanner preFlushScannerOpen(Store store, KeyValueScanner memstoreScanner)
+ throws IOException {
ObserverContext<RegionCoprocessorEnvironment> ctx = null;
InternalScanner s = null;
for (RegionEnvironment env : coprocessors) {
if (env.getInstance() instanceof RegionObserver) {
ctx = ObserverContext.createAndPrepare(env, ctx);
try {
- s = ((RegionObserver) env.getInstance()).preFlushScannerOpen(ctx, store, memstoreScanner, s);
+ s = ((RegionObserver) env.getInstance())
+ .preFlushScannerOpen(ctx, store, memstoreScanner, s);
} catch (Throwable e) {
handleCoprocessorThrowable(env, e);
}
@@ -567,7 +573,7 @@ public class RegionCoprocessorHost
* Invoked after a memstore flush
* @throws IOException
*/
- public void postFlush(final HStore store, final StoreFile storeFile) throws IOException {
+ public void postFlush(final Store store, final StoreFile storeFile) throws IOException {
ObserverContext<RegionCoprocessorEnvironment> ctx = null;
for (RegionEnvironment env: coprocessors) {
if (env.getInstance() instanceof RegionObserver) {
@@ -1243,9 +1249,10 @@ public class RegionCoprocessorHost
/**
* See
- * {@link RegionObserver#preStoreScannerOpen(ObserverContext, HStore, Scan, NavigableSet, KeyValueScanner)}
+ * {@link RegionObserver#preStoreScannerOpen(ObserverContext,
+ * Store, Scan, NavigableSet, KeyValueScanner)}
*/
- public KeyValueScanner preStoreScannerOpen(HStore store, Scan scan,
+ public KeyValueScanner preStoreScannerOpen(Store store, Scan scan,
final NavigableSet<byte[]> targetCols) throws IOException {
KeyValueScanner s = null;
ObserverContext<RegionCoprocessorEnvironment> ctx = null;
Added: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ScanInfo.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ScanInfo.java?rev=1445434&view=auto
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ScanInfo.java (added)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ScanInfo.java Wed Feb 13 00:08:47 2013
@@ -0,0 +1,106 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements. See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership. The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.hadoop.hbase.regionserver;
+
+import org.apache.hadoop.classification.InterfaceAudience;
+import org.apache.hadoop.hbase.HColumnDescriptor;
+import org.apache.hadoop.hbase.KeyValue.KVComparator;
+import org.apache.hadoop.hbase.util.Bytes;
+import org.apache.hadoop.hbase.util.ClassSize;
+
+/**
+ * Immutable information for scans over a store.
+ */
+@InterfaceAudience.Private
+public class ScanInfo {
+ private byte[] family;
+ private int minVersions;
+ private int maxVersions;
+ private long ttl;
+ private boolean keepDeletedCells;
+ private long timeToPurgeDeletes;
+ private KVComparator comparator;
+
+ public static final long FIXED_OVERHEAD = ClassSize.align(ClassSize.OBJECT
+ + (2 * ClassSize.REFERENCE) + (2 * Bytes.SIZEOF_INT)
+ + Bytes.SIZEOF_LONG + Bytes.SIZEOF_BOOLEAN);
+
+ /**
+ * @param family {@link HColumnDescriptor} describing the column family
+ * @param ttl Store's TTL (in ms)
+ * @param timeToPurgeDeletes duration in ms after which a delete marker can
+ * be purged during a major compaction.
+ * @param comparator The store's comparator
+ */
+ public ScanInfo(final HColumnDescriptor family, final long ttl, final long timeToPurgeDeletes,
+ final KVComparator comparator) {
+ this(family.getName(), family.getMinVersions(), family.getMaxVersions(), ttl, family
+ .getKeepDeletedCells(), timeToPurgeDeletes, comparator);
+ }
+
+ /**
+ * @param family Name of this store's column family
+ * @param minVersions Store's MIN_VERSIONS setting
+ * @param maxVersions Store's VERSIONS setting
+ * @param ttl Store's TTL (in ms)
+ * @param timeToPurgeDeletes duration in ms after which a delete marker can
+ * be purged during a major compaction.
+ * @param keepDeletedCells Store's keepDeletedCells setting
+ * @param comparator The store's comparator
+ */
+ public ScanInfo(final byte[] family, final int minVersions, final int maxVersions,
+ final long ttl, final boolean keepDeletedCells, final long timeToPurgeDeletes,
+ final KVComparator comparator) {
+ this.family = family;
+ this.minVersions = minVersions;
+ this.maxVersions = maxVersions;
+ this.ttl = ttl;
+ this.keepDeletedCells = keepDeletedCells;
+ this.timeToPurgeDeletes = timeToPurgeDeletes;
+ this.comparator = comparator;
+ }
+
+ public byte[] getFamily() {
+ return family;
+ }
+
+ public int getMinVersions() {
+ return minVersions;
+ }
+
+ public int getMaxVersions() {
+ return maxVersions;
+ }
+
+ public long getTtl() {
+ return ttl;
+ }
+
+ public boolean getKeepDeletedCells() {
+ return keepDeletedCells;
+ }
+
+ public long getTimeToPurgeDeletes() {
+ return timeToPurgeDeletes;
+ }
+
+ public KVComparator getComparator() {
+ return comparator;
+ }
+}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ScanQueryMatcher.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ScanQueryMatcher.java?rev=1445434&r1=1445433&r2=1445434&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ScanQueryMatcher.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ScanQueryMatcher.java Wed Feb 13 00:08:47 2013
@@ -136,7 +136,7 @@ public class ScanQueryMatcher {
* @param oldestUnexpiredTS the oldest timestamp we are interested in,
* based on TTL
*/
- public ScanQueryMatcher(Scan scan, HStore.ScanInfo scanInfo,
+ public ScanQueryMatcher(Scan scan, ScanInfo scanInfo,
NavigableSet<byte[]> columns, ScanType scanType,
long readPointToUse, long earliestPutTs, long oldestUnexpiredTS) {
this.tr = scan.getTimeRange();
@@ -182,7 +182,7 @@ public class ScanQueryMatcher {
/*
* Constructor for tests
*/
- ScanQueryMatcher(Scan scan, HStore.ScanInfo scanInfo,
+ ScanQueryMatcher(Scan scan, ScanInfo scanInfo,
NavigableSet<byte[]> columns, long oldestUnexpiredTS) {
this(scan, scanInfo, columns, ScanType.USER_SCAN,
Long.MAX_VALUE, /* max Readpoint to track versions */
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java?rev=1445434&r1=1445433&r2=1445434&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java Wed Feb 13 00:08:47 2013
@@ -24,11 +24,14 @@ import java.util.NavigableSet;
import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.classification.InterfaceStability;
+import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hbase.HColumnDescriptor;
+import org.apache.hadoop.hbase.HRegionInfo;
import org.apache.hadoop.hbase.KeyValue;
import org.apache.hadoop.hbase.client.Scan;
import org.apache.hadoop.hbase.io.HeapSize;
+import org.apache.hadoop.hbase.io.compress.Compression;
import org.apache.hadoop.hbase.io.hfile.CacheConfig;
import org.apache.hadoop.hbase.io.hfile.HFileDataBlockEncoder;
import org.apache.hadoop.hbase.regionserver.compactions.CompactionProgress;
@@ -75,6 +78,23 @@ public interface Store extends HeapSize,
throws IOException;
/**
+ * Get all scanners with no filtering based on TTL (that happens further down
+ * the line).
+ * @param cacheBlocks
+ * @param isGet
+ * @param isCompaction
+ * @param matcher
+ * @param startRow
+ * @param stopRow
+ * @return all scanners for this store
+ */
+ public List<KeyValueScanner> getScanners(boolean cacheBlocks,
+ boolean isGet, boolean isCompaction, ScanQueryMatcher matcher, byte[] startRow,
+ byte[] stopRow) throws IOException;
+
+ public ScanInfo getScanInfo();
+
+ /**
* Adds or replaces the specified KeyValues.
* <p>
* For each KeyValue specified, if a cell with the same row, family, and qualifier exists in
@@ -117,6 +137,17 @@ public interface Store extends HeapSize,
*/
public KeyValue getRowKeyAtOrBefore(final byte[] row) throws IOException;
+ public FileSystem getFileSystem();
+
+ /*
+ * @param maxKeyCount
+ * @param compression Compression algorithm to use
+ * @param isCompaction whether we are creating a new file in a compaction
+ * @return Writer for a new StoreFile in the tmp dir.
+ */
+ public StoreFile.Writer createWriterInTmp(int maxKeyCount,
+ Compression.Algorithm compression, boolean isCompaction) throws IOException;
+
// Compaction oriented methods
public boolean throttleCompaction(long compactionSize);
@@ -259,11 +290,32 @@ public interface Store extends HeapSize,
public CacheConfig getCacheConfig();
/**
- * @return the parent region hosting this store
+ * @return the parent region info hosting this store
*/
- public HRegion getHRegion();
+ public HRegionInfo getRegionInfo();
+
+ public RegionCoprocessorHost getCoprocessorHost();
+
+ public boolean areWritesEnabled();
+
+ /**
+ * @return The smallest mvcc readPoint across all the scanners in this
+ * region. Writes older than this readPoint, are included in every
+ * read operation.
+ */
+ public long getSmallestReadPoint();
public String getColumnFamilyName();
public String getTableName();
+
+ /*
+ * @param o Observer who wants to know about changes in set of Readers
+ */
+ public void addChangedReaderObserver(ChangedReadersObserver o);
+
+ /*
+ * @param o Observer no longer interested in changes in set of Readers.
+ */
+ public void deleteChangedReaderObserver(ChangedReadersObserver o);
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreScanner.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreScanner.java?rev=1445434&r1=1445433&r2=1445434&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreScanner.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreScanner.java Wed Feb 13 00:08:47 2013
@@ -32,19 +32,19 @@ import org.apache.hadoop.hbase.HConstant
import org.apache.hadoop.hbase.KeyValue;
import org.apache.hadoop.hbase.client.Scan;
import org.apache.hadoop.hbase.filter.Filter;
-import org.apache.hadoop.hbase.regionserver.HStore.ScanInfo;
+import org.apache.hadoop.hbase.regionserver.ScanInfo;
import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
/**
- * Scanner scans both the memstore and the HStore. Coalesce KeyValue stream
+ * Scanner scans both the memstore and the Store. Coalesce KeyValue stream
* into List<KeyValue> for a single row.
*/
@InterfaceAudience.Private
public class StoreScanner extends NonLazyKeyValueScanner
implements KeyValueScanner, InternalScanner, ChangedReadersObserver {
static final Log LOG = LogFactory.getLog(StoreScanner.class);
- protected HStore store;
+ protected Store store;
protected ScanQueryMatcher matcher;
protected KeyValueHeap heap;
protected boolean cacheBlocks;
@@ -75,7 +75,7 @@ public class StoreScanner extends NonLaz
protected KeyValue lastTop = null;
/** An internal constructor. */
- protected StoreScanner(HStore store, boolean cacheBlocks, Scan scan,
+ protected StoreScanner(Store store, boolean cacheBlocks, Scan scan,
final NavigableSet<byte[]> columns, long ttl, int minVersions) {
this.store = store;
this.cacheBlocks = cacheBlocks;
@@ -103,7 +103,7 @@ public class StoreScanner extends NonLaz
* @param columns which columns we are scanning
* @throws IOException
*/
- public StoreScanner(HStore store, ScanInfo scanInfo, Scan scan, final NavigableSet<byte[]> columns)
+ public StoreScanner(Store store, ScanInfo scanInfo, Scan scan, final NavigableSet<byte[]> columns)
throws IOException {
this(store, scan.getCacheBlocks(), scan, columns, scanInfo.getTtl(),
scanInfo.getMinVersions());
@@ -139,7 +139,7 @@ public class StoreScanner extends NonLaz
this.storeOffset = scan.getRowOffsetPerColumnFamily();
// Combine all seeked scanners with a heap
- heap = new KeyValueHeap(scanners, store.comparator);
+ heap = new KeyValueHeap(scanners, store.getComparator());
this.store.addChangedReaderObserver(this);
}
@@ -154,7 +154,7 @@ public class StoreScanner extends NonLaz
* @param smallestReadPoint the readPoint that we should use for tracking
* versions
*/
- public StoreScanner(HStore store, ScanInfo scanInfo, Scan scan,
+ public StoreScanner(Store store, ScanInfo scanInfo, Scan scan,
List<? extends KeyValueScanner> scanners, ScanType scanType,
long smallestReadPoint, long earliestPutTs) throws IOException {
this(store, false, scan, null, scanInfo.getTtl(),
@@ -171,11 +171,11 @@ public class StoreScanner extends NonLaz
}
// Combine all seeked scanners with a heap
- heap = new KeyValueHeap(scanners, store.comparator);
+ heap = new KeyValueHeap(scanners, store.getComparator());
}
/** Constructor for testing. */
- StoreScanner(final Scan scan, HStore.ScanInfo scanInfo,
+ StoreScanner(final Scan scan, ScanInfo scanInfo,
ScanType scanType, final NavigableSet<byte[]> columns,
final List<KeyValueScanner> scanners) throws IOException {
this(scan, scanInfo, scanType, columns, scanners,
@@ -183,7 +183,7 @@ public class StoreScanner extends NonLaz
}
// Constructor for testing.
- StoreScanner(final Scan scan, HStore.ScanInfo scanInfo,
+ StoreScanner(final Scan scan, ScanInfo scanInfo,
ScanType scanType, final NavigableSet<byte[]> columns,
final List<KeyValueScanner> scanners, long earliestPutTs)
throws IOException {
@@ -325,7 +325,8 @@ public class StoreScanner extends NonLaz
byte[] row = peeked.getBuffer();
int offset = peeked.getRowOffset();
short length = peeked.getRowLength();
- if ((matcher.row == null) || !Bytes.equals(row, offset, length, matcher.row, matcher.rowOffset, matcher.rowLength)) {
+ if ((matcher.row == null) || !Bytes.equals(row, offset, length, matcher.row,
+ matcher.rowOffset, matcher.rowLength)) {
this.countPerRow = 0;
matcher.setRow(row, offset, length);
}
@@ -490,7 +491,7 @@ public class StoreScanner extends NonLaz
if (this.heap == null && this.lastTop != null) {
resetScannerStack(this.lastTop);
if (this.heap.peek() == null
- || store.comparator.compareRows(this.lastTop, this.heap.peek()) != 0) {
+ || store.getComparator().compareRows(this.lastTop, this.heap.peek()) != 0) {
LOG.debug("Storescanner.peek() is changed where before = "
+ this.lastTop.toString() + ",and after = " + this.heap.peek());
this.lastTop = null;
@@ -517,7 +518,7 @@ public class StoreScanner extends NonLaz
}
// Combine all seeked scanners with a heap
- heap = new KeyValueHeap(scanners, store.comparator);
+ heap = new KeyValueHeap(scanners, store.getComparator());
// Reset the state of the Query Matcher and set to top row.
// Only reset and call setRow if the row changes; avoids confusing the
@@ -529,7 +530,8 @@ public class StoreScanner extends NonLaz
byte[] row = kv.getBuffer();
int offset = kv.getRowOffset();
short length = kv.getRowLength();
- if ((matcher.row == null) || !Bytes.equals(row, offset, length, matcher.row, matcher.rowOffset, matcher.rowLength)) {
+ if ((matcher.row == null) || !Bytes.equals(row, offset, length, matcher.row,
+ matcher.rowOffset, matcher.rowLength)) {
this.countPerRow = 0;
matcher.reset();
matcher.setRow(row, offset, length);
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/DefaultCompactor.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/DefaultCompactor.java?rev=1445434&r1=1445433&r2=1445434&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/DefaultCompactor.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/DefaultCompactor.java Wed Feb 13 00:08:47 2013
@@ -31,6 +31,7 @@ import org.apache.hadoop.hbase.HConstant
import org.apache.hadoop.hbase.KeyValue;
import org.apache.hadoop.hbase.client.Scan;
import org.apache.hadoop.hbase.io.compress.Compression;
+import org.apache.hadoop.hbase.regionserver.Store;
import org.apache.hadoop.hbase.regionserver.HStore;
import org.apache.hadoop.hbase.regionserver.InternalScanner;
import org.apache.hadoop.hbase.regionserver.MultiVersionConsistencyControl;
@@ -71,7 +72,7 @@ class DefaultCompactor extends Compactor
// Calculate maximum key count after compaction (for blooms)
// Also calculate earliest put timestamp if major compaction
int maxKeyCount = 0;
- HStore store = policy.store;
+ Store store = policy.store;
long earliestPutTs = HConstants.LATEST_TIMESTAMP;
for (StoreFile file: filesToCompact) {
StoreFile.Reader r = file.getReader();
@@ -127,14 +128,13 @@ class DefaultCompactor extends Compactor
StoreFile.Writer writer = null;
List<Path> newFiles = new ArrayList<Path>();
// Find the smallest read point across all the Scanners.
- long smallestReadPoint = store.getHRegion().getSmallestReadPoint();
+ long smallestReadPoint = store.getSmallestReadPoint();
MultiVersionConsistencyControl.setThreadReadPoint(smallestReadPoint);
try {
InternalScanner scanner = null;
try {
- if (store.getHRegion().getCoprocessorHost() != null) {
+ if (store.getCoprocessorHost() != null) {
scanner = store
- .getHRegion()
.getCoprocessorHost()
.preCompactScannerOpen(store, scanners,
majorCompaction ? ScanType.MAJOR_COMPACT : ScanType.MINOR_COMPACT, earliestPutTs);
@@ -147,9 +147,9 @@ class DefaultCompactor extends Compactor
scanner = new StoreScanner(store, store.getScanInfo(), scan, scanners,
scanType, smallestReadPoint, earliestPutTs);
}
- if (store.getHRegion().getCoprocessorHost() != null) {
+ if (store.getCoprocessorHost() != null) {
InternalScanner cpScanner =
- store.getHRegion().getCoprocessorHost().preCompact(store, scanner, scanType);
+ store.getCoprocessorHost().preCompact(store, scanner, scanType);
// NULL scanner returned from coprocessor hooks means skip normal processing
if (cpScanner == null) {
return newFiles; // an empty list
@@ -209,13 +209,14 @@ class DefaultCompactor extends Compactor
return newFiles;
}
- void isInterrupted(final HStore store, final StoreFile.Writer writer)
+ void isInterrupted(final Store store, final StoreFile.Writer writer)
throws IOException {
- if (store.getHRegion().areWritesEnabled()) return;
+ if (store.areWritesEnabled()) return;
// Else cleanup.
writer.close();
store.getFileSystem().delete(writer.getPath(), false);
throw new InterruptedIOException( "Aborting compaction of store " + store +
- " in region " + store.getHRegion() + " because it was interrupted.");
+ " in region " + store.getRegionInfo().getRegionNameAsString() +
+ " because it was interrupted.");
}
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java?rev=1445434&r1=1445433&r2=1445434&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java Wed Feb 13 00:08:47 2013
@@ -57,7 +57,7 @@ import org.apache.hadoop.hbase.protobuf.
import org.apache.hadoop.hbase.regionserver.HRegion;
import org.apache.hadoop.hbase.regionserver.InternalScanner;
import org.apache.hadoop.hbase.regionserver.RegionScanner;
-import org.apache.hadoop.hbase.regionserver.HStore;
+import org.apache.hadoop.hbase.regionserver.Store;
import org.apache.hadoop.hbase.regionserver.ScanType;
import org.apache.hadoop.hbase.regionserver.StoreFile;
import org.apache.hadoop.hbase.regionserver.wal.WALEdit;
@@ -766,7 +766,7 @@ public class AccessController extends Ba
@Override
public InternalScanner preCompact(ObserverContext<RegionCoprocessorEnvironment> e,
- final HStore store, final InternalScanner scanner, final ScanType scanType)
+ final Store store, final InternalScanner scanner, final ScanType scanType)
throws IOException {
requirePermission("compact", getTableName(e.getEnvironment()), null, null, Action.ADMIN);
return scanner;
@@ -774,7 +774,7 @@ public class AccessController extends Ba
@Override
public void preCompactSelection(final ObserverContext<RegionCoprocessorEnvironment> e,
- final HStore store, final List<StoreFile> candidates) throws IOException {
+ final Store store, final List<StoreFile> candidates) throws IOException {
requirePermission("compact", getTableName(e.getEnvironment()), null, null, Action.ADMIN);
}
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/SimpleRegionObserver.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/SimpleRegionObserver.java?rev=1445434&r1=1445433&r2=1445434&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/SimpleRegionObserver.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/SimpleRegionObserver.java Wed Feb 13 00:08:47 2013
@@ -47,7 +47,7 @@ import org.apache.hadoop.hbase.regionser
import org.apache.hadoop.hbase.regionserver.Leases;
import org.apache.hadoop.hbase.regionserver.RegionScanner;
import org.apache.hadoop.hbase.regionserver.ScanType;
-import org.apache.hadoop.hbase.regionserver.HStore;
+import org.apache.hadoop.hbase.regionserver.Store;
import org.apache.hadoop.hbase.regionserver.StoreFile;
import org.apache.hadoop.hbase.regionserver.wal.WALEdit;
import org.apache.hadoop.hbase.util.Bytes;
@@ -137,20 +137,22 @@ public class SimpleRegionObserver extend
}
@Override
- public InternalScanner preFlush(ObserverContext<RegionCoprocessorEnvironment> c, HStore store, InternalScanner scanner) {
+ public InternalScanner preFlush(ObserverContext<RegionCoprocessorEnvironment> c,
+ Store store, InternalScanner scanner) {
hadPreFlush = true;
return scanner;
}
@Override
public InternalScanner preFlushScannerOpen(final ObserverContext<RegionCoprocessorEnvironment> c,
- HStore store, KeyValueScanner memstoreScanner, InternalScanner s) throws IOException {
+ Store store, KeyValueScanner memstoreScanner, InternalScanner s) throws IOException {
hadPreFlushScannerOpen = true;
return null;
}
@Override
- public void postFlush(ObserverContext<RegionCoprocessorEnvironment> c, HStore store, StoreFile resultFile) {
+ public void postFlush(ObserverContext<RegionCoprocessorEnvironment> c,
+ Store store, StoreFile resultFile) {
hadPostFlush = true;
}
@@ -174,26 +176,27 @@ public class SimpleRegionObserver extend
@Override
public void preCompactSelection(ObserverContext<RegionCoprocessorEnvironment> c,
- HStore store, List<StoreFile> candidates) {
+ Store store, List<StoreFile> candidates) {
hadPreCompactSelect = true;
}
@Override
public void postCompactSelection(ObserverContext<RegionCoprocessorEnvironment> c,
- HStore store, ImmutableList<StoreFile> selected) {
+ Store store, ImmutableList<StoreFile> selected) {
hadPostCompactSelect = true;
}
@Override
public InternalScanner preCompact(ObserverContext<RegionCoprocessorEnvironment> e,
- HStore store, InternalScanner scanner, ScanType scanType) {
+ Store store, InternalScanner scanner, ScanType scanType) {
hadPreCompact = true;
return scanner;
}
@Override
- public InternalScanner preCompactScannerOpen(final ObserverContext<RegionCoprocessorEnvironment> c,
- HStore store, List<? extends KeyValueScanner> scanners, ScanType scanType, long earliestPutTs,
+ public InternalScanner preCompactScannerOpen(
+ final ObserverContext<RegionCoprocessorEnvironment> c,
+ Store store, List<? extends KeyValueScanner> scanners, ScanType scanType, long earliestPutTs,
InternalScanner s) throws IOException {
hadPreCompactScanner = true;
return null;
@@ -201,7 +204,7 @@ public class SimpleRegionObserver extend
@Override
public void postCompact(ObserverContext<RegionCoprocessorEnvironment> e,
- HStore store, StoreFile resultFile) {
+ Store store, StoreFile resultFile) {
hadPostCompact = true;
}
@@ -219,7 +222,7 @@ public class SimpleRegionObserver extend
@Override
public KeyValueScanner preStoreScannerOpen(final ObserverContext<RegionCoprocessorEnvironment> c,
- final HStore store, final Scan scan, final NavigableSet<byte[]> targetCols,
+ final Store store, final Scan scan, final NavigableSet<byte[]> targetCols,
final KeyValueScanner s) throws IOException {
hadPreStoreScannerOpen = true;
return null;
@@ -450,7 +453,7 @@ public class SimpleRegionObserver extend
@Override
public boolean postBulkLoadHFile(ObserverContext<RegionCoprocessorEnvironment> ctx,
- List<Pair<byte[], String>> familyPaths, boolean hasLoaded) throws IOException {
+ List<Pair<byte[], String>> familyPaths, boolean hasLoaded) throws IOException {
RegionCoprocessorEnvironment e = ctx.getEnvironment();
assertNotNull(e);
assertNotNull(e.getRegion());
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorInterface.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorInterface.java?rev=1445434&r1=1445433&r2=1445434&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorInterface.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorInterface.java Wed Feb 13 00:08:47 2013
@@ -46,7 +46,7 @@ import org.apache.hadoop.hbase.SmallTest
import org.apache.hadoop.hbase.client.Get;
import org.apache.hadoop.hbase.client.Scan;
import org.apache.hadoop.hbase.regionserver.HRegion;
-import org.apache.hadoop.hbase.regionserver.HStore;
+import org.apache.hadoop.hbase.regionserver.Store;
import org.apache.hadoop.hbase.regionserver.InternalScanner;
import org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost;
import org.apache.hadoop.hbase.regionserver.RegionScanner;
@@ -187,13 +187,13 @@ public class TestCoprocessorInterface ex
}
@Override
public InternalScanner preCompact(ObserverContext<RegionCoprocessorEnvironment> e,
- HStore store, InternalScanner scanner, ScanType scanType) {
+ Store store, InternalScanner scanner, ScanType scanType) {
preCompactCalled = true;
return scanner;
}
@Override
public void postCompact(ObserverContext<RegionCoprocessorEnvironment> e,
- HStore store, StoreFile resultFile) {
+ Store store, StoreFile resultFile) {
postCompactCalled = true;
}
@Override
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverInterface.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverInterface.java?rev=1445434&r1=1445433&r2=1445434&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverInterface.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverInterface.java Wed Feb 13 00:08:47 2013
@@ -60,7 +60,7 @@ import org.apache.hadoop.hbase.protobuf.
import org.apache.hadoop.hbase.regionserver.HRegion;
import org.apache.hadoop.hbase.regionserver.InternalScanner;
import org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost;
-import org.apache.hadoop.hbase.regionserver.HStore;
+import org.apache.hadoop.hbase.regionserver.Store;
import org.apache.hadoop.hbase.regionserver.ScanType;
import org.apache.hadoop.hbase.regionserver.StoreFile;
import org.apache.hadoop.hbase.util.Bytes;
@@ -316,7 +316,7 @@ public class TestRegionObserverInterface
@Override
public InternalScanner preCompact(ObserverContext<RegionCoprocessorEnvironment> e,
- HStore store, final InternalScanner scanner, final ScanType scanType) {
+ Store store, final InternalScanner scanner, final ScanType scanType) {
return new InternalScanner() {
@Override
public boolean next(List<KeyValue> results) throws IOException {
@@ -368,7 +368,7 @@ public class TestRegionObserverInterface
@Override
public void postCompact(ObserverContext<RegionCoprocessorEnvironment> e,
- HStore store, StoreFile resultFile) {
+ Store store, StoreFile resultFile) {
lastCompaction = EnvironmentEdgeManager.currentTimeMillis();
}
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/NoOpScanPolicyObserver.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/NoOpScanPolicyObserver.java?rev=1445434&r1=1445433&r2=1445434&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/NoOpScanPolicyObserver.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/NoOpScanPolicyObserver.java Wed Feb 13 00:08:47 2013
@@ -25,15 +25,14 @@ public class NoOpScanPolicyObserver exte
*/
@Override
public InternalScanner preFlushScannerOpen(final ObserverContext<RegionCoprocessorEnvironment> c,
- HStore store, KeyValueScanner memstoreScanner, InternalScanner s) throws IOException {
- HStore.ScanInfo oldSI = store.getScanInfo();
- HStore.ScanInfo scanInfo = new HStore.ScanInfo(store.getFamily(), oldSI.getTtl(),
+ Store store, KeyValueScanner memstoreScanner, InternalScanner s) throws IOException {
+ ScanInfo oldSI = store.getScanInfo();
+ ScanInfo scanInfo = new ScanInfo(store.getFamily(), oldSI.getTtl(),
oldSI.getTimeToPurgeDeletes(), oldSI.getComparator());
Scan scan = new Scan();
scan.setMaxVersions(oldSI.getMaxVersions());
return new StoreScanner(store, scanInfo, scan, Collections.singletonList(memstoreScanner),
- ScanType.MINOR_COMPACT, store.getHRegion().getSmallestReadPoint(),
- HConstants.OLDEST_TIMESTAMP);
+ ScanType.MINOR_COMPACT, store.getSmallestReadPoint(), HConstants.OLDEST_TIMESTAMP);
}
/**
@@ -41,21 +40,21 @@ public class NoOpScanPolicyObserver exte
*/
@Override
public InternalScanner preCompactScannerOpen(final ObserverContext<RegionCoprocessorEnvironment> c,
- HStore store, List<? extends KeyValueScanner> scanners, ScanType scanType, long earliestPutTs,
+ Store store, List<? extends KeyValueScanner> scanners, ScanType scanType, long earliestPutTs,
InternalScanner s) throws IOException {
// this demonstrates how to override the scanners default behavior
- HStore.ScanInfo oldSI = store.getScanInfo();
- HStore.ScanInfo scanInfo = new HStore.ScanInfo(store.getFamily(), oldSI.getTtl(),
+ ScanInfo oldSI = store.getScanInfo();
+ ScanInfo scanInfo = new ScanInfo(store.getFamily(), oldSI.getTtl(),
oldSI.getTimeToPurgeDeletes(), oldSI.getComparator());
Scan scan = new Scan();
scan.setMaxVersions(oldSI.getMaxVersions());
- return new StoreScanner(store, scanInfo, scan, scanners, scanType, store.getHRegion()
- .getSmallestReadPoint(), earliestPutTs);
+ return new StoreScanner(store, scanInfo, scan, scanners, scanType,
+ store.getSmallestReadPoint(), earliestPutTs);
}
@Override
public KeyValueScanner preStoreScannerOpen(final ObserverContext<RegionCoprocessorEnvironment> c,
- HStore store, final Scan scan, final NavigableSet<byte[]> targetCols, KeyValueScanner s)
+ Store store, final Scan scan, final NavigableSet<byte[]> targetCols, KeyValueScanner s)
throws IOException {
return new StoreScanner(store, store.getScanInfo(), scan, targetCols);
}
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompaction.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompaction.java?rev=1445434&r1=1445433&r2=1445434&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompaction.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompaction.java Wed Feb 13 00:08:47 2013
@@ -281,8 +281,8 @@ public class TestCompaction extends HBas
final int ttl = 1000;
for (Store hstore : this.r.stores.values()) {
HStore store = ((HStore) hstore);
- HStore.ScanInfo old = store.getScanInfo();
- HStore.ScanInfo si = new HStore.ScanInfo(old.getFamily(),
+ ScanInfo old = store.getScanInfo();
+ ScanInfo si = new ScanInfo(old.getFamily(),
old.getMinVersions(), old.getMaxVersions(), ttl,
old.getKeepDeletedCells(), 0, old.getComparator());
store.setScanInfo(si);
@@ -540,8 +540,8 @@ public class TestCompaction extends HBas
final int ttl = 1000;
for (Store hstore: this.r.stores.values()) {
HStore store = (HStore)hstore;
- HStore.ScanInfo old = store.getScanInfo();
- HStore.ScanInfo si = new HStore.ScanInfo(old.getFamily(),
+ ScanInfo old = store.getScanInfo();
+ ScanInfo si = new ScanInfo(old.getFamily(),
old.getMinVersions(), old.getMaxVersions(), ttl,
old.getKeepDeletedCells(), 0, old.getComparator());
store.setScanInfo(si);
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMemStore.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMemStore.java?rev=1445434&r1=1445433&r2=1445434&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMemStore.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMemStore.java Wed Feb 13 00:08:47 2013
@@ -34,7 +34,7 @@ import org.apache.commons.logging.LogFac
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hbase.*;
import org.apache.hadoop.hbase.client.Scan;
-import org.apache.hadoop.hbase.regionserver.HStore.ScanInfo;
+import org.apache.hadoop.hbase.regionserver.ScanInfo;
import org.apache.hadoop.hbase.util.Bytes;
import com.google.common.base.Joiner;
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestQueryMatcher.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestQueryMatcher.java?rev=1445434&r1=1445433&r2=1445434&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestQueryMatcher.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestQueryMatcher.java Wed Feb 13 00:08:47 2013
@@ -98,7 +98,7 @@ public class TestQueryMatcher extends HB
// 2,4,5
- ScanQueryMatcher qm = new ScanQueryMatcher(scan, new HStore.ScanInfo(fam2,
+ ScanQueryMatcher qm = new ScanQueryMatcher(scan, new ScanInfo(fam2,
0, 1, ttl, false, 0, rowComparator), get.getFamilyMap().get(fam2),
EnvironmentEdgeManager.currentTimeMillis() - ttl);
@@ -144,7 +144,7 @@ public class TestQueryMatcher extends HB
expected.add(ScanQueryMatcher.MatchCode.INCLUDE);
expected.add(ScanQueryMatcher.MatchCode.DONE);
- ScanQueryMatcher qm = new ScanQueryMatcher(scan, new HStore.ScanInfo(fam2,
+ ScanQueryMatcher qm = new ScanQueryMatcher(scan, new ScanInfo(fam2,
0, 1, ttl, false, 0, rowComparator), null,
EnvironmentEdgeManager.currentTimeMillis() - ttl);
@@ -198,7 +198,7 @@ public class TestQueryMatcher extends HB
};
long now = EnvironmentEdgeManager.currentTimeMillis();
- ScanQueryMatcher qm = new ScanQueryMatcher(scan, new HStore.ScanInfo(fam2,
+ ScanQueryMatcher qm = new ScanQueryMatcher(scan, new ScanInfo(fam2,
0, 1, testTTL, false, 0, rowComparator), get.getFamilyMap().get(fam2),
now - testTTL);
@@ -252,7 +252,7 @@ public class TestQueryMatcher extends HB
};
long now = EnvironmentEdgeManager.currentTimeMillis();
- ScanQueryMatcher qm = new ScanQueryMatcher(scan, new HStore.ScanInfo(fam2,
+ ScanQueryMatcher qm = new ScanQueryMatcher(scan, new ScanInfo(fam2,
0, 1, testTTL, false, 0, rowComparator), null,
now - testTTL);
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStore.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStore.java?rev=1445434&r1=1445433&r2=1445434&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStore.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStore.java Wed Feb 13 00:08:47 2013
@@ -324,7 +324,7 @@ public class TestStore extends TestCase
this.store = new HStore(storedir.getParent().getParent(),
this.store.getHRegion(),
this.store.getFamily(), fs, c);
- System.out.println(this.store.getHRegionInfo().getEncodedName());
+ System.out.println(this.store.getRegionInfo().getEncodedName());
assertEquals(2, this.store.getStorefilesCount());
result = HBaseTestingUtility.getFromStoreFile(store,
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreScanner.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreScanner.java?rev=1445434&r1=1445433&r2=1445434&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreScanner.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreScanner.java Wed Feb 13 00:08:47 2013
@@ -36,7 +36,7 @@ import org.apache.hadoop.hbase.KeyValue;
import org.apache.hadoop.hbase.KeyValueTestUtil;
import org.apache.hadoop.hbase.MediumTests;
import org.apache.hadoop.hbase.client.Scan;
-import org.apache.hadoop.hbase.regionserver.HStore.ScanInfo;
+import org.apache.hadoop.hbase.regionserver.ScanInfo;
import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.hbase.util.EnvironmentEdge;
import org.apache.hadoop.hbase.util.EnvironmentEdgeManagerTestHelper;
@@ -547,7 +547,7 @@ public class TestStoreScanner extends Te
List<KeyValueScanner> scanners = scanFixture(kvs);
Scan scan = new Scan();
scan.setMaxVersions(2);
- HStore.ScanInfo scanInfo = new HStore.ScanInfo(Bytes.toBytes("cf"),
+ ScanInfo scanInfo = new ScanInfo(Bytes.toBytes("cf"),
0 /* minVersions */,
2 /* maxVersions */, 500 /* ttl */,
false /* keepDeletedCells */,
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestCoprocessorScanPolicy.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestCoprocessorScanPolicy.java?rev=1445434&r1=1445433&r2=1445434&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestCoprocessorScanPolicy.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestCoprocessorScanPolicy.java Wed Feb 13 00:08:47 2013
@@ -47,7 +47,8 @@ import org.apache.hadoop.hbase.coprocess
import org.apache.hadoop.hbase.regionserver.InternalScanner;
import org.apache.hadoop.hbase.regionserver.KeyValueScanner;
import org.apache.hadoop.hbase.regionserver.ScanType;
-import org.apache.hadoop.hbase.regionserver.HStore;
+import org.apache.hadoop.hbase.regionserver.Store;
+import org.apache.hadoop.hbase.regionserver.ScanInfo;
import org.apache.hadoop.hbase.regionserver.StoreScanner;
import org.apache.hadoop.hbase.regionserver.wal.WALEdit;
import org.apache.hadoop.hbase.util.Bytes;
@@ -204,53 +205,55 @@ public class TestCoprocessorScanPolicy {
}
@Override
- public InternalScanner preFlushScannerOpen(final ObserverContext<RegionCoprocessorEnvironment> c,
- HStore store, KeyValueScanner memstoreScanner, InternalScanner s) throws IOException {
+ public InternalScanner preFlushScannerOpen(
+ final ObserverContext<RegionCoprocessorEnvironment> c,
+ Store store, KeyValueScanner memstoreScanner, InternalScanner s) throws IOException {
Long newTtl = ttls.get(store.getTableName());
if (newTtl != null) {
System.out.println("PreFlush:" + newTtl);
}
Integer newVersions = versions.get(store.getTableName());
- HStore.ScanInfo oldSI = store.getScanInfo();
+ ScanInfo oldSI = store.getScanInfo();
HColumnDescriptor family = store.getFamily();
- HStore.ScanInfo scanInfo = new HStore.ScanInfo(family.getName(), family.getMinVersions(),
+ ScanInfo scanInfo = new ScanInfo(family.getName(), family.getMinVersions(),
newVersions == null ? family.getMaxVersions() : newVersions,
newTtl == null ? oldSI.getTtl() : newTtl, family.getKeepDeletedCells(),
oldSI.getTimeToPurgeDeletes(), oldSI.getComparator());
Scan scan = new Scan();
scan.setMaxVersions(newVersions == null ? oldSI.getMaxVersions() : newVersions);
return new StoreScanner(store, scanInfo, scan, Collections.singletonList(memstoreScanner),
- ScanType.MINOR_COMPACT, store.getHRegion().getSmallestReadPoint(),
+ ScanType.MINOR_COMPACT, store.getSmallestReadPoint(),
HConstants.OLDEST_TIMESTAMP);
}
@Override
- public InternalScanner preCompactScannerOpen(final ObserverContext<RegionCoprocessorEnvironment> c,
- HStore store, List<? extends KeyValueScanner> scanners, ScanType scanType,
+ public InternalScanner preCompactScannerOpen(
+ final ObserverContext<RegionCoprocessorEnvironment> c,
+ Store store, List<? extends KeyValueScanner> scanners, ScanType scanType,
long earliestPutTs, InternalScanner s) throws IOException {
Long newTtl = ttls.get(store.getTableName());
Integer newVersions = versions.get(store.getTableName());
- HStore.ScanInfo oldSI = store.getScanInfo();
+ ScanInfo oldSI = store.getScanInfo();
HColumnDescriptor family = store.getFamily();
- HStore.ScanInfo scanInfo = new HStore.ScanInfo(family.getName(), family.getMinVersions(),
+ ScanInfo scanInfo = new ScanInfo(family.getName(), family.getMinVersions(),
newVersions == null ? family.getMaxVersions() : newVersions,
newTtl == null ? oldSI.getTtl() : newTtl, family.getKeepDeletedCells(),
oldSI.getTimeToPurgeDeletes(), oldSI.getComparator());
Scan scan = new Scan();
scan.setMaxVersions(newVersions == null ? oldSI.getMaxVersions() : newVersions);
- return new StoreScanner(store, scanInfo, scan, scanners, scanType, store.getHRegion()
- .getSmallestReadPoint(), earliestPutTs);
+ return new StoreScanner(store, scanInfo, scan, scanners, scanType,
+ store.getSmallestReadPoint(), earliestPutTs);
}
@Override
public KeyValueScanner preStoreScannerOpen(
- final ObserverContext<RegionCoprocessorEnvironment> c, HStore store, final Scan scan,
+ final ObserverContext<RegionCoprocessorEnvironment> c, Store store, final Scan scan,
final NavigableSet<byte[]> targetCols, KeyValueScanner s) throws IOException {
Long newTtl = ttls.get(store.getTableName());
Integer newVersions = versions.get(store.getTableName());
- HStore.ScanInfo oldSI = store.getScanInfo();
+ ScanInfo oldSI = store.getScanInfo();
HColumnDescriptor family = store.getFamily();
- HStore.ScanInfo scanInfo = new HStore.ScanInfo(family.getName(), family.getMinVersions(),
+ ScanInfo scanInfo = new ScanInfo(family.getName(), family.getMinVersions(),
newVersions == null ? family.getMaxVersions() : newVersions,
newTtl == null ? oldSI.getTtl() : newTtl, family.getKeepDeletedCells(),
oldSI.getTimeToPurgeDeletes(), oldSI.getComparator());