You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2013/07/10 20:50:23 UTC
svn commit: r1501909 [3/4] - in /hbase/trunk:
hbase-client/src/main/java/org/apache/hadoop/hbase/
hbase-client/src/main/java/org/apache/hadoop/hbase/catalog/
hbase-client/src/main/java/org/apache/hadoop/hbase/client/
hbase-client/src/main/java/org/apac...
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/handler/TotesHRegionInfo.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/handler/TotesHRegionInfo.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/handler/TotesHRegionInfo.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/handler/TotesHRegionInfo.java Wed Jul 10 18:50:21 2013
@@ -33,5 +33,5 @@ public interface TotesHRegionInfo {
/**
* @return HRegionInfo instance.
*/
- HRegionInfo getHRegionInfo();
+ public HRegionInfo getHRegionInfo();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandler.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandler.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandler.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandler.java Wed Jul 10 18:50:21 2013
@@ -29,16 +29,16 @@ import com.google.protobuf.Message;
*/
@InterfaceAudience.Private
public interface MonitoredRPCHandler extends MonitoredTask {
- String getRPC();
- String getRPC(boolean withParams);
- long getRPCPacketLength();
- String getClient();
- long getRPCStartTime();
- long getRPCQueueTime();
- boolean isRPCRunning();
- boolean isOperationRunning();
+ public abstract String getRPC();
+ public abstract String getRPC(boolean withParams);
+ public abstract long getRPCPacketLength();
+ public abstract String getClient();
+ public abstract long getRPCStartTime();
+ public abstract long getRPCQueueTime();
+ public abstract boolean isRPCRunning();
+ public abstract boolean isOperationRunning();
- void setRPC(String methodName, Object[] params, long queueTime);
- void setRPCPacket(Message param);
- void setConnection(String clientAddress, int remotePort);
+ public abstract void setRPC(String methodName, Object [] params, long queueTime);
+ public abstract void setRPCPacket(Message param);
+ public abstract void setConnection(String clientAddress, int remotePort);
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java Wed Jul 10 18:50:21 2013
@@ -32,47 +32,47 @@ public interface MonitoredTask extends C
ABORTED;
}
- long getStartTime();
- String getDescription();
- String getStatus();
- long getStatusTime();
- State getState();
- long getStateTime();
- long getCompletionTimestamp();
-
- void markComplete(String msg);
- void pause(String msg);
- void resume(String msg);
- void abort(String msg);
- void expireNow();
+ public abstract long getStartTime();
+ public abstract String getDescription();
+ public abstract String getStatus();
+ public abstract long getStatusTime();
+ public abstract State getState();
+ public abstract long getStateTime();
+ public abstract long getCompletionTimestamp();
+
+ public abstract void markComplete(String msg);
+ public abstract void pause(String msg);
+ public abstract void resume(String msg);
+ public abstract void abort(String msg);
+ public abstract void expireNow();
- void setStatus(String status);
- void setDescription(String description);
+ public abstract void setStatus(String status);
+ public abstract void setDescription(String description);
/**
* Explicitly mark this status as able to be cleaned up,
* even though it might not be complete.
*/
- void cleanup();
+ public abstract void cleanup();
/**
* Public exposure of Object.clone() in order to allow clients to easily
* capture current state.
* @return a copy of the object whose references will not change
*/
- MonitoredTask clone();
+ public abstract MonitoredTask clone();
/**
* Creates a string map of internal details for extensible exposure of
* monitored tasks.
* @return A Map containing information for this task.
*/
- Map<String, Object> toMap() throws IOException;
+ public abstract Map<String, Object> toMap() throws IOException;
/**
* Creates a JSON object for parseable exposure of monitored tasks.
* @return An encoded JSON object containing information for this task.
*/
- String toJSON() throws IOException;
+ public abstract String toJSON() throws IOException;
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinatorRpcs.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinatorRpcs.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinatorRpcs.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinatorRpcs.java Wed Jul 10 18:50:21 2013
@@ -39,7 +39,7 @@ public interface ProcedureCoordinatorRpc
* @param listener
* @return true if succeed, false if encountered initialization errors.
*/
- boolean start(final ProcedureCoordinator listener);
+ public boolean start(final ProcedureCoordinator listener);
/**
* Notify the members that the coordinator has aborted the procedure and that it should release
@@ -50,7 +50,7 @@ public interface ProcedureCoordinatorRpc
* @throws IOException if the rpcs can't reach the other members of the procedure (and can't
* recover).
*/
- void sendAbortToMembers(Procedure procName, ForeignException cause) throws IOException;
+ public void sendAbortToMembers(Procedure procName, ForeignException cause) throws IOException;
/**
* Notify the members to acquire barrier for the procedure
@@ -61,7 +61,7 @@ public interface ProcedureCoordinatorRpc
* @throws IllegalArgumentException if the procedure was already marked as failed
* @throws IOException if we can't reach the remote notification mechanism
*/
- void sendGlobalBarrierAcquire(Procedure procName, byte[] info, List<String> members)
+ public void sendGlobalBarrierAcquire(Procedure procName, byte[] info, List<String> members)
throws IOException, IllegalArgumentException;
/**
@@ -74,12 +74,12 @@ public interface ProcedureCoordinatorRpc
* @param members members to tell we have reached in-barrier phase
* @throws IOException if we can't reach the remote notification mechanism
*/
- void sendGlobalBarrierReached(Procedure procName, List<String> members) throws IOException;
+ public void sendGlobalBarrierReached(Procedure procName, List<String> members) throws IOException;
/**
* Notify Members to reset the distributed state for procedure
* @param procName name of the procedure to reset
* @throws IOException if the remote notification mechanism cannot be reached
*/
- void resetMembers(Procedure procName) throws IOException;
+ public void resetMembers(Procedure procName) throws IOException;
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureMemberRpcs.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureMemberRpcs.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureMemberRpcs.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureMemberRpcs.java Wed Jul 10 18:50:21 2013
@@ -35,13 +35,13 @@ public interface ProcedureMemberRpcs ext
/**
* Initialize and start any threads or connections the member needs.
*/
- void start(final String memberName, final ProcedureMember member);
+ public void start(final String memberName, final ProcedureMember member);
/**
* Each subprocedure is being executed on a member. This is the identifier for the member.
* @return the member name
*/
- String getMemberName();
+ public String getMemberName();
/**
* Notify the coordinator that we aborted the specified {@link Subprocedure}
@@ -51,7 +51,7 @@ public interface ProcedureMemberRpcs ext
* @throws IOException thrown when the rpcs can't reach the other members of the procedure (and
* thus can't recover).
*/
- void sendMemberAborted(Subprocedure sub, ForeignException cause) throws IOException;
+ public void sendMemberAborted(Subprocedure sub, ForeignException cause) throws IOException;
/**
* Notify the coordinator that the specified {@link Subprocedure} has acquired the locally required
@@ -60,7 +60,7 @@ public interface ProcedureMemberRpcs ext
* @param sub the specified {@link Subprocedure}
* @throws IOException if we can't reach the coordinator
*/
- void sendMemberAcquired(Subprocedure sub) throws IOException;
+ public void sendMemberAcquired(Subprocedure sub) throws IOException;
/**
* Notify the coordinator that the specified {@link Subprocedure} has completed the work that
@@ -69,5 +69,5 @@ public interface ProcedureMemberRpcs ext
* @param sub the specified {@link Subprocedure}
* @throws IOException if we can't reach the coordinator
*/
- void sendMemberCompleted(Subprocedure sub) throws IOException;
-}
+ public void sendMemberCompleted(Subprocedure sub) throws IOException;
+}
\ No newline at end of file
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/SubprocedureFactory.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/SubprocedureFactory.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/SubprocedureFactory.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/SubprocedureFactory.java Wed Jul 10 18:50:21 2013
@@ -36,5 +36,5 @@ public interface SubprocedureFactory {
* request
* @throws IllegalStateException if the current runner cannot accept any more new requests
*/
- Subprocedure buildSubprocedure(String procName, byte[] procArgs);
+ public Subprocedure buildSubprocedure(String procName, byte[] procArgs);
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ColumnTracker.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ColumnTracker.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ColumnTracker.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ColumnTracker.java Wed Jul 10 18:50:21 2013
@@ -56,26 +56,25 @@ public interface ColumnTracker {
* @throws IOException in case there is an internal consistency problem
* caused by a data corruption.
*/
- ScanQueryMatcher.MatchCode checkColumn(
- byte[] bytes, int offset, int length, long ttl, byte type, boolean ignoreCount
- )
+ public ScanQueryMatcher.MatchCode checkColumn(byte[] bytes, int offset,
+ int length, long ttl, byte type, boolean ignoreCount)
throws IOException;
/**
* Updates internal variables in between files
*/
- void update();
+ public void update();
/**
* Resets the Matcher
*/
- void reset();
+ public void reset();
/**
*
* @return <code>true</code> when done.
*/
- boolean done();
+ public boolean done();
/**
* Used by matcher and scan/get to get a hint of the next column
@@ -88,14 +87,13 @@ public interface ColumnTracker {
*
* @return null, or a ColumnCount that we should seek to
*/
- ColumnCount getColumnHint();
+ public ColumnCount getColumnHint();
/**
* Retrieve the MatchCode for the next row or column
*/
- MatchCode getNextRowOrNextColumn(
- byte[] bytes, int offset, int qualLength
- );
+ public MatchCode getNextRowOrNextColumn(byte[] bytes, int offset,
+ int qualLength);
/**
* Give the tracker a chance to declare it's done based on only the timestamp
@@ -104,5 +102,5 @@ public interface ColumnTracker {
* @param timestamp
* @return <code>true</code> to early out based on timestamp.
*/
- boolean isDone(long timestamp);
+ public boolean isDone(long timestamp);
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactionRequestor.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactionRequestor.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactionRequestor.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactionRequestor.java Wed Jul 10 18:50:21 2013
@@ -34,7 +34,7 @@ public interface CompactionRequestor {
* compactions were started
* @throws IOException
*/
- List<CompactionRequest> requestCompaction(final HRegion r, final String why)
+ public List<CompactionRequest> requestCompaction(final HRegion r, final String why)
throws IOException;
/**
@@ -47,9 +47,8 @@ public interface CompactionRequestor {
* compactions were started
* @throws IOException
*/
- List<CompactionRequest> requestCompaction(
- final HRegion r, final String why, List<Pair<CompactionRequest, Store>> requests
- )
+ public List<CompactionRequest> requestCompaction(final HRegion r, final String why,
+ List<Pair<CompactionRequest, Store>> requests)
throws IOException;
/**
@@ -61,9 +60,8 @@ public interface CompactionRequestor {
* @return The created {@link CompactionRequest} or <tt>null</tt> if no compaction was started.
* @throws IOException
*/
- CompactionRequest requestCompaction(
- final HRegion r, final Store s, final String why, CompactionRequest request
- ) throws IOException;
+ public CompactionRequest requestCompaction(final HRegion r, final Store s, final String why,
+ CompactionRequest request) throws IOException;
/**
* @param r Region to compact
@@ -76,9 +74,8 @@ public interface CompactionRequestor {
* compactions were started.
* @throws IOException
*/
- List<CompactionRequest> requestCompaction(
- final HRegion r, final String why, int pri, List<Pair<CompactionRequest, Store>> requests
- ) throws IOException;
+ public List<CompactionRequest> requestCompaction(final HRegion r, final String why, int pri,
+ List<Pair<CompactionRequest, Store>> requests) throws IOException;
/**
* @param r Region to compact
@@ -90,7 +87,6 @@ public interface CompactionRequestor {
* @return The created {@link CompactionRequest} or <tt>null</tt> if no compaction was started
* @throws IOException
*/
- CompactionRequest requestCompaction(
- final HRegion r, final Store s, final String why, int pri, CompactionRequest request
- ) throws IOException;
+ public CompactionRequest requestCompaction(final HRegion r, final Store s, final String why,
+ int pri, CompactionRequest request) throws IOException;
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DeleteTracker.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DeleteTracker.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DeleteTracker.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DeleteTracker.java Wed Jul 10 18:50:21 2013
@@ -43,9 +43,8 @@ public interface DeleteTracker {
* @param timestamp timestamp
* @param type delete type as byte
*/
- void add(
- byte[] buffer, int qualifierOffset, int qualifierLength, long timestamp, byte type
- );
+ public void add(byte [] buffer, int qualifierOffset, int qualifierLength,
+ long timestamp, byte type);
/**
* Check if the specified KeyValue buffer has been deleted by a previously
@@ -56,14 +55,13 @@ public interface DeleteTracker {
* @param timestamp timestamp
* @return deleteResult The result tells whether the KeyValue is deleted and why
*/
- DeleteResult isDeleted(
- byte[] buffer, int qualifierOffset, int qualifierLength, long timestamp
- );
+ public DeleteResult isDeleted(byte [] buffer, int qualifierOffset,
+ int qualifierLength, long timestamp);
/**
* @return true if there are no current delete, false otherwise
*/
- boolean isEmpty();
+ public boolean isEmpty();
/**
* Called at the end of every StoreFile.
@@ -71,14 +69,14 @@ public interface DeleteTracker {
* Many optimized implementations of Trackers will require an update at
* when the end of each StoreFile is reached.
*/
- void update();
+ public void update();
/**
* Called between rows.
* <p>
* This clears everything as if a new DeleteTracker was instantiated.
*/
- void reset();
+ public void reset();
/**
@@ -104,7 +102,7 @@ public interface DeleteTracker {
* Based on the delete result, the ScanQueryMatcher will decide the next
* operation
*/
- enum DeleteResult {
+ public static enum DeleteResult {
FAMILY_DELETED, // The KeyValue is deleted by a delete family.
FAMILY_VERSION_DELETED, // The KeyValue is deleted by a delete family version.
COLUMN_DELETED, // The KeyValue is deleted by a delete column.
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java Wed Jul 10 18:50:21 2013
@@ -5602,7 +5602,7 @@ public class HRegion implements HeapSize
* bulkLoadHFile() to perform any necessary
* pre/post processing of a given bulkload call
*/
- public interface BulkLoadListener {
+ public static interface BulkLoadListener {
/**
* Called before an HFile is actually loaded
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/InternalScanner.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/InternalScanner.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/InternalScanner.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/InternalScanner.java Wed Jul 10 18:50:21 2013
@@ -47,7 +47,7 @@ public interface InternalScanner extends
* @return true if more rows exist after this one, false if scanner is done
* @throws IOException e
*/
- boolean next(List<KeyValue> results) throws IOException;
+ public boolean next(List<KeyValue> results) throws IOException;
/**
* Grab the next row's worth of values with a limit on the number of values
@@ -57,11 +57,11 @@ public interface InternalScanner extends
* @return true if more rows exist after this one, false if scanner is done
* @throws IOException e
*/
- boolean next(List<KeyValue> result, int limit) throws IOException;
+ public boolean next(List<KeyValue> result, int limit) throws IOException;
/**
* Closes the scanner and releases any resources it has allocated
* @throws IOException
*/
- void close() throws IOException;
+ public void close() throws IOException;
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/KeyValueScanner.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/KeyValueScanner.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/KeyValueScanner.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/KeyValueScanner.java Wed Jul 10 18:50:21 2013
@@ -34,20 +34,20 @@ public interface KeyValueScanner {
* Look at the next KeyValue in this scanner, but do not iterate scanner.
* @return the next KeyValue
*/
- KeyValue peek();
+ public KeyValue peek();
/**
* Return the next KeyValue in this scanner, iterating the scanner
* @return the next KeyValue
*/
- KeyValue next() throws IOException;
+ public KeyValue next() throws IOException;
/**
* Seek the scanner at or after the specified KeyValue.
* @param key seek value
* @return true if scanner has values left, false if end of scanner
*/
- boolean seek(KeyValue key) throws IOException;
+ public boolean seek(KeyValue key) throws IOException;
/**
* Reseek the scanner at or after the specified KeyValue.
@@ -57,7 +57,7 @@ public interface KeyValueScanner {
* @param key seek value (should be non-null)
* @return true if scanner has values left, false if end of scanner
*/
- boolean reseek(KeyValue key) throws IOException;
+ public boolean reseek(KeyValue key) throws IOException;
/**
* Get the sequence id associated with this KeyValueScanner. This is required
@@ -65,12 +65,12 @@ public interface KeyValueScanner {
* The default implementation for this would be to return 0. A file having
* lower sequence id will be considered to be the older one.
*/
- long getSequenceID();
+ public long getSequenceID();
/**
* Close the KeyValue scanner.
*/
- void close();
+ public void close();
/**
* Allows to filter out scanners (both StoreFile and memstore) that we don't
@@ -82,9 +82,8 @@ public interface KeyValueScanner {
* this query, based on TTL
* @return true if the scanner should be included in the query
*/
- boolean shouldUseScanner(
- Scan scan, SortedSet<byte[]> columns, long oldestUnexpiredTS
- );
+ public boolean shouldUseScanner(Scan scan, SortedSet<byte[]> columns,
+ long oldestUnexpiredTS);
// "Lazy scanner" optimizations
@@ -98,7 +97,7 @@ public interface KeyValueScanner {
* @param forward do a forward-only "reseek" instead of a random-access seek
* @param useBloom whether to enable multi-column Bloom filter optimization
*/
- boolean requestSeek(KeyValue kv, boolean forward, boolean useBloom)
+ public boolean requestSeek(KeyValue kv, boolean forward, boolean useBloom)
throws IOException;
/**
@@ -107,7 +106,7 @@ public interface KeyValueScanner {
* store scanner bubbles up to the top of the key-value heap. This method is
* then used to ensure the top store file scanner has done a seek operation.
*/
- boolean realSeekDone();
+ public boolean realSeekDone();
/**
* Does the real seek operation in case it was skipped by
@@ -116,11 +115,11 @@ public interface KeyValueScanner {
* of the scanners). The easiest way to achieve this is to call
* {@link #realSeekDone()} first.
*/
- void enforceSeek() throws IOException;
+ public void enforceSeek() throws IOException;
/**
* @return true if this is a file scanner. Otherwise a memory scanner is
* assumed.
*/
- boolean isFileScanner();
+ public boolean isFileScanner();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/LastSequenceId.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/LastSequenceId.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/LastSequenceId.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/LastSequenceId.java Wed Jul 10 18:50:21 2013
@@ -29,5 +29,5 @@ public interface LastSequenceId {
* @param regionName Encoded region name
* @return Last flushed sequence Id for regionName or -1 if it can't be determined
*/
- long getLastSequenceId(byte[] regionName);
+ public long getLastSequenceId(byte[] regionName);
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/LeaseListener.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/LeaseListener.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/LeaseListener.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/LeaseListener.java Wed Jul 10 18:50:21 2013
@@ -32,5 +32,5 @@ import org.apache.hadoop.classification.
@InterfaceAudience.Private
public interface LeaseListener {
/** When a lease expires, this method is called. */
- void leaseExpired();
+ public void leaseExpired();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/OnlineRegions.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/OnlineRegions.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/OnlineRegions.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/OnlineRegions.java Wed Jul 10 18:50:21 2013
@@ -35,7 +35,7 @@ interface OnlineRegions extends Server {
* Add to online regions.
* @param r
*/
- void addToOnlineRegions(final HRegion r);
+ public void addToOnlineRegions(final HRegion r);
/**
* This method removes HRegion corresponding to hri from the Map of onlineRegions.
@@ -44,7 +44,7 @@ interface OnlineRegions extends Server {
* @param destination Destination, if any, null otherwise.
* @return True if we removed a region from online list.
*/
- boolean removeFromOnlineRegions(final HRegion r, ServerName destination);
+ public boolean removeFromOnlineRegions(final HRegion r, ServerName destination);
/**
* Return {@link HRegion} instance.
@@ -54,7 +54,7 @@ interface OnlineRegions extends Server {
* @return HRegion for the passed encoded <code>encodedRegionName</code> or
* null if named region is not member of the online regions.
*/
- HRegion getFromOnlineRegions(String encodedRegionName);
+ public HRegion getFromOnlineRegions(String encodedRegionName);
/**
* Get all online regions of a table in this RS.
@@ -62,5 +62,5 @@ interface OnlineRegions extends Server {
* @return List of HRegion
* @throws java.io.IOException
*/
- List<HRegion> getOnlineRegions(byte[] tableName) throws IOException;
+ public List<HRegion> getOnlineRegions(byte[] tableName) throws IOException;
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionScanner.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionScanner.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionScanner.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionScanner.java Wed Jul 10 18:50:21 2013
@@ -34,13 +34,13 @@ public interface RegionScanner extends I
/**
* @return The RegionInfo for this scanner.
*/
- HRegionInfo getRegionInfo();
+ public HRegionInfo getRegionInfo();
/**
* @return True if a filter indicates that this scanner will return no further rows.
* @throws IOException in case of I/O failure on a filter.
*/
- boolean isFilterDone() throws IOException;
+ public boolean isFilterDone() throws IOException;
/**
* Do a reseek to the required row. Should not be used to seek to a key which
@@ -52,17 +52,17 @@ public interface RegionScanner extends I
* if row is null
*
*/
- boolean reseek(byte[] row) throws IOException;
+ public boolean reseek(byte[] row) throws IOException;
/**
* @return The preferred max buffersize. See {@link Scan#setMaxResultSize(long)}
*/
- long getMaxResultSize();
+ public long getMaxResultSize();
/**
* @return The Scanner's MVCC readPt see {@link MultiVersionConsistencyControl}
*/
- long getMvccReadPoint();
+ public long getMvccReadPoint();
/**
* Grab the next row's worth of values with the default limit on the number of values
@@ -74,7 +74,7 @@ public interface RegionScanner extends I
* @return true if more rows exist after this one, false if scanner is done
* @throws IOException e
*/
- boolean nextRaw(List<KeyValue> result) throws IOException;
+ public boolean nextRaw(List<KeyValue> result) throws IOException;
/**
* Grab the next row's worth of values with a limit on the number of values
@@ -102,5 +102,5 @@ public interface RegionScanner extends I
* @return true if more rows exist after this one, false if scanner is done
* @throws IOException e
*/
- boolean nextRaw(List<KeyValue> result, int limit) throws IOException;
+ public boolean nextRaw(List<KeyValue> result, int limit) throws IOException;
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerServices.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerServices.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerServices.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerServices.java Wed Jul 10 18:50:21 2013
@@ -40,31 +40,31 @@ public interface RegionServerServices ex
/**
* @return True if this regionserver is stopping.
*/
- boolean isStopping();
+ public boolean isStopping();
/** @return the HLog for a particular region. Pass null for getting the
* default (common) WAL */
- HLog getWAL(HRegionInfo regionInfo) throws IOException;
+ public HLog getWAL(HRegionInfo regionInfo) throws IOException;
/**
* @return Implementation of {@link CompactionRequestor} or null.
*/
- CompactionRequestor getCompactionRequester();
+ public CompactionRequestor getCompactionRequester();
/**
* @return Implementation of {@link FlushRequester} or null.
*/
- FlushRequester getFlushRequester();
+ public FlushRequester getFlushRequester();
/**
* @return the RegionServerAccounting for this Region Server
*/
- RegionServerAccounting getRegionServerAccounting();
+ public RegionServerAccounting getRegionServerAccounting();
/**
* @return RegionServer's instance of {@link TableLockManager}
*/
- TableLockManager getTableLockManager();
+ public TableLockManager getTableLockManager();
/**
* Tasks to perform after region open to complete deploy of region on
@@ -75,42 +75,42 @@ public interface RegionServerServices ex
* @throws KeeperException
* @throws IOException
*/
- void postOpenDeployTasks(final HRegion r, final CatalogTracker ct)
+ public void postOpenDeployTasks(final HRegion r, final CatalogTracker ct)
throws KeeperException, IOException;
/**
* Returns a reference to the region server's RPC server
*/
- RpcServerInterface getRpcServer();
+ public RpcServerInterface getRpcServer();
/**
* Get the regions that are currently being opened or closed in the RS
* @return map of regions in transition in this RS
*/
- ConcurrentMap<byte[], Boolean> getRegionsInTransitionInRS();
+ public ConcurrentMap<byte[], Boolean> getRegionsInTransitionInRS();
/**
* @return Return the FileSystem object used by the regionserver
*/
- FileSystem getFileSystem();
+ public FileSystem getFileSystem();
/**
* @return The RegionServer's "Leases" service
*/
- Leases getLeases();
+ public Leases getLeases();
/**
* @return hbase executor service
*/
- ExecutorService getExecutorService();
+ public ExecutorService getExecutorService();
/**
* @return The RegionServer's CatalogTracker
*/
- CatalogTracker getCatalogTracker();
+ public CatalogTracker getCatalogTracker();
/**
* @return set of recovering regions on the hosting region server
*/
- Map<String, HRegion> getRecoveringRegions();
+ public Map<String, HRegion> getRecoveringRegions();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationService.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationService.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationService.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationService.java Wed Jul 10 18:50:21 2013
@@ -38,18 +38,17 @@ public interface ReplicationService {
* Initializes the replication service object.
* @throws IOException
*/
- void initialize(
- Server rs, FileSystem fs, Path logdir, Path oldLogDir
- ) throws IOException;
+ public void initialize(Server rs, FileSystem fs, Path logdir,
+ Path oldLogDir) throws IOException;
/**
* Start replication services.
* @throws IOException
*/
- void startReplicationService() throws IOException;
+ public void startReplicationService() throws IOException;
/**
* Stops replication service.
*/
- void stopReplicationService();
+ public void stopReplicationService();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSinkService.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSinkService.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSinkService.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSinkService.java Wed Jul 10 18:50:21 2013
@@ -38,5 +38,5 @@ public interface ReplicationSinkService
* @param cells Cells that the WALEntries refer to (if cells is non-null)
* @throws IOException
*/
- void replicateLogEntries(List<WALEntry> entries, CellScanner cells) throws IOException;
-}
+ public void replicateLogEntries(List<WALEntry> entries, CellScanner cells) throws IOException;
+}
\ No newline at end of file
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.java Wed Jul 10 18:50:21 2013
@@ -32,5 +32,5 @@ public interface ReplicationSourceServic
* Returns a WALObserver for the service. This is needed to
* observe log rolls and log archival events.
*/
- WALActionsListener getWALActionsListener();
-}
+ public WALActionsListener getWALActionsListener();
+}
\ No newline at end of file
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java Wed Jul 10 18:50:21 2013
@@ -640,13 +640,13 @@ public class SplitLogWorker extends ZooK
* is better to have workers prepare the task and then have the
* {@link SplitLogManager} commit the work in SplitLogManager.TaskFinisher
*/
- public interface TaskExecutor {
- enum Status {
+ static public interface TaskExecutor {
+ static public enum Status {
DONE(),
ERR(),
RESIGNED(),
PREEMPTED()
}
- Status exec(String name, CancelableProgressable p);
+ public Status exec(String name, CancelableProgressable p);
}
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java Wed Jul 10 18:50:21 2013
@@ -50,13 +50,14 @@ public interface Store extends HeapSize,
/* The default priority for user-specified compaction requests.
* The user gets top priority unless we have blocking compactions. (Pri <= 0)
- */ int PRIORITY_USER = 1;
- int NO_PRIORITY = Integer.MIN_VALUE;
+ */
+ public static final int PRIORITY_USER = 1;
+ public static final int NO_PRIORITY = Integer.MIN_VALUE;
// General Accessors
- KeyValue.KVComparator getComparator();
+ public KeyValue.KVComparator getComparator();
- Collection<StoreFile> getStorefiles();
+ public Collection<StoreFile> getStorefiles();
/**
* Close all the readers We don't need to worry about subsequent requests because the HRegion
@@ -64,7 +65,7 @@ public interface Store extends HeapSize,
* @return the {@link StoreFile StoreFiles} that were previously being used.
* @throws IOException on failure
*/
- Collection<StoreFile> close() throws IOException;
+ public Collection<StoreFile> close() throws IOException;
/**
* Return a scanner for both the memstore and the HStore files. Assumes we are not in a
@@ -74,7 +75,7 @@ public interface Store extends HeapSize,
* @return a scanner over the current key values
* @throws IOException on failure
*/
- KeyValueScanner getScanner(Scan scan, final NavigableSet<byte[]> targetCols)
+ public KeyValueScanner getScanner(Scan scan, final NavigableSet<byte[]> targetCols)
throws IOException;
/**
@@ -88,16 +89,11 @@ public interface Store extends HeapSize,
* @param stopRow
* @return all scanners for this store
*/
- List<KeyValueScanner> getScanners(
- boolean cacheBlocks,
- boolean isGet,
- boolean isCompaction,
- ScanQueryMatcher matcher,
- byte[] startRow,
- byte[] stopRow
- ) throws IOException;
+ public List<KeyValueScanner> getScanners(boolean cacheBlocks,
+ boolean isGet, boolean isCompaction, ScanQueryMatcher matcher, byte[] startRow,
+ byte[] stopRow) throws IOException;
- ScanInfo getScanInfo();
+ public ScanInfo getScanInfo();
/**
* Adds or replaces the specified KeyValues.
@@ -112,14 +108,14 @@ public interface Store extends HeapSize,
* @return memstore size delta
* @throws IOException
*/
- long upsert(Iterable<? extends Cell> cells, long readpoint) throws IOException;
+ public long upsert(Iterable<? extends Cell> cells, long readpoint) throws IOException;
/**
* Adds a value to the memstore
* @param kv
* @return memstore size delta
*/
- long add(KeyValue kv);
+ public long add(KeyValue kv);
/**
* When was the last edit done in the memstore
@@ -131,7 +127,7 @@ public interface Store extends HeapSize,
* key & memstoreTS value of the kv parameter.
* @param kv
*/
- void rollback(final KeyValue kv);
+ public void rollback(final KeyValue kv);
/**
* Find the key that matches <i>row</i> exactly, or the one that immediately precedes it. WARNING:
@@ -145,9 +141,9 @@ public interface Store extends HeapSize,
* @return Found keyvalue or null if none found.
* @throws IOException
*/
- KeyValue getRowKeyAtOrBefore(final byte[] row) throws IOException;
+ public KeyValue getRowKeyAtOrBefore(final byte[] row) throws IOException;
- FileSystem getFileSystem();
+ public FileSystem getFileSystem();
/*
* @param maxKeyCount
@@ -156,48 +152,44 @@ public interface Store extends HeapSize,
* @param includeMVCCReadpoint whether we should out the MVCC readpoint
* @return Writer for a new StoreFile in the tmp dir.
*/
- StoreFile.Writer createWriterInTmp(
- long maxKeyCount,
- Compression.Algorithm compression,
- boolean isCompaction,
- boolean includeMVCCReadpoint
- ) throws IOException;
+ public StoreFile.Writer createWriterInTmp(long maxKeyCount, Compression.Algorithm compression,
+ boolean isCompaction, boolean includeMVCCReadpoint) throws IOException;
// Compaction oriented methods
- boolean throttleCompaction(long compactionSize);
+ public boolean throttleCompaction(long compactionSize);
/**
* getter for CompactionProgress object
* @return CompactionProgress object; can be null
*/
- CompactionProgress getCompactionProgress();
+ public CompactionProgress getCompactionProgress();
- CompactionContext requestCompaction() throws IOException;
+ public CompactionContext requestCompaction() throws IOException;
- CompactionContext requestCompaction(int priority, CompactionRequest baseRequest)
+ public CompactionContext requestCompaction(int priority, CompactionRequest baseRequest)
throws IOException;
- void cancelRequestedCompaction(CompactionContext compaction);
+ public void cancelRequestedCompaction(CompactionContext compaction);
- List<StoreFile> compact(CompactionContext compaction) throws IOException;
+ public List<StoreFile> compact(CompactionContext compaction) throws IOException;
/**
* @return true if we should run a major compaction.
*/
- boolean isMajorCompaction() throws IOException;
+ public boolean isMajorCompaction() throws IOException;
- void triggerMajorCompaction();
+ public void triggerMajorCompaction();
/**
* See if there's too much store files in this store
* @return true if number of store files is greater than the number defined in minFilesToCompact
*/
- boolean needsCompaction();
+ public boolean needsCompaction();
- int getCompactPriority();
+ public int getCompactPriority();
- StoreFlushContext createFlushContext(long cacheFlushId);
+ public StoreFlushContext createFlushContext(long cacheFlushId);
/**
* Call to complete a compaction. Its for the case where we find in the WAL a compaction
@@ -205,18 +197,18 @@ public interface Store extends HeapSize,
* See HBASE-2331.
* @param compaction
*/
- void completeCompactionMarker(CompactionDescriptor compaction)
+ public void completeCompactionMarker(CompactionDescriptor compaction)
throws IOException;
// Split oriented methods
- boolean canSplit();
+ public boolean canSplit();
/**
* Determines if Store should be split
* @return byte[] if store should be split, null otherwise.
*/
- byte[] getSplitPoint();
+ public byte[] getSplitPoint();
// Bulk Load methods
@@ -224,7 +216,7 @@ public interface Store extends HeapSize,
* This throws a WrongRegionException if the HFile does not fit in this region, or an
* InvalidHFileException if the HFile is not valid.
*/
- void assertBulkLoadHFileOk(Path srcPath) throws IOException;
+ public void assertBulkLoadHFileOk(Path srcPath) throws IOException;
/**
* This method should only be called from HRegion. It is assumed that the ranges of values in the
@@ -233,7 +225,7 @@ public interface Store extends HeapSize,
* @param srcPathStr
* @param sequenceId sequence Id associated with the HFile
*/
- void bulkLoadHFile(String srcPathStr, long sequenceId) throws IOException;
+ public void bulkLoadHFile(String srcPathStr, long sequenceId) throws IOException;
// General accessors into the state of the store
// TODO abstract some of this out into a metrics class
@@ -241,50 +233,50 @@ public interface Store extends HeapSize,
/**
* @return <tt>true</tt> if the store has any underlying reference files to older HFiles
*/
- boolean hasReferences();
+ public boolean hasReferences();
/**
* @return The size of this store's memstore, in bytes
*/
- long getMemStoreSize();
+ public long getMemStoreSize();
- HColumnDescriptor getFamily();
+ public HColumnDescriptor getFamily();
/**
* @return The maximum memstoreTS in all store files.
*/
- long getMaxMemstoreTS();
+ public long getMaxMemstoreTS();
/**
* @return the data block encoder
*/
- HFileDataBlockEncoder getDataBlockEncoder();
+ public HFileDataBlockEncoder getDataBlockEncoder();
/** @return aggregate size of all HStores used in the last compaction */
- long getLastCompactSize();
+ public long getLastCompactSize();
/** @return aggregate size of HStore */
- long getSize();
+ public long getSize();
/**
* @return Count of store files
*/
- int getStorefilesCount();
+ public int getStorefilesCount();
/**
* @return The size of the store files, in bytes, uncompressed.
*/
- long getStoreSizeUncompressed();
+ public long getStoreSizeUncompressed();
/**
* @return The size of the store files, in bytes.
*/
- long getStorefilesSize();
+ public long getStorefilesSize();
/**
* @return The size of the store file indexes, in bytes.
*/
- long getStorefilesIndexSize();
+ public long getStorefilesIndexSize();
/**
* Returns the total size of all index blocks in the data block indexes, including the root level,
@@ -292,14 +284,14 @@ public interface Store extends HeapSize,
* single-level indexes.
* @return the total size of block indexes in the store
*/
- long getTotalStaticIndexSize();
+ public long getTotalStaticIndexSize();
/**
* Returns the total byte size of all Bloom filter bit arrays. For compound Bloom filters even the
* Bloom blocks currently not loaded into the block cache are counted.
* @return the total size of all Bloom filters in the store
*/
- long getTotalStaticBloomSize();
+ public long getTotalStaticBloomSize();
// Test-helper methods
@@ -307,40 +299,40 @@ public interface Store extends HeapSize,
* Used for tests.
* @return cache configuration for this Store.
*/
- CacheConfig getCacheConfig();
+ public CacheConfig getCacheConfig();
/**
* @return the parent region info hosting this store
*/
- HRegionInfo getRegionInfo();
+ public HRegionInfo getRegionInfo();
- RegionCoprocessorHost getCoprocessorHost();
+ public RegionCoprocessorHost getCoprocessorHost();
- boolean areWritesEnabled();
+ public boolean areWritesEnabled();
/**
* @return The smallest mvcc readPoint across all the scanners in this
* region. Writes older than this readPoint, are included in every
* read operation.
*/
- long getSmallestReadPoint();
+ public long getSmallestReadPoint();
- String getColumnFamilyName();
+ public String getColumnFamilyName();
- String getTableName();
+ public String getTableName();
/*
* @param o Observer who wants to know about changes in set of Readers
*/
- void addChangedReaderObserver(ChangedReadersObserver o);
+ public void addChangedReaderObserver(ChangedReadersObserver o);
/*
* @param o Observer no longer interested in changes in set of Readers.
*/
- void deleteChangedReaderObserver(ChangedReadersObserver o);
+ public void deleteChangedReaderObserver(ChangedReadersObserver o);
/**
* @return Whether this store has too many store files.
*/
- boolean hasTooManyStoreFiles();
+ public boolean hasTooManyStoreFiles();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreConfigInformation.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreConfigInformation.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreConfigInformation.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreConfigInformation.java Wed Jul 10 18:50:21 2013
@@ -34,22 +34,22 @@ public interface StoreConfigInformation
* TODO: remove after HBASE-7252 is fixed.
* @return Gets the Memstore flush size for the region that this store works with.
*/
- long getMemstoreFlushSize();
+ public long getMemstoreFlushSize();
/**
* @return Gets the cf-specific time-to-live for store files.
*/
- long getStoreFileTtl();
+ public long getStoreFileTtl();
/**
* @return Gets the cf-specific compaction check frequency multiplier.
* The need for compaction (outside of normal checks during flush, open, etc.) will
* be ascertained every multiplier * HConstants.THREAD_WAKE_FREQUENCY milliseconds.
*/
- long getCompactionCheckMultiplier();
+ public long getCompactionCheckMultiplier();
/**
* The number of files required before flushes for this store will be blocked.
*/
- long getBlockingFileCount();
+ public long getBlockingFileCount();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFileManager.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFileManager.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFileManager.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFileManager.java Wed Jul 10 18:50:21 2013
@@ -44,41 +44,40 @@ public interface StoreFileManager {
* Loads the initial store files into empty StoreFileManager.
* @param storeFiles The files to load.
*/
- void loadFiles(List<StoreFile> storeFiles);
+ public abstract void loadFiles(List<StoreFile> storeFiles);
/**
* Adds new file, either for from MemStore flush or bulk insert, into the structure.
* @param sf New store file.
*/
- void insertNewFile(StoreFile sf);
+ public abstract void insertNewFile(StoreFile sf);
/**
* Adds compaction results into the structure.
* @param compactedFiles The input files for the compaction.
* @param results The resulting files for the compaction.
*/
- void addCompactionResults(
- Collection<StoreFile> compactedFiles, Collection<StoreFile> results
- );
+ public abstract void addCompactionResults(
+ Collection<StoreFile> compactedFiles, Collection<StoreFile> results);
/**
* Clears all the files currently in use and returns them.
* @return The files previously in use.
*/
- ImmutableCollection<StoreFile> clearFiles();
+ public abstract ImmutableCollection<StoreFile> clearFiles();
/**
* Gets the snapshot of the store files currently in use. Can be used for things like metrics
* and checks; should not assume anything about relations between store files in the list.
* @return The list of StoreFiles.
*/
- Collection<StoreFile> getStorefiles();
+ public abstract Collection<StoreFile> getStorefiles();
/**
* Returns the number of files currently in use.
* @return The number of files.
*/
- int getStorefileCount();
+ public abstract int getStorefileCount();
/**
* Gets the store files to scan for a Scan or Get request.
@@ -87,9 +86,8 @@ public interface StoreFileManager {
* @param stopRow Stop row of the request.
* @return The list of files that are to be read for this request.
*/
- Collection<StoreFile> getFilesForScanOrGet(
- boolean isGet, byte[] startRow, byte[] stopRow
- );
+ public abstract Collection<StoreFile> getFilesForScanOrGet(boolean isGet,
+ byte[] startRow, byte[] stopRow);
/**
* Gets initial, full list of candidate store files to check for row-key-before.
@@ -97,9 +95,8 @@ public interface StoreFileManager {
* @return The files that may have the key less than or equal to targetKey, in reverse
* order of new-ness, and preference for target key.
*/
- Iterator<StoreFile> getCandidateFilesForRowKeyBefore(
- KeyValue targetKey
- );
+ public abstract Iterator<StoreFile> getCandidateFilesForRowKeyBefore(
+ KeyValue targetKey);
/**
* Updates the candidate list for finding row key before. Based on the list of candidates
@@ -112,9 +109,8 @@ public interface StoreFileManager {
* @param candidate The current best candidate found.
* @return The list to replace candidateFiles.
*/
- Iterator<StoreFile> updateCandidateFilesForRowKeyBefore(
- Iterator<StoreFile> candidateFiles, KeyValue targetKey, KeyValue candidate
- );
+ public abstract Iterator<StoreFile> updateCandidateFilesForRowKeyBefore(
+ Iterator<StoreFile> candidateFiles, KeyValue targetKey, KeyValue candidate);
/**
@@ -122,10 +118,10 @@ public interface StoreFileManager {
* @return The mid-point, or null if no split is possible.
* @throws IOException
*/
- byte[] getSplitPoint() throws IOException;
+ public abstract byte[] getSplitPoint() throws IOException;
/**
* @return The store compaction priority.
*/
- int getStoreCompactionPriority();
+ public abstract int getStoreCompactionPriority();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/Dictionary.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/Dictionary.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/Dictionary.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/Dictionary.java Wed Jul 10 18:50:21 2013
@@ -28,7 +28,7 @@ import org.apache.hadoop.classification.
*/
@InterfaceAudience.Private
interface Dictionary {
- byte NOT_IN_DICTIONARY = -1;
+ static final byte NOT_IN_DICTIONARY = -1;
/**
* Gets an entry from the dictionary.
@@ -36,7 +36,7 @@ interface Dictionary {
* @param idx index of the entry
* @return the entry, or null if non existent
*/
- byte[] getEntry(short idx);
+ public byte[] getEntry(short idx);
/**
* Finds the index of an entry.
@@ -47,7 +47,7 @@ interface Dictionary {
* @param length Length beyond <code>offset</code> that comprises entry; must be > 0.
* @return the index of the entry, or {@link #NOT_IN_DICTIONARY} if not found
*/
- short findEntry(byte[] data, int offset, int length);
+ public short findEntry(byte[] data, int offset, int length);
/**
* Adds an entry to the dictionary.
@@ -62,10 +62,10 @@ interface Dictionary {
* @return the index of the entry
*/
- short addEntry(byte[] data, int offset, int length);
+ public short addEntry(byte[] data, int offset, int length);
/**
* Flushes the dictionary, empties all values.
*/
- void clear();
+ public void clear();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java Wed Jul 10 18:50:21 2013
@@ -41,26 +41,26 @@ import org.apache.hadoop.io.Writable;
@InterfaceAudience.Private
public interface HLog {
- Log LOG = LogFactory.getLog(HLog.class);
+ public static final Log LOG = LogFactory.getLog(HLog.class);
/** File Extension used while splitting an HLog into regions (HBASE-2312) */
- String SPLITTING_EXT = "-splitting";
- boolean SPLIT_SKIP_ERRORS_DEFAULT = false;
+ public static final String SPLITTING_EXT = "-splitting";
+ public static final boolean SPLIT_SKIP_ERRORS_DEFAULT = false;
/** The META region's HLog filename extension */
- String META_HLOG_FILE_EXTN = ".meta";
+ public static final String META_HLOG_FILE_EXTN = ".meta";
/**
* Configuration name of HLog Trailer's warning size. If a waltrailer's size is greater than the
* configured size, a warning is logged. This is used with Protobuf reader/writer.
*/
- String WAL_TRAILER_WARN_SIZE =
+ public static final String WAL_TRAILER_WARN_SIZE =
"hbase.regionserver.waltrailer.warn.size";
- int DEFAULT_WAL_TRAILER_WARN_SIZE = 1024*1024; // 1MB
+ public static final int DEFAULT_WAL_TRAILER_WARN_SIZE = 1024*1024; // 1MB
- Pattern EDITFILES_NAME_PATTERN = Pattern.compile("-?[0-9]+");
- String RECOVERED_LOG_TMPFILE_SUFFIX = ".temp";
+ static final Pattern EDITFILES_NAME_PATTERN = Pattern.compile("-?[0-9]+");
+ public static final String RECOVERED_LOG_TMPFILE_SUFFIX = ".temp";
- interface Reader {
+ public interface Reader {
/**
* @param fs File system.
@@ -88,7 +88,7 @@ public interface HLog {
WALTrailer getWALTrailer();
}
- interface Writer {
+ public interface Writer {
void init(FileSystem fs, Path path, Configuration c) throws IOException;
void close() throws IOException;
@@ -110,7 +110,7 @@ public interface HLog {
* Utility class that lets us keep track of the edit with it's key Only used
* when splitting logs
*/
- class Entry implements Writable {
+ public static class Entry implements Writable {
private WALEdit edit;
private HLogKey key;
@@ -185,19 +185,19 @@ public interface HLog {
*
* @param listener
*/
- void registerWALActionsListener(final WALActionsListener listener);
+ public void registerWALActionsListener(final WALActionsListener listener);
/**
* unregisters WALActionsListener
*
* @param listener
*/
- boolean unregisterWALActionsListener(final WALActionsListener listener);
+ public boolean unregisterWALActionsListener(final WALActionsListener listener);
/**
* @return Current state of the monotonically increasing file id.
*/
- long getFilenum();
+ public long getFilenum();
/**
* Called by HRegionServer when it opens a new region to ensure that log
@@ -208,12 +208,12 @@ public interface HLog {
* We'll set log edit/sequence number to this value if it is greater
* than the current value.
*/
- void setSequenceNumber(final long newvalue);
+ public void setSequenceNumber(final long newvalue);
/**
* @return log sequence number
*/
- long getSequenceNumber();
+ public long getSequenceNumber();
/**
* Roll the log writer. That is, start writing log messages to a new file.
@@ -228,7 +228,7 @@ public interface HLog {
* @throws org.apache.hadoop.hbase.exceptions.FailedLogCloseException
* @throws IOException
*/
- byte[][] rollWriter() throws FailedLogCloseException, IOException;
+ public byte[][] rollWriter() throws FailedLogCloseException, IOException;
/**
* Roll the log writer. That is, start writing log messages to a new file.
@@ -246,7 +246,7 @@ public interface HLog {
* @throws org.apache.hadoop.hbase.exceptions.FailedLogCloseException
* @throws IOException
*/
- byte[][] rollWriter(boolean force) throws FailedLogCloseException,
+ public byte[][] rollWriter(boolean force) throws FailedLogCloseException,
IOException;
/**
@@ -254,22 +254,21 @@ public interface HLog {
*
* @throws IOException
*/
- void close() throws IOException;
+ public void close() throws IOException;
/**
* Shut down the log and delete the log directory
*
* @throws IOException
*/
- void closeAndDelete() throws IOException;
+ public void closeAndDelete() throws IOException;
/**
* Same as {@link #appendNoSync(HRegionInfo, byte[], WALEdit, UUID, long, HTableDescriptor)},
* except it causes a sync on the log
*/
- void append(
- HRegionInfo info, byte[] tableName, WALEdit edits, final long now, HTableDescriptor htd
- ) throws IOException;
+ public void append(HRegionInfo info, byte[] tableName, WALEdit edits,
+ final long now, HTableDescriptor htd) throws IOException;
/**
* Append a set of edits to the log. Log edits are keyed by (encoded)
@@ -282,14 +281,8 @@ public interface HLog {
* @param htd
* @param isInMemstore Whether the record is in memstore. False for system records.
*/
- void append(
- HRegionInfo info,
- byte[] tableName,
- WALEdit edits,
- final long now,
- HTableDescriptor htd,
- boolean isInMemstore
- ) throws IOException;
+ public void append(HRegionInfo info, byte[] tableName, WALEdit edits,
+ final long now, HTableDescriptor htd, boolean isInMemstore) throws IOException;
/**
* Append a set of edits to the log. Log edits are keyed by (encoded)
@@ -306,27 +299,21 @@ public interface HLog {
* @return txid of this transaction
* @throws IOException
*/
- long appendNoSync(
- HRegionInfo info,
- byte[] tableName,
- WALEdit edits,
- UUID clusterId,
- final long now,
- HTableDescriptor htd
- ) throws IOException;
+ public long appendNoSync(HRegionInfo info, byte[] tableName, WALEdit edits,
+ UUID clusterId, final long now, HTableDescriptor htd) throws IOException;
- void hsync() throws IOException;
+ public void hsync() throws IOException;
- void hflush() throws IOException;
+ public void hflush() throws IOException;
- void sync() throws IOException;
+ public void sync() throws IOException;
- void sync(long txid) throws IOException;
+ public void sync(long txid) throws IOException;
/**
* Obtain a log sequence number.
*/
- long obtainSeqNum();
+ public long obtainSeqNum();
/**
* WAL keeps track of the sequence numbers that were not yet flushed from memstores
@@ -343,13 +330,13 @@ public interface HLog {
* the resulting file as an upper-bound seqNum for that file), or NULL if flush
* should not be started.
*/
- Long startCacheFlush(final byte[] encodedRegionName);
+ public Long startCacheFlush(final byte[] encodedRegionName);
/**
* Complete the cache flush.
* @param encodedRegionName Encoded region name.
*/
- void completeCacheFlush(final byte[] encodedRegionName);
+ public void completeCacheFlush(final byte[] encodedRegionName);
/**
* Abort a cache flush. Call if the flush fails. Note that the only recovery
@@ -357,24 +344,24 @@ public interface HLog {
* snapshot content dropped by the failure gets restored to the memstore.v
* @param encodedRegionName Encoded region name.
*/
- void abortCacheFlush(byte[] encodedRegionName);
+ public void abortCacheFlush(byte[] encodedRegionName);
/**
* @return Coprocessor host.
*/
- WALCoprocessorHost getCoprocessorHost();
+ public WALCoprocessorHost getCoprocessorHost();
/**
* Get LowReplication-Roller status
*
* @return lowReplicationRollEnabled
*/
- boolean isLowReplicationRollEnabled();
+ public boolean isLowReplicationRollEnabled();
/** Gets the earliest sequence number in the memstore for this particular region.
* This can serve as best-effort "recent" WAL number for this region.
* @param encodedRegionName The region to get the number for.
* @return The number if present, HConstants.NO_SEQNUM if absent.
*/
- long getEarliestMemstoreSeqNum(byte[] encodedRegionName);
+ public long getEarliestMemstoreSeqNum(byte[] encodedRegionName);
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALActionsListener.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALActionsListener.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALActionsListener.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALActionsListener.java Wed Jul 10 18:50:21 2013
@@ -38,7 +38,7 @@ public interface WALActionsListener {
* @param oldPath the path to the old hlog
* @param newPath the path to the new hlog
*/
- void preLogRoll(Path oldPath, Path newPath) throws IOException;
+ public void preLogRoll(Path oldPath, Path newPath) throws IOException;
/**
* The WAL has been rolled. The oldPath can be null if this is
@@ -46,31 +46,31 @@ public interface WALActionsListener {
* @param oldPath the path to the old hlog
* @param newPath the path to the new hlog
*/
- void postLogRoll(Path oldPath, Path newPath) throws IOException;
+ public void postLogRoll(Path oldPath, Path newPath) throws IOException;
/**
* The WAL is going to be archived.
* @param oldPath the path to the old hlog
* @param newPath the path to the new hlog
*/
- void preLogArchive(Path oldPath, Path newPath) throws IOException;
+ public void preLogArchive(Path oldPath, Path newPath) throws IOException;
/**
* The WAL has been archived.
* @param oldPath the path to the old hlog
* @param newPath the path to the new hlog
*/
- void postLogArchive(Path oldPath, Path newPath) throws IOException;
+ public void postLogArchive(Path oldPath, Path newPath) throws IOException;
/**
* A request was made that the WAL be rolled.
*/
- void logRollRequested();
+ public void logRollRequested();
/**
* The WAL is about to close.
*/
- void logCloseRequested();
+ public void logCloseRequested();
/**
* Called before each write.
@@ -78,9 +78,8 @@ public interface WALActionsListener {
* @param logKey
* @param logEdit
*/
- void visitLogEntryBeforeWrite(
- HRegionInfo info, HLogKey logKey, WALEdit logEdit
- );
+ public void visitLogEntryBeforeWrite(HRegionInfo info, HLogKey logKey,
+ WALEdit logEdit);
/**
*
@@ -88,8 +87,7 @@ public interface WALActionsListener {
* @param logKey
* @param logEdit
*/
- void visitLogEntryBeforeWrite(
- HTableDescriptor htd, HLogKey logKey, WALEdit logEdit
- );
+ public void visitLogEntryBeforeWrite(HTableDescriptor htd, HLogKey logKey,
+ WALEdit logEdit);
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceInterface.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceInterface.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceInterface.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceInterface.java Wed Jul 10 18:50:21 2013
@@ -41,63 +41,61 @@ public interface ReplicationSourceInterf
* @param peerClusterId the id of the peer cluster
* @throws IOException
*/
- void init(
- final Configuration conf,
- final FileSystem fs,
- final ReplicationSourceManager manager,
- final Stoppable stopper,
- final String peerClusterId
- ) throws IOException;
+ public void init(final Configuration conf,
+ final FileSystem fs,
+ final ReplicationSourceManager manager,
+ final Stoppable stopper,
+ final String peerClusterId) throws IOException;
/**
* Add a log to the list of logs to replicate
* @param log path to the log to replicate
*/
- void enqueueLog(Path log);
+ public void enqueueLog(Path log);
/**
* Get the current log that's replicated
* @return the current log
*/
- Path getCurrentPath();
+ public Path getCurrentPath();
/**
* Start the replication
*/
- void startup();
+ public void startup();
/**
* End the replication
* @param reason why it's terminating
*/
- void terminate(String reason);
+ public void terminate(String reason);
/**
* End the replication
* @param reason why it's terminating
* @param cause the error that's causing it
*/
- void terminate(String reason, Exception cause);
+ public void terminate(String reason, Exception cause);
/**
* Get the id that the source is replicating to
*
* @return peer cluster id
*/
- String getPeerClusterZnode();
+ public String getPeerClusterZnode();
/**
* Get the id that the source is replicating to.
*
* @return peer cluster id
*/
- String getPeerClusterId();
+ public String getPeerClusterId();
/**
* Get a string representation of the current statistics
* for this source
* @return printable stats
*/
- String getStats();
+ public String getStats();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/Constants.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/Constants.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/Constants.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/Constants.java Wed Jul 10 18:50:21 2013
@@ -28,19 +28,19 @@ import org.apache.hadoop.classification.
@InterfaceAudience.Public
@InterfaceStability.Stable
public interface Constants {
- String VERSION_STRING = "0.0.2";
+ public static final String VERSION_STRING = "0.0.2";
- int DEFAULT_MAX_AGE = 60 * 60 * 4; // 4 hours
+ public static final int DEFAULT_MAX_AGE = 60 * 60 * 4; // 4 hours
- int DEFAULT_LISTEN_PORT = 8080;
+ public static final int DEFAULT_LISTEN_PORT = 8080;
- String MIMETYPE_TEXT = "text/plain";
- String MIMETYPE_HTML = "text/html";
- String MIMETYPE_XML = "text/xml";
- String MIMETYPE_BINARY = "application/octet-stream";
- String MIMETYPE_PROTOBUF = "application/x-protobuf";
- String MIMETYPE_PROTOBUF_IETF = "application/protobuf";
- String MIMETYPE_JSON = "application/json";
+ public static final String MIMETYPE_TEXT = "text/plain";
+ public static final String MIMETYPE_HTML = "text/html";
+ public static final String MIMETYPE_XML = "text/xml";
+ public static final String MIMETYPE_BINARY = "application/octet-stream";
+ public static final String MIMETYPE_PROTOBUF = "application/x-protobuf";
+ public static final String MIMETYPE_PROTOBUF_IETF = "application/protobuf";
+ public static final String MIMETYPE_JSON = "application/json";
- String CRLF = "\r\n";
+ public static final String CRLF = "\r\n";
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/ProtobufMessageHandler.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/ProtobufMessageHandler.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/ProtobufMessageHandler.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/ProtobufMessageHandler.java Wed Jul 10 18:50:21 2013
@@ -29,11 +29,11 @@ import org.apache.hadoop.classification.
* ProtobufMessageBodyProducer adapters.
*/
@InterfaceAudience.Private
-public interface ProtobufMessageHandler {
+public abstract interface ProtobufMessageHandler {
/**
* @return the protobuf represention of the model
*/
- byte[] createProtobufOutput();
+ public byte[] createProtobufOutput();
/**
* Initialize the model from a protobuf representation.
@@ -41,6 +41,6 @@ public interface ProtobufMessageHandler
* @return reference to self for convenience
* @throws IOException
*/
- ProtobufMessageHandler getObjectFromMessage(byte[] message)
+ public ProtobufMessageHandler getObjectFromMessage(byte[] message)
throws IOException;
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/CodeToClassAndBackFor96Migration.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/CodeToClassAndBackFor96Migration.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/CodeToClassAndBackFor96Migration.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/CodeToClassAndBackFor96Migration.java Wed Jul 10 18:50:21 2013
@@ -39,13 +39,13 @@ interface CodeToClassAndBackFor96Migrati
/**
* Static map that contains mapping from code to class
*/
- Map<Byte, Class<?>> CODE_TO_CLASS =
+ static final Map<Byte, Class<?>> CODE_TO_CLASS =
new HashMap<Byte, Class<?>>();
/**
* Static map that contains mapping from class to code
*/
- Map<Class<?>, Byte> CLASS_TO_CODE =
+ static final Map<Class<?>, Byte> CLASS_TO_CODE =
new HashMap<Class<?>, Byte>();
/**
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/thrift/IncrementCoalescerMBean.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/thrift/IncrementCoalescerMBean.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/thrift/IncrementCoalescerMBean.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/thrift/IncrementCoalescerMBean.java Wed Jul 10 18:50:21 2013
@@ -19,31 +19,31 @@
package org.apache.hadoop.hbase.thrift;
public interface IncrementCoalescerMBean {
- int getQueueSize();
+ public int getQueueSize();
- int getMaxQueueSize();
+ public int getMaxQueueSize();
- void setMaxQueueSize(int newSize);
+ public void setMaxQueueSize(int newSize);
- long getPoolCompletedTaskCount();
+ public long getPoolCompletedTaskCount();
- long getPoolTaskCount();
+ public long getPoolTaskCount();
- int getPoolLargestPoolSize();
+ public int getPoolLargestPoolSize();
- int getCorePoolSize();
+ public int getCorePoolSize();
- void setCorePoolSize(int newCoreSize);
+ public void setCorePoolSize(int newCoreSize);
- int getMaxPoolSize();
+ public int getMaxPoolSize();
- void setMaxPoolSize(int newMaxSize);
+ public void setMaxPoolSize(int newMaxSize);
- long getFailedIncrements();
+ public long getFailedIncrements();
- long getSuccessfulCoalescings();
+ public long getSuccessfulCoalescings();
- long getTotalIncrements();
+ public long getTotalIncrements();
- long getCountersMapSize();
+ public long getCountersMapSize();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java Wed Jul 10 18:50:21 2013
@@ -47,9 +47,9 @@ import org.apache.hadoop.hbase.client.HB
public final class Canary implements Tool {
// Sink interface used by the canary to outputs information
public interface Sink {
- void publishReadFailure(HRegionInfo region);
- void publishReadFailure(HRegionInfo region, HColumnDescriptor column);
- void publishReadTiming(HRegionInfo region, HColumnDescriptor column, long msTime);
+ public void publishReadFailure(HRegionInfo region);
+ public void publishReadFailure(HRegionInfo region, HColumnDescriptor column);
+ public void publishReadTiming(HRegionInfo region, HColumnDescriptor column, long msTime);
}
// Simple implementation of canary sink that allows to plot on
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/CancelableProgressable.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/CancelableProgressable.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/CancelableProgressable.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/CancelableProgressable.java Wed Jul 10 18:50:21 2013
@@ -34,6 +34,6 @@ public interface CancelableProgressable
* operation should be canceled and rolled back.
* @return whether to continue (true) or cancel (false) the operation
*/
- boolean progress();
+ public boolean progress();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java Wed Jul 10 18:50:21 2013
@@ -2903,7 +2903,7 @@ public class HBaseFsck extends Configure
}
public interface ErrorReporter {
- enum ERROR_CODE {
+ public static enum ERROR_CODE {
UNKNOWN, NO_META_REGION, NULL_META_REGION, NO_VERSION_FILE, NOT_IN_META_HDFS, NOT_IN_META,
NOT_IN_META_OR_DEPLOYED, NOT_IN_HDFS_OR_DEPLOYED, NOT_IN_HDFS, SERVER_DOES_NOT_MATCH_META, NOT_DEPLOYED,
MULTI_DEPLOYED, SHOULD_NOT_BE_DEPLOYED, MULTI_META_REGION, RS_CONNECT_FAILURE,
@@ -2912,26 +2912,20 @@ public class HBaseFsck extends Configure
ORPHAN_HDFS_REGION, LINGERING_SPLIT_PARENT, NO_TABLEINFO_FILE, LINGERING_REFERENCE_HFILE,
WRONG_USAGE, EMPTY_META_CELL, EXPIRED_TABLE_LOCK
}
- void clear();
- void report(String message);
- void reportError(String message);
- void reportError(ERROR_CODE errorCode, String message);
- void reportError(ERROR_CODE errorCode, String message, TableInfo table);
- void reportError(ERROR_CODE errorCode, String message, TableInfo table, HbckInfo info);
- void reportError(
- ERROR_CODE errorCode,
- String message,
- TableInfo table,
- HbckInfo info1,
- HbckInfo info2
- );
- int summarize();
- void detail(String details);
- ArrayList<ERROR_CODE> getErrorList();
- void progress();
- void print(String message);
- void resetErrors();
- boolean tableHasErrors(TableInfo table);
+ public void clear();
+ public void report(String message);
+ public void reportError(String message);
+ public void reportError(ERROR_CODE errorCode, String message);
+ public void reportError(ERROR_CODE errorCode, String message, TableInfo table);
+ public void reportError(ERROR_CODE errorCode, String message, TableInfo table, HbckInfo info);
+ public void reportError(ERROR_CODE errorCode, String message, TableInfo table, HbckInfo info1, HbckInfo info2);
+ public int summarize();
+ public void detail(String details);
+ public ArrayList<ERROR_CODE> getErrorList();
+ public void progress();
+ public void print(String message);
+ public void resetErrors();
+ public boolean tableHasErrors(TableInfo table);
}
static class PrintingErrorReporter implements ErrorReporter {
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/KeyRange.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/KeyRange.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/KeyRange.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/KeyRange.java Wed Jul 10 18:50:21 2013
@@ -25,7 +25,7 @@ import org.apache.hadoop.classification.
*/
@InterfaceAudience.Private
public interface KeyRange {
- byte[] getStartKey();
+ abstract byte[] getStartKey();
- byte[] getEndKey();
+ abstract byte[] getEndKey();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ModifyRegionUtils.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ModifyRegionUtils.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ModifyRegionUtils.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ModifyRegionUtils.java Wed Jul 10 18:50:21 2013
@@ -60,7 +60,7 @@ public abstract class ModifyRegionUtils
}
public interface RegionFillTask {
- void fillRegion(final HRegion region) throws IOException;
+ public void fillRegion(final HRegion region) throws IOException;
}
/**
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/RegionSplitter.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/RegionSplitter.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/RegionSplitter.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/RegionSplitter.java Wed Jul 10 18:50:21 2013
@@ -147,7 +147,7 @@ public class RegionSplitter {
* {@link RegionSplitter#rollingSplit(String, SplitAlgorithm, Configuration)} with the
* argument splitClassName giving the name of your class.
*/
- public interface SplitAlgorithm {
+ public static interface SplitAlgorithm {
/**
* Split a pre-existing region into 2 regions.
*
@@ -1046,4 +1046,4 @@ public class RegionSplitter {
+ "," + rowToStr(lastRow()) + "]";
}
}
-}
+}
\ No newline at end of file
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestCase.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestCase.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestCase.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestCase.java Wed Jul 10 18:50:21 2013
@@ -382,11 +382,11 @@ public abstract class HBaseTestCase exte
/**
* Implementors can flushcache.
*/
- public interface FlushCache {
+ public static interface FlushCache {
/**
* @throws IOException
*/
- void flushcache() throws IOException;
+ public void flushcache() throws IOException;
}
/**
@@ -395,23 +395,23 @@ public abstract class HBaseTestCase exte
*
* TOOD: Come up w/ a better name for this interface.
*/
- public interface Incommon {
+ public static interface Incommon {
/**
*
* @param delete
* @param writeToWAL
* @throws IOException
*/
- void delete(Delete delete, boolean writeToWAL)
+ public void delete(Delete delete, boolean writeToWAL)
throws IOException;
/**
* @param put
* @throws IOException
*/
- void put(Put put) throws IOException;
+ public void put(Put put) throws IOException;
- Result get(Get get) throws IOException;
+ public Result get(Get get) throws IOException;
/**
* @param family
@@ -421,9 +421,8 @@ public abstract class HBaseTestCase exte
* @return scanner for specified columns, first row and timestamp
* @throws IOException
*/
- ScannerIncommon getScanner(
- byte[] family, byte[][] qualifiers, byte[] firstRow, long ts
- )
+ public ScannerIncommon getScanner(byte [] family, byte [][] qualifiers,
+ byte [] firstRow, long ts)
throws IOException;
}
@@ -521,10 +520,10 @@ public abstract class HBaseTestCase exte
public interface ScannerIncommon
extends Iterable<Result> {
- boolean next(List<KeyValue> values)
+ public boolean next(List<KeyValue> values)
throws IOException;
- void close() throws IOException;
+ public void close() throws IOException;
}
public static class ClientScannerIncommon implements ScannerIncommon {
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java Wed Jul 10 18:50:21 2013
@@ -198,7 +198,7 @@ public class PerformanceEvaluation exten
/**
* Implementations can have their status set.
*/
- interface Status {
+ static interface Status {
/**
* Sets status
* @param msg status message