You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2013/07/10 19:53:35 UTC
svn commit: r1501880 [3/4] - in /hbase/trunk:
hbase-client/src/main/java/org/apache/hadoop/hbase/
hbase-client/src/main/java/org/apache/hadoop/hbase/catalog/
hbase-client/src/main/java/org/apache/hadoop/hbase/client/
hbase-client/src/main/java/org/apac...
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/handler/TotesHRegionInfo.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/handler/TotesHRegionInfo.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/handler/TotesHRegionInfo.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/handler/TotesHRegionInfo.java Wed Jul 10 17:53:33 2013
@@ -33,5 +33,5 @@ public interface TotesHRegionInfo {
/**
* @return HRegionInfo instance.
*/
- public HRegionInfo getHRegionInfo();
+ HRegionInfo getHRegionInfo();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandler.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandler.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandler.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandler.java Wed Jul 10 17:53:33 2013
@@ -29,16 +29,16 @@ import com.google.protobuf.Message;
*/
@InterfaceAudience.Private
public interface MonitoredRPCHandler extends MonitoredTask {
- public abstract String getRPC();
- public abstract String getRPC(boolean withParams);
- public abstract long getRPCPacketLength();
- public abstract String getClient();
- public abstract long getRPCStartTime();
- public abstract long getRPCQueueTime();
- public abstract boolean isRPCRunning();
- public abstract boolean isOperationRunning();
+ String getRPC();
+ String getRPC(boolean withParams);
+ long getRPCPacketLength();
+ String getClient();
+ long getRPCStartTime();
+ long getRPCQueueTime();
+ boolean isRPCRunning();
+ boolean isOperationRunning();
- public abstract void setRPC(String methodName, Object [] params, long queueTime);
- public abstract void setRPCPacket(Message param);
- public abstract void setConnection(String clientAddress, int remotePort);
+ void setRPC(String methodName, Object[] params, long queueTime);
+ void setRPCPacket(Message param);
+ void setConnection(String clientAddress, int remotePort);
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java Wed Jul 10 17:53:33 2013
@@ -32,47 +32,47 @@ public interface MonitoredTask extends C
ABORTED;
}
- public abstract long getStartTime();
- public abstract String getDescription();
- public abstract String getStatus();
- public abstract long getStatusTime();
- public abstract State getState();
- public abstract long getStateTime();
- public abstract long getCompletionTimestamp();
-
- public abstract void markComplete(String msg);
- public abstract void pause(String msg);
- public abstract void resume(String msg);
- public abstract void abort(String msg);
- public abstract void expireNow();
+ long getStartTime();
+ String getDescription();
+ String getStatus();
+ long getStatusTime();
+ State getState();
+ long getStateTime();
+ long getCompletionTimestamp();
+
+ void markComplete(String msg);
+ void pause(String msg);
+ void resume(String msg);
+ void abort(String msg);
+ void expireNow();
- public abstract void setStatus(String status);
- public abstract void setDescription(String description);
+ void setStatus(String status);
+ void setDescription(String description);
/**
* Explicitly mark this status as able to be cleaned up,
* even though it might not be complete.
*/
- public abstract void cleanup();
+ void cleanup();
/**
* Public exposure of Object.clone() in order to allow clients to easily
* capture current state.
* @return a copy of the object whose references will not change
*/
- public abstract MonitoredTask clone();
+ MonitoredTask clone();
/**
* Creates a string map of internal details for extensible exposure of
* monitored tasks.
* @return A Map containing information for this task.
*/
- public abstract Map<String, Object> toMap() throws IOException;
+ Map<String, Object> toMap() throws IOException;
/**
* Creates a JSON object for parseable exposure of monitored tasks.
* @return An encoded JSON object containing information for this task.
*/
- public abstract String toJSON() throws IOException;
+ String toJSON() throws IOException;
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinatorRpcs.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinatorRpcs.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinatorRpcs.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinatorRpcs.java Wed Jul 10 17:53:33 2013
@@ -39,7 +39,7 @@ public interface ProcedureCoordinatorRpc
* @param listener
* @return true if succeed, false if encountered initialization errors.
*/
- public boolean start(final ProcedureCoordinator listener);
+ boolean start(final ProcedureCoordinator listener);
/**
* Notify the members that the coordinator has aborted the procedure and that it should release
@@ -50,7 +50,7 @@ public interface ProcedureCoordinatorRpc
* @throws IOException if the rpcs can't reach the other members of the procedure (and can't
* recover).
*/
- public void sendAbortToMembers(Procedure procName, ForeignException cause) throws IOException;
+ void sendAbortToMembers(Procedure procName, ForeignException cause) throws IOException;
/**
* Notify the members to acquire barrier for the procedure
@@ -61,7 +61,7 @@ public interface ProcedureCoordinatorRpc
* @throws IllegalArgumentException if the procedure was already marked as failed
* @throws IOException if we can't reach the remote notification mechanism
*/
- public void sendGlobalBarrierAcquire(Procedure procName, byte[] info, List<String> members)
+ void sendGlobalBarrierAcquire(Procedure procName, byte[] info, List<String> members)
throws IOException, IllegalArgumentException;
/**
@@ -74,12 +74,12 @@ public interface ProcedureCoordinatorRpc
* @param members members to tell we have reached in-barrier phase
* @throws IOException if we can't reach the remote notification mechanism
*/
- public void sendGlobalBarrierReached(Procedure procName, List<String> members) throws IOException;
+ void sendGlobalBarrierReached(Procedure procName, List<String> members) throws IOException;
/**
* Notify Members to reset the distributed state for procedure
* @param procName name of the procedure to reset
* @throws IOException if the remote notification mechanism cannot be reached
*/
- public void resetMembers(Procedure procName) throws IOException;
+ void resetMembers(Procedure procName) throws IOException;
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureMemberRpcs.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureMemberRpcs.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureMemberRpcs.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureMemberRpcs.java Wed Jul 10 17:53:33 2013
@@ -35,13 +35,13 @@ public interface ProcedureMemberRpcs ext
/**
* Initialize and start any threads or connections the member needs.
*/
- public void start(final String memberName, final ProcedureMember member);
+ void start(final String memberName, final ProcedureMember member);
/**
* Each subprocedure is being executed on a member. This is the identifier for the member.
* @return the member name
*/
- public String getMemberName();
+ String getMemberName();
/**
* Notify the coordinator that we aborted the specified {@link Subprocedure}
@@ -51,7 +51,7 @@ public interface ProcedureMemberRpcs ext
* @throws IOException thrown when the rpcs can't reach the other members of the procedure (and
* thus can't recover).
*/
- public void sendMemberAborted(Subprocedure sub, ForeignException cause) throws IOException;
+ void sendMemberAborted(Subprocedure sub, ForeignException cause) throws IOException;
/**
* Notify the coordinator that the specified {@link Subprocedure} has acquired the locally required
@@ -60,7 +60,7 @@ public interface ProcedureMemberRpcs ext
* @param sub the specified {@link Subprocedure}
* @throws IOException if we can't reach the coordinator
*/
- public void sendMemberAcquired(Subprocedure sub) throws IOException;
+ void sendMemberAcquired(Subprocedure sub) throws IOException;
/**
* Notify the coordinator that the specified {@link Subprocedure} has completed the work that
@@ -69,5 +69,5 @@ public interface ProcedureMemberRpcs ext
* @param sub the specified {@link Subprocedure}
* @throws IOException if we can't reach the coordinator
*/
- public void sendMemberCompleted(Subprocedure sub) throws IOException;
-}
\ No newline at end of file
+ void sendMemberCompleted(Subprocedure sub) throws IOException;
+}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/SubprocedureFactory.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/SubprocedureFactory.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/SubprocedureFactory.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/SubprocedureFactory.java Wed Jul 10 17:53:33 2013
@@ -36,5 +36,5 @@ public interface SubprocedureFactory {
* request
* @throws IllegalStateException if the current runner cannot accept any more new requests
*/
- public Subprocedure buildSubprocedure(String procName, byte[] procArgs);
+ Subprocedure buildSubprocedure(String procName, byte[] procArgs);
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ColumnTracker.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ColumnTracker.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ColumnTracker.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ColumnTracker.java Wed Jul 10 17:53:33 2013
@@ -56,25 +56,26 @@ public interface ColumnTracker {
* @throws IOException in case there is an internal consistency problem
* caused by a data corruption.
*/
- public ScanQueryMatcher.MatchCode checkColumn(byte[] bytes, int offset,
- int length, long ttl, byte type, boolean ignoreCount)
+ ScanQueryMatcher.MatchCode checkColumn(
+ byte[] bytes, int offset, int length, long ttl, byte type, boolean ignoreCount
+ )
throws IOException;
/**
* Updates internal variables in between files
*/
- public void update();
+ void update();
/**
* Resets the Matcher
*/
- public void reset();
+ void reset();
/**
*
* @return <code>true</code> when done.
*/
- public boolean done();
+ boolean done();
/**
* Used by matcher and scan/get to get a hint of the next column
@@ -87,13 +88,14 @@ public interface ColumnTracker {
*
* @return null, or a ColumnCount that we should seek to
*/
- public ColumnCount getColumnHint();
+ ColumnCount getColumnHint();
/**
* Retrieve the MatchCode for the next row or column
*/
- public MatchCode getNextRowOrNextColumn(byte[] bytes, int offset,
- int qualLength);
+ MatchCode getNextRowOrNextColumn(
+ byte[] bytes, int offset, int qualLength
+ );
/**
* Give the tracker a chance to declare it's done based on only the timestamp
@@ -102,5 +104,5 @@ public interface ColumnTracker {
* @param timestamp
* @return <code>true</code> to early out based on timestamp.
*/
- public boolean isDone(long timestamp);
+ boolean isDone(long timestamp);
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactionRequestor.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactionRequestor.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactionRequestor.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactionRequestor.java Wed Jul 10 17:53:33 2013
@@ -34,7 +34,7 @@ public interface CompactionRequestor {
* compactions were started
* @throws IOException
*/
- public List<CompactionRequest> requestCompaction(final HRegion r, final String why)
+ List<CompactionRequest> requestCompaction(final HRegion r, final String why)
throws IOException;
/**
@@ -47,8 +47,9 @@ public interface CompactionRequestor {
* compactions were started
* @throws IOException
*/
- public List<CompactionRequest> requestCompaction(final HRegion r, final String why,
- List<Pair<CompactionRequest, Store>> requests)
+ List<CompactionRequest> requestCompaction(
+ final HRegion r, final String why, List<Pair<CompactionRequest, Store>> requests
+ )
throws IOException;
/**
@@ -60,8 +61,9 @@ public interface CompactionRequestor {
* @return The created {@link CompactionRequest} or <tt>null</tt> if no compaction was started.
* @throws IOException
*/
- public CompactionRequest requestCompaction(final HRegion r, final Store s, final String why,
- CompactionRequest request) throws IOException;
+ CompactionRequest requestCompaction(
+ final HRegion r, final Store s, final String why, CompactionRequest request
+ ) throws IOException;
/**
* @param r Region to compact
@@ -74,8 +76,9 @@ public interface CompactionRequestor {
* compactions were started.
* @throws IOException
*/
- public List<CompactionRequest> requestCompaction(final HRegion r, final String why, int pri,
- List<Pair<CompactionRequest, Store>> requests) throws IOException;
+ List<CompactionRequest> requestCompaction(
+ final HRegion r, final String why, int pri, List<Pair<CompactionRequest, Store>> requests
+ ) throws IOException;
/**
* @param r Region to compact
@@ -87,6 +90,7 @@ public interface CompactionRequestor {
* @return The created {@link CompactionRequest} or <tt>null</tt> if no compaction was started
* @throws IOException
*/
- public CompactionRequest requestCompaction(final HRegion r, final Store s, final String why,
- int pri, CompactionRequest request) throws IOException;
+ CompactionRequest requestCompaction(
+ final HRegion r, final Store s, final String why, int pri, CompactionRequest request
+ ) throws IOException;
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DeleteTracker.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DeleteTracker.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DeleteTracker.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DeleteTracker.java Wed Jul 10 17:53:33 2013
@@ -43,8 +43,9 @@ public interface DeleteTracker {
* @param timestamp timestamp
* @param type delete type as byte
*/
- public void add(byte [] buffer, int qualifierOffset, int qualifierLength,
- long timestamp, byte type);
+ void add(
+ byte[] buffer, int qualifierOffset, int qualifierLength, long timestamp, byte type
+ );
/**
* Check if the specified KeyValue buffer has been deleted by a previously
@@ -55,13 +56,14 @@ public interface DeleteTracker {
* @param timestamp timestamp
* @return deleteResult The result tells whether the KeyValue is deleted and why
*/
- public DeleteResult isDeleted(byte [] buffer, int qualifierOffset,
- int qualifierLength, long timestamp);
+ DeleteResult isDeleted(
+ byte[] buffer, int qualifierOffset, int qualifierLength, long timestamp
+ );
/**
* @return true if there are no current delete, false otherwise
*/
- public boolean isEmpty();
+ boolean isEmpty();
/**
* Called at the end of every StoreFile.
@@ -69,14 +71,14 @@ public interface DeleteTracker {
* Many optimized implementations of Trackers will require an update at
* when the end of each StoreFile is reached.
*/
- public void update();
+ void update();
/**
* Called between rows.
* <p>
* This clears everything as if a new DeleteTracker was instantiated.
*/
- public void reset();
+ void reset();
/**
@@ -102,7 +104,7 @@ public interface DeleteTracker {
* Based on the delete result, the ScanQueryMatcher will decide the next
* operation
*/
- public static enum DeleteResult {
+ enum DeleteResult {
FAMILY_DELETED, // The KeyValue is deleted by a delete family.
FAMILY_VERSION_DELETED, // The KeyValue is deleted by a delete family version.
COLUMN_DELETED, // The KeyValue is deleted by a delete column.
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java Wed Jul 10 17:53:33 2013
@@ -5602,7 +5602,7 @@ public class HRegion implements HeapSize
* bulkLoadHFile() to perform any necessary
* pre/post processing of a given bulkload call
*/
- public static interface BulkLoadListener {
+ public interface BulkLoadListener {
/**
* Called before an HFile is actually loaded
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/InternalScanner.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/InternalScanner.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/InternalScanner.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/InternalScanner.java Wed Jul 10 17:53:33 2013
@@ -47,7 +47,7 @@ public interface InternalScanner extends
* @return true if more rows exist after this one, false if scanner is done
* @throws IOException e
*/
- public boolean next(List<KeyValue> results) throws IOException;
+ boolean next(List<KeyValue> results) throws IOException;
/**
* Grab the next row's worth of values with a limit on the number of values
@@ -57,11 +57,11 @@ public interface InternalScanner extends
* @return true if more rows exist after this one, false if scanner is done
* @throws IOException e
*/
- public boolean next(List<KeyValue> result, int limit) throws IOException;
+ boolean next(List<KeyValue> result, int limit) throws IOException;
/**
* Closes the scanner and releases any resources it has allocated
* @throws IOException
*/
- public void close() throws IOException;
+ void close() throws IOException;
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/KeyValueScanner.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/KeyValueScanner.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/KeyValueScanner.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/KeyValueScanner.java Wed Jul 10 17:53:33 2013
@@ -34,20 +34,20 @@ public interface KeyValueScanner {
* Look at the next KeyValue in this scanner, but do not iterate scanner.
* @return the next KeyValue
*/
- public KeyValue peek();
+ KeyValue peek();
/**
* Return the next KeyValue in this scanner, iterating the scanner
* @return the next KeyValue
*/
- public KeyValue next() throws IOException;
+ KeyValue next() throws IOException;
/**
* Seek the scanner at or after the specified KeyValue.
* @param key seek value
* @return true if scanner has values left, false if end of scanner
*/
- public boolean seek(KeyValue key) throws IOException;
+ boolean seek(KeyValue key) throws IOException;
/**
* Reseek the scanner at or after the specified KeyValue.
@@ -57,7 +57,7 @@ public interface KeyValueScanner {
* @param key seek value (should be non-null)
* @return true if scanner has values left, false if end of scanner
*/
- public boolean reseek(KeyValue key) throws IOException;
+ boolean reseek(KeyValue key) throws IOException;
/**
* Get the sequence id associated with this KeyValueScanner. This is required
@@ -65,12 +65,12 @@ public interface KeyValueScanner {
* The default implementation for this would be to return 0. A file having
* lower sequence id will be considered to be the older one.
*/
- public long getSequenceID();
+ long getSequenceID();
/**
* Close the KeyValue scanner.
*/
- public void close();
+ void close();
/**
* Allows to filter out scanners (both StoreFile and memstore) that we don't
@@ -82,8 +82,9 @@ public interface KeyValueScanner {
* this query, based on TTL
* @return true if the scanner should be included in the query
*/
- public boolean shouldUseScanner(Scan scan, SortedSet<byte[]> columns,
- long oldestUnexpiredTS);
+ boolean shouldUseScanner(
+ Scan scan, SortedSet<byte[]> columns, long oldestUnexpiredTS
+ );
// "Lazy scanner" optimizations
@@ -97,7 +98,7 @@ public interface KeyValueScanner {
* @param forward do a forward-only "reseek" instead of a random-access seek
* @param useBloom whether to enable multi-column Bloom filter optimization
*/
- public boolean requestSeek(KeyValue kv, boolean forward, boolean useBloom)
+ boolean requestSeek(KeyValue kv, boolean forward, boolean useBloom)
throws IOException;
/**
@@ -106,7 +107,7 @@ public interface KeyValueScanner {
* store scanner bubbles up to the top of the key-value heap. This method is
* then used to ensure the top store file scanner has done a seek operation.
*/
- public boolean realSeekDone();
+ boolean realSeekDone();
/**
* Does the real seek operation in case it was skipped by
@@ -115,11 +116,11 @@ public interface KeyValueScanner {
* of the scanners). The easiest way to achieve this is to call
* {@link #realSeekDone()} first.
*/
- public void enforceSeek() throws IOException;
+ void enforceSeek() throws IOException;
/**
* @return true if this is a file scanner. Otherwise a memory scanner is
* assumed.
*/
- public boolean isFileScanner();
+ boolean isFileScanner();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/LastSequenceId.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/LastSequenceId.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/LastSequenceId.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/LastSequenceId.java Wed Jul 10 17:53:33 2013
@@ -29,5 +29,5 @@ public interface LastSequenceId {
* @param regionName Encoded region name
* @return Last flushed sequence Id for regionName or -1 if it can't be determined
*/
- public long getLastSequenceId(byte[] regionName);
+ long getLastSequenceId(byte[] regionName);
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/LeaseListener.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/LeaseListener.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/LeaseListener.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/LeaseListener.java Wed Jul 10 17:53:33 2013
@@ -32,5 +32,5 @@ import org.apache.hadoop.classification.
@InterfaceAudience.Private
public interface LeaseListener {
/** When a lease expires, this method is called. */
- public void leaseExpired();
+ void leaseExpired();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/OnlineRegions.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/OnlineRegions.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/OnlineRegions.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/OnlineRegions.java Wed Jul 10 17:53:33 2013
@@ -35,7 +35,7 @@ interface OnlineRegions extends Server {
* Add to online regions.
* @param r
*/
- public void addToOnlineRegions(final HRegion r);
+ void addToOnlineRegions(final HRegion r);
/**
* This method removes HRegion corresponding to hri from the Map of onlineRegions.
@@ -44,7 +44,7 @@ interface OnlineRegions extends Server {
* @param destination Destination, if any, null otherwise.
* @return True if we removed a region from online list.
*/
- public boolean removeFromOnlineRegions(final HRegion r, ServerName destination);
+ boolean removeFromOnlineRegions(final HRegion r, ServerName destination);
/**
* Return {@link HRegion} instance.
@@ -54,7 +54,7 @@ interface OnlineRegions extends Server {
* @return HRegion for the passed encoded <code>encodedRegionName</code> or
* null if named region is not member of the online regions.
*/
- public HRegion getFromOnlineRegions(String encodedRegionName);
+ HRegion getFromOnlineRegions(String encodedRegionName);
/**
* Get all online regions of a table in this RS.
@@ -62,5 +62,5 @@ interface OnlineRegions extends Server {
* @return List of HRegion
* @throws java.io.IOException
*/
- public List<HRegion> getOnlineRegions(byte[] tableName) throws IOException;
+ List<HRegion> getOnlineRegions(byte[] tableName) throws IOException;
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionScanner.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionScanner.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionScanner.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionScanner.java Wed Jul 10 17:53:33 2013
@@ -34,13 +34,13 @@ public interface RegionScanner extends I
/**
* @return The RegionInfo for this scanner.
*/
- public HRegionInfo getRegionInfo();
+ HRegionInfo getRegionInfo();
/**
* @return True if a filter indicates that this scanner will return no further rows.
* @throws IOException in case of I/O failure on a filter.
*/
- public boolean isFilterDone() throws IOException;
+ boolean isFilterDone() throws IOException;
/**
* Do a reseek to the required row. Should not be used to seek to a key which
@@ -52,17 +52,17 @@ public interface RegionScanner extends I
* if row is null
*
*/
- public boolean reseek(byte[] row) throws IOException;
+ boolean reseek(byte[] row) throws IOException;
/**
* @return The preferred max buffersize. See {@link Scan#setMaxResultSize(long)}
*/
- public long getMaxResultSize();
+ long getMaxResultSize();
/**
* @return The Scanner's MVCC readPt see {@link MultiVersionConsistencyControl}
*/
- public long getMvccReadPoint();
+ long getMvccReadPoint();
/**
* Grab the next row's worth of values with the default limit on the number of values
@@ -74,7 +74,7 @@ public interface RegionScanner extends I
* @return true if more rows exist after this one, false if scanner is done
* @throws IOException e
*/
- public boolean nextRaw(List<KeyValue> result) throws IOException;
+ boolean nextRaw(List<KeyValue> result) throws IOException;
/**
* Grab the next row's worth of values with a limit on the number of values
@@ -102,5 +102,5 @@ public interface RegionScanner extends I
* @return true if more rows exist after this one, false if scanner is done
* @throws IOException e
*/
- public boolean nextRaw(List<KeyValue> result, int limit) throws IOException;
+ boolean nextRaw(List<KeyValue> result, int limit) throws IOException;
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerServices.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerServices.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerServices.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerServices.java Wed Jul 10 17:53:33 2013
@@ -40,31 +40,31 @@ public interface RegionServerServices ex
/**
* @return True if this regionserver is stopping.
*/
- public boolean isStopping();
+ boolean isStopping();
/** @return the HLog for a particular region. Pass null for getting the
* default (common) WAL */
- public HLog getWAL(HRegionInfo regionInfo) throws IOException;
+ HLog getWAL(HRegionInfo regionInfo) throws IOException;
/**
* @return Implementation of {@link CompactionRequestor} or null.
*/
- public CompactionRequestor getCompactionRequester();
+ CompactionRequestor getCompactionRequester();
/**
* @return Implementation of {@link FlushRequester} or null.
*/
- public FlushRequester getFlushRequester();
+ FlushRequester getFlushRequester();
/**
* @return the RegionServerAccounting for this Region Server
*/
- public RegionServerAccounting getRegionServerAccounting();
+ RegionServerAccounting getRegionServerAccounting();
/**
* @return RegionServer's instance of {@link TableLockManager}
*/
- public TableLockManager getTableLockManager();
+ TableLockManager getTableLockManager();
/**
* Tasks to perform after region open to complete deploy of region on
@@ -75,42 +75,42 @@ public interface RegionServerServices ex
* @throws KeeperException
* @throws IOException
*/
- public void postOpenDeployTasks(final HRegion r, final CatalogTracker ct)
+ void postOpenDeployTasks(final HRegion r, final CatalogTracker ct)
throws KeeperException, IOException;
/**
* Returns a reference to the region server's RPC server
*/
- public RpcServerInterface getRpcServer();
+ RpcServerInterface getRpcServer();
/**
* Get the regions that are currently being opened or closed in the RS
* @return map of regions in transition in this RS
*/
- public ConcurrentMap<byte[], Boolean> getRegionsInTransitionInRS();
+ ConcurrentMap<byte[], Boolean> getRegionsInTransitionInRS();
/**
* @return Return the FileSystem object used by the regionserver
*/
- public FileSystem getFileSystem();
+ FileSystem getFileSystem();
/**
* @return The RegionServer's "Leases" service
*/
- public Leases getLeases();
+ Leases getLeases();
/**
* @return hbase executor service
*/
- public ExecutorService getExecutorService();
+ ExecutorService getExecutorService();
/**
* @return The RegionServer's CatalogTracker
*/
- public CatalogTracker getCatalogTracker();
+ CatalogTracker getCatalogTracker();
/**
* @return set of recovering regions on the hosting region server
*/
- public Map<String, HRegion> getRecoveringRegions();
+ Map<String, HRegion> getRecoveringRegions();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationService.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationService.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationService.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationService.java Wed Jul 10 17:53:33 2013
@@ -38,17 +38,18 @@ public interface ReplicationService {
* Initializes the replication service object.
* @throws IOException
*/
- public void initialize(Server rs, FileSystem fs, Path logdir,
- Path oldLogDir) throws IOException;
+ void initialize(
+ Server rs, FileSystem fs, Path logdir, Path oldLogDir
+ ) throws IOException;
/**
* Start replication services.
* @throws IOException
*/
- public void startReplicationService() throws IOException;
+ void startReplicationService() throws IOException;
/**
* Stops replication service.
*/
- public void stopReplicationService();
+ void stopReplicationService();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSinkService.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSinkService.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSinkService.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSinkService.java Wed Jul 10 17:53:33 2013
@@ -38,5 +38,5 @@ public interface ReplicationSinkService
* @param cells Cells that the WALEntries refer to (if cells is non-null)
* @throws IOException
*/
- public void replicateLogEntries(List<WALEntry> entries, CellScanner cells) throws IOException;
-}
\ No newline at end of file
+ void replicateLogEntries(List<WALEntry> entries, CellScanner cells) throws IOException;
+}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.java Wed Jul 10 17:53:33 2013
@@ -32,5 +32,5 @@ public interface ReplicationSourceServic
* Returns a WALObserver for the service. This is needed to
* observe log rolls and log archival events.
*/
- public WALActionsListener getWALActionsListener();
-}
\ No newline at end of file
+ WALActionsListener getWALActionsListener();
+}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java Wed Jul 10 17:53:33 2013
@@ -640,13 +640,13 @@ public class SplitLogWorker extends ZooK
* is better to have workers prepare the task and then have the
* {@link SplitLogManager} commit the work in SplitLogManager.TaskFinisher
*/
- static public interface TaskExecutor {
- static public enum Status {
+ public interface TaskExecutor {
+ enum Status {
DONE(),
ERR(),
RESIGNED(),
PREEMPTED()
}
- public Status exec(String name, CancelableProgressable p);
+ Status exec(String name, CancelableProgressable p);
}
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java Wed Jul 10 17:53:33 2013
@@ -50,14 +50,13 @@ public interface Store extends HeapSize,
/* The default priority for user-specified compaction requests.
* The user gets top priority unless we have blocking compactions. (Pri <= 0)
- */
- public static final int PRIORITY_USER = 1;
- public static final int NO_PRIORITY = Integer.MIN_VALUE;
+ */ int PRIORITY_USER = 1;
+ int NO_PRIORITY = Integer.MIN_VALUE;
// General Accessors
- public KeyValue.KVComparator getComparator();
+ KeyValue.KVComparator getComparator();
- public Collection<StoreFile> getStorefiles();
+ Collection<StoreFile> getStorefiles();
/**
* Close all the readers We don't need to worry about subsequent requests because the HRegion
@@ -65,7 +64,7 @@ public interface Store extends HeapSize,
* @return the {@link StoreFile StoreFiles} that were previously being used.
* @throws IOException on failure
*/
- public Collection<StoreFile> close() throws IOException;
+ Collection<StoreFile> close() throws IOException;
/**
* Return a scanner for both the memstore and the HStore files. Assumes we are not in a
@@ -75,7 +74,7 @@ public interface Store extends HeapSize,
* @return a scanner over the current key values
* @throws IOException on failure
*/
- public KeyValueScanner getScanner(Scan scan, final NavigableSet<byte[]> targetCols)
+ KeyValueScanner getScanner(Scan scan, final NavigableSet<byte[]> targetCols)
throws IOException;
/**
@@ -89,11 +88,16 @@ public interface Store extends HeapSize,
* @param stopRow
* @return all scanners for this store
*/
- public List<KeyValueScanner> getScanners(boolean cacheBlocks,
- boolean isGet, boolean isCompaction, ScanQueryMatcher matcher, byte[] startRow,
- byte[] stopRow) throws IOException;
+ List<KeyValueScanner> getScanners(
+ boolean cacheBlocks,
+ boolean isGet,
+ boolean isCompaction,
+ ScanQueryMatcher matcher,
+ byte[] startRow,
+ byte[] stopRow
+ ) throws IOException;
- public ScanInfo getScanInfo();
+ ScanInfo getScanInfo();
/**
* Adds or replaces the specified KeyValues.
@@ -108,14 +112,14 @@ public interface Store extends HeapSize,
* @return memstore size delta
* @throws IOException
*/
- public long upsert(Iterable<? extends Cell> cells, long readpoint) throws IOException;
+ long upsert(Iterable<? extends Cell> cells, long readpoint) throws IOException;
/**
* Adds a value to the memstore
* @param kv
* @return memstore size delta
*/
- public long add(KeyValue kv);
+ long add(KeyValue kv);
/**
* When was the last edit done in the memstore
@@ -127,7 +131,7 @@ public interface Store extends HeapSize,
* key & memstoreTS value of the kv parameter.
* @param kv
*/
- public void rollback(final KeyValue kv);
+ void rollback(final KeyValue kv);
/**
* Find the key that matches <i>row</i> exactly, or the one that immediately precedes it. WARNING:
@@ -141,9 +145,9 @@ public interface Store extends HeapSize,
* @return Found keyvalue or null if none found.
* @throws IOException
*/
- public KeyValue getRowKeyAtOrBefore(final byte[] row) throws IOException;
+ KeyValue getRowKeyAtOrBefore(final byte[] row) throws IOException;
- public FileSystem getFileSystem();
+ FileSystem getFileSystem();
/*
* @param maxKeyCount
@@ -152,44 +156,48 @@ public interface Store extends HeapSize,
* @param includeMVCCReadpoint whether we should out the MVCC readpoint
* @return Writer for a new StoreFile in the tmp dir.
*/
- public StoreFile.Writer createWriterInTmp(long maxKeyCount, Compression.Algorithm compression,
- boolean isCompaction, boolean includeMVCCReadpoint) throws IOException;
+ StoreFile.Writer createWriterInTmp(
+ long maxKeyCount,
+ Compression.Algorithm compression,
+ boolean isCompaction,
+ boolean includeMVCCReadpoint
+ ) throws IOException;
// Compaction oriented methods
- public boolean throttleCompaction(long compactionSize);
+ boolean throttleCompaction(long compactionSize);
/**
* getter for CompactionProgress object
* @return CompactionProgress object; can be null
*/
- public CompactionProgress getCompactionProgress();
+ CompactionProgress getCompactionProgress();
- public CompactionContext requestCompaction() throws IOException;
+ CompactionContext requestCompaction() throws IOException;
- public CompactionContext requestCompaction(int priority, CompactionRequest baseRequest)
+ CompactionContext requestCompaction(int priority, CompactionRequest baseRequest)
throws IOException;
- public void cancelRequestedCompaction(CompactionContext compaction);
+ void cancelRequestedCompaction(CompactionContext compaction);
- public List<StoreFile> compact(CompactionContext compaction) throws IOException;
+ List<StoreFile> compact(CompactionContext compaction) throws IOException;
/**
* @return true if we should run a major compaction.
*/
- public boolean isMajorCompaction() throws IOException;
+ boolean isMajorCompaction() throws IOException;
- public void triggerMajorCompaction();
+ void triggerMajorCompaction();
/**
* See if there's too much store files in this store
* @return true if number of store files is greater than the number defined in minFilesToCompact
*/
- public boolean needsCompaction();
+ boolean needsCompaction();
- public int getCompactPriority();
+ int getCompactPriority();
- public StoreFlushContext createFlushContext(long cacheFlushId);
+ StoreFlushContext createFlushContext(long cacheFlushId);
/**
* Call to complete a compaction. Its for the case where we find in the WAL a compaction
@@ -197,18 +205,18 @@ public interface Store extends HeapSize,
* See HBASE-2331.
* @param compaction
*/
- public void completeCompactionMarker(CompactionDescriptor compaction)
+ void completeCompactionMarker(CompactionDescriptor compaction)
throws IOException;
// Split oriented methods
- public boolean canSplit();
+ boolean canSplit();
/**
* Determines if Store should be split
* @return byte[] if store should be split, null otherwise.
*/
- public byte[] getSplitPoint();
+ byte[] getSplitPoint();
// Bulk Load methods
@@ -216,7 +224,7 @@ public interface Store extends HeapSize,
* This throws a WrongRegionException if the HFile does not fit in this region, or an
* InvalidHFileException if the HFile is not valid.
*/
- public void assertBulkLoadHFileOk(Path srcPath) throws IOException;
+ void assertBulkLoadHFileOk(Path srcPath) throws IOException;
/**
* This method should only be called from HRegion. It is assumed that the ranges of values in the
@@ -225,7 +233,7 @@ public interface Store extends HeapSize,
* @param srcPathStr
* @param sequenceId sequence Id associated with the HFile
*/
- public void bulkLoadHFile(String srcPathStr, long sequenceId) throws IOException;
+ void bulkLoadHFile(String srcPathStr, long sequenceId) throws IOException;
// General accessors into the state of the store
// TODO abstract some of this out into a metrics class
@@ -233,50 +241,50 @@ public interface Store extends HeapSize,
/**
* @return <tt>true</tt> if the store has any underlying reference files to older HFiles
*/
- public boolean hasReferences();
+ boolean hasReferences();
/**
* @return The size of this store's memstore, in bytes
*/
- public long getMemStoreSize();
+ long getMemStoreSize();
- public HColumnDescriptor getFamily();
+ HColumnDescriptor getFamily();
/**
* @return The maximum memstoreTS in all store files.
*/
- public long getMaxMemstoreTS();
+ long getMaxMemstoreTS();
/**
* @return the data block encoder
*/
- public HFileDataBlockEncoder getDataBlockEncoder();
+ HFileDataBlockEncoder getDataBlockEncoder();
/** @return aggregate size of all HStores used in the last compaction */
- public long getLastCompactSize();
+ long getLastCompactSize();
/** @return aggregate size of HStore */
- public long getSize();
+ long getSize();
/**
* @return Count of store files
*/
- public int getStorefilesCount();
+ int getStorefilesCount();
/**
* @return The size of the store files, in bytes, uncompressed.
*/
- public long getStoreSizeUncompressed();
+ long getStoreSizeUncompressed();
/**
* @return The size of the store files, in bytes.
*/
- public long getStorefilesSize();
+ long getStorefilesSize();
/**
* @return The size of the store file indexes, in bytes.
*/
- public long getStorefilesIndexSize();
+ long getStorefilesIndexSize();
/**
* Returns the total size of all index blocks in the data block indexes, including the root level,
@@ -284,14 +292,14 @@ public interface Store extends HeapSize,
* single-level indexes.
* @return the total size of block indexes in the store
*/
- public long getTotalStaticIndexSize();
+ long getTotalStaticIndexSize();
/**
* Returns the total byte size of all Bloom filter bit arrays. For compound Bloom filters even the
* Bloom blocks currently not loaded into the block cache are counted.
* @return the total size of all Bloom filters in the store
*/
- public long getTotalStaticBloomSize();
+ long getTotalStaticBloomSize();
// Test-helper methods
@@ -299,40 +307,40 @@ public interface Store extends HeapSize,
* Used for tests.
* @return cache configuration for this Store.
*/
- public CacheConfig getCacheConfig();
+ CacheConfig getCacheConfig();
/**
* @return the parent region info hosting this store
*/
- public HRegionInfo getRegionInfo();
+ HRegionInfo getRegionInfo();
- public RegionCoprocessorHost getCoprocessorHost();
+ RegionCoprocessorHost getCoprocessorHost();
- public boolean areWritesEnabled();
+ boolean areWritesEnabled();
/**
* @return The smallest mvcc readPoint across all the scanners in this
* region. Writes older than this readPoint, are included in every
* read operation.
*/
- public long getSmallestReadPoint();
+ long getSmallestReadPoint();
- public String getColumnFamilyName();
+ String getColumnFamilyName();
- public String getTableName();
+ String getTableName();
/*
* @param o Observer who wants to know about changes in set of Readers
*/
- public void addChangedReaderObserver(ChangedReadersObserver o);
+ void addChangedReaderObserver(ChangedReadersObserver o);
/*
* @param o Observer no longer interested in changes in set of Readers.
*/
- public void deleteChangedReaderObserver(ChangedReadersObserver o);
+ void deleteChangedReaderObserver(ChangedReadersObserver o);
/**
* @return Whether this store has too many store files.
*/
- public boolean hasTooManyStoreFiles();
+ boolean hasTooManyStoreFiles();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreConfigInformation.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreConfigInformation.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreConfigInformation.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreConfigInformation.java Wed Jul 10 17:53:33 2013
@@ -34,22 +34,22 @@ public interface StoreConfigInformation
* TODO: remove after HBASE-7252 is fixed.
* @return Gets the Memstore flush size for the region that this store works with.
*/
- public long getMemstoreFlushSize();
+ long getMemstoreFlushSize();
/**
* @return Gets the cf-specific time-to-live for store files.
*/
- public long getStoreFileTtl();
+ long getStoreFileTtl();
/**
* @return Gets the cf-specific compaction check frequency multiplier.
* The need for compaction (outside of normal checks during flush, open, etc.) will
* be ascertained every multiplier * HConstants.THREAD_WAKE_FREQUENCY milliseconds.
*/
- public long getCompactionCheckMultiplier();
+ long getCompactionCheckMultiplier();
/**
* The number of files required before flushes for this store will be blocked.
*/
- public long getBlockingFileCount();
+ long getBlockingFileCount();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFileManager.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFileManager.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFileManager.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFileManager.java Wed Jul 10 17:53:33 2013
@@ -44,40 +44,41 @@ public interface StoreFileManager {
* Loads the initial store files into empty StoreFileManager.
* @param storeFiles The files to load.
*/
- public abstract void loadFiles(List<StoreFile> storeFiles);
+ void loadFiles(List<StoreFile> storeFiles);
/**
* Adds new file, either for from MemStore flush or bulk insert, into the structure.
* @param sf New store file.
*/
- public abstract void insertNewFile(StoreFile sf);
+ void insertNewFile(StoreFile sf);
/**
* Adds compaction results into the structure.
* @param compactedFiles The input files for the compaction.
* @param results The resulting files for the compaction.
*/
- public abstract void addCompactionResults(
- Collection<StoreFile> compactedFiles, Collection<StoreFile> results);
+ void addCompactionResults(
+ Collection<StoreFile> compactedFiles, Collection<StoreFile> results
+ );
/**
* Clears all the files currently in use and returns them.
* @return The files previously in use.
*/
- public abstract ImmutableCollection<StoreFile> clearFiles();
+ ImmutableCollection<StoreFile> clearFiles();
/**
* Gets the snapshot of the store files currently in use. Can be used for things like metrics
* and checks; should not assume anything about relations between store files in the list.
* @return The list of StoreFiles.
*/
- public abstract Collection<StoreFile> getStorefiles();
+ Collection<StoreFile> getStorefiles();
/**
* Returns the number of files currently in use.
* @return The number of files.
*/
- public abstract int getStorefileCount();
+ int getStorefileCount();
/**
* Gets the store files to scan for a Scan or Get request.
@@ -86,8 +87,9 @@ public interface StoreFileManager {
* @param stopRow Stop row of the request.
* @return The list of files that are to be read for this request.
*/
- public abstract Collection<StoreFile> getFilesForScanOrGet(boolean isGet,
- byte[] startRow, byte[] stopRow);
+ Collection<StoreFile> getFilesForScanOrGet(
+ boolean isGet, byte[] startRow, byte[] stopRow
+ );
/**
* Gets initial, full list of candidate store files to check for row-key-before.
@@ -95,8 +97,9 @@ public interface StoreFileManager {
* @return The files that may have the key less than or equal to targetKey, in reverse
* order of new-ness, and preference for target key.
*/
- public abstract Iterator<StoreFile> getCandidateFilesForRowKeyBefore(
- KeyValue targetKey);
+ Iterator<StoreFile> getCandidateFilesForRowKeyBefore(
+ KeyValue targetKey
+ );
/**
* Updates the candidate list for finding row key before. Based on the list of candidates
@@ -109,8 +112,9 @@ public interface StoreFileManager {
* @param candidate The current best candidate found.
* @return The list to replace candidateFiles.
*/
- public abstract Iterator<StoreFile> updateCandidateFilesForRowKeyBefore(
- Iterator<StoreFile> candidateFiles, KeyValue targetKey, KeyValue candidate);
+ Iterator<StoreFile> updateCandidateFilesForRowKeyBefore(
+ Iterator<StoreFile> candidateFiles, KeyValue targetKey, KeyValue candidate
+ );
/**
@@ -118,10 +122,10 @@ public interface StoreFileManager {
* @return The mid-point, or null if no split is possible.
* @throws IOException
*/
- public abstract byte[] getSplitPoint() throws IOException;
+ byte[] getSplitPoint() throws IOException;
/**
* @return The store compaction priority.
*/
- public abstract int getStoreCompactionPriority();
+ int getStoreCompactionPriority();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/Dictionary.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/Dictionary.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/Dictionary.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/Dictionary.java Wed Jul 10 17:53:33 2013
@@ -28,7 +28,7 @@ import org.apache.hadoop.classification.
*/
@InterfaceAudience.Private
interface Dictionary {
- static final byte NOT_IN_DICTIONARY = -1;
+ byte NOT_IN_DICTIONARY = -1;
/**
* Gets an entry from the dictionary.
@@ -36,7 +36,7 @@ interface Dictionary {
* @param idx index of the entry
* @return the entry, or null if non existent
*/
- public byte[] getEntry(short idx);
+ byte[] getEntry(short idx);
/**
* Finds the index of an entry.
@@ -47,7 +47,7 @@ interface Dictionary {
* @param length Length beyond <code>offset</code> that comprises entry; must be > 0.
* @return the index of the entry, or {@link #NOT_IN_DICTIONARY} if not found
*/
- public short findEntry(byte[] data, int offset, int length);
+ short findEntry(byte[] data, int offset, int length);
/**
* Adds an entry to the dictionary.
@@ -62,10 +62,10 @@ interface Dictionary {
* @return the index of the entry
*/
- public short addEntry(byte[] data, int offset, int length);
+ short addEntry(byte[] data, int offset, int length);
/**
* Flushes the dictionary, empties all values.
*/
- public void clear();
+ void clear();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java Wed Jul 10 17:53:33 2013
@@ -41,26 +41,26 @@ import org.apache.hadoop.io.Writable;
@InterfaceAudience.Private
public interface HLog {
- public static final Log LOG = LogFactory.getLog(HLog.class);
+ Log LOG = LogFactory.getLog(HLog.class);
/** File Extension used while splitting an HLog into regions (HBASE-2312) */
- public static final String SPLITTING_EXT = "-splitting";
- public static final boolean SPLIT_SKIP_ERRORS_DEFAULT = false;
+ String SPLITTING_EXT = "-splitting";
+ boolean SPLIT_SKIP_ERRORS_DEFAULT = false;
/** The META region's HLog filename extension */
- public static final String META_HLOG_FILE_EXTN = ".meta";
+ String META_HLOG_FILE_EXTN = ".meta";
/**
* Configuration name of HLog Trailer's warning size. If a waltrailer's size is greater than the
* configured size, a warning is logged. This is used with Protobuf reader/writer.
*/
- public static final String WAL_TRAILER_WARN_SIZE =
+ String WAL_TRAILER_WARN_SIZE =
"hbase.regionserver.waltrailer.warn.size";
- public static final int DEFAULT_WAL_TRAILER_WARN_SIZE = 1024*1024; // 1MB
+ int DEFAULT_WAL_TRAILER_WARN_SIZE = 1024*1024; // 1MB
- static final Pattern EDITFILES_NAME_PATTERN = Pattern.compile("-?[0-9]+");
- public static final String RECOVERED_LOG_TMPFILE_SUFFIX = ".temp";
+ Pattern EDITFILES_NAME_PATTERN = Pattern.compile("-?[0-9]+");
+ String RECOVERED_LOG_TMPFILE_SUFFIX = ".temp";
- public interface Reader {
+ interface Reader {
/**
* @param fs File system.
@@ -88,7 +88,7 @@ public interface HLog {
WALTrailer getWALTrailer();
}
- public interface Writer {
+ interface Writer {
void init(FileSystem fs, Path path, Configuration c) throws IOException;
void close() throws IOException;
@@ -110,7 +110,7 @@ public interface HLog {
* Utility class that lets us keep track of the edit with it's key Only used
* when splitting logs
*/
- public static class Entry implements Writable {
+ class Entry implements Writable {
private WALEdit edit;
private HLogKey key;
@@ -185,19 +185,19 @@ public interface HLog {
*
* @param listener
*/
- public void registerWALActionsListener(final WALActionsListener listener);
+ void registerWALActionsListener(final WALActionsListener listener);
/**
* unregisters WALActionsListener
*
* @param listener
*/
- public boolean unregisterWALActionsListener(final WALActionsListener listener);
+ boolean unregisterWALActionsListener(final WALActionsListener listener);
/**
* @return Current state of the monotonically increasing file id.
*/
- public long getFilenum();
+ long getFilenum();
/**
* Called by HRegionServer when it opens a new region to ensure that log
@@ -208,12 +208,12 @@ public interface HLog {
* We'll set log edit/sequence number to this value if it is greater
* than the current value.
*/
- public void setSequenceNumber(final long newvalue);
+ void setSequenceNumber(final long newvalue);
/**
* @return log sequence number
*/
- public long getSequenceNumber();
+ long getSequenceNumber();
/**
* Roll the log writer. That is, start writing log messages to a new file.
@@ -228,7 +228,7 @@ public interface HLog {
* @throws org.apache.hadoop.hbase.exceptions.FailedLogCloseException
* @throws IOException
*/
- public byte[][] rollWriter() throws FailedLogCloseException, IOException;
+ byte[][] rollWriter() throws FailedLogCloseException, IOException;
/**
* Roll the log writer. That is, start writing log messages to a new file.
@@ -246,7 +246,7 @@ public interface HLog {
* @throws org.apache.hadoop.hbase.exceptions.FailedLogCloseException
* @throws IOException
*/
- public byte[][] rollWriter(boolean force) throws FailedLogCloseException,
+ byte[][] rollWriter(boolean force) throws FailedLogCloseException,
IOException;
/**
@@ -254,21 +254,22 @@ public interface HLog {
*
* @throws IOException
*/
- public void close() throws IOException;
+ void close() throws IOException;
/**
* Shut down the log and delete the log directory
*
* @throws IOException
*/
- public void closeAndDelete() throws IOException;
+ void closeAndDelete() throws IOException;
/**
* Same as {@link #appendNoSync(HRegionInfo, byte[], WALEdit, UUID, long, HTableDescriptor)},
* except it causes a sync on the log
*/
- public void append(HRegionInfo info, byte[] tableName, WALEdit edits,
- final long now, HTableDescriptor htd) throws IOException;
+ void append(
+ HRegionInfo info, byte[] tableName, WALEdit edits, final long now, HTableDescriptor htd
+ ) throws IOException;
/**
* Append a set of edits to the log. Log edits are keyed by (encoded)
@@ -281,8 +282,14 @@ public interface HLog {
* @param htd
* @param isInMemstore Whether the record is in memstore. False for system records.
*/
- public void append(HRegionInfo info, byte[] tableName, WALEdit edits,
- final long now, HTableDescriptor htd, boolean isInMemstore) throws IOException;
+ void append(
+ HRegionInfo info,
+ byte[] tableName,
+ WALEdit edits,
+ final long now,
+ HTableDescriptor htd,
+ boolean isInMemstore
+ ) throws IOException;
/**
* Append a set of edits to the log. Log edits are keyed by (encoded)
@@ -299,21 +306,27 @@ public interface HLog {
* @return txid of this transaction
* @throws IOException
*/
- public long appendNoSync(HRegionInfo info, byte[] tableName, WALEdit edits,
- UUID clusterId, final long now, HTableDescriptor htd) throws IOException;
+ long appendNoSync(
+ HRegionInfo info,
+ byte[] tableName,
+ WALEdit edits,
+ UUID clusterId,
+ final long now,
+ HTableDescriptor htd
+ ) throws IOException;
- public void hsync() throws IOException;
+ void hsync() throws IOException;
- public void hflush() throws IOException;
+ void hflush() throws IOException;
- public void sync() throws IOException;
+ void sync() throws IOException;
- public void sync(long txid) throws IOException;
+ void sync(long txid) throws IOException;
/**
* Obtain a log sequence number.
*/
- public long obtainSeqNum();
+ long obtainSeqNum();
/**
* WAL keeps track of the sequence numbers that were not yet flushed from memstores
@@ -330,13 +343,13 @@ public interface HLog {
* the resulting file as an upper-bound seqNum for that file), or NULL if flush
* should not be started.
*/
- public Long startCacheFlush(final byte[] encodedRegionName);
+ Long startCacheFlush(final byte[] encodedRegionName);
/**
* Complete the cache flush.
* @param encodedRegionName Encoded region name.
*/
- public void completeCacheFlush(final byte[] encodedRegionName);
+ void completeCacheFlush(final byte[] encodedRegionName);
/**
* Abort a cache flush. Call if the flush fails. Note that the only recovery
@@ -344,24 +357,24 @@ public interface HLog {
* snapshot content dropped by the failure gets restored to the memstore.v
* @param encodedRegionName Encoded region name.
*/
- public void abortCacheFlush(byte[] encodedRegionName);
+ void abortCacheFlush(byte[] encodedRegionName);
/**
* @return Coprocessor host.
*/
- public WALCoprocessorHost getCoprocessorHost();
+ WALCoprocessorHost getCoprocessorHost();
/**
* Get LowReplication-Roller status
*
* @return lowReplicationRollEnabled
*/
- public boolean isLowReplicationRollEnabled();
+ boolean isLowReplicationRollEnabled();
/** Gets the earliest sequence number in the memstore for this particular region.
* This can serve as best-effort "recent" WAL number for this region.
* @param encodedRegionName The region to get the number for.
* @return The number if present, HConstants.NO_SEQNUM if absent.
*/
- public long getEarliestMemstoreSeqNum(byte[] encodedRegionName);
+ long getEarliestMemstoreSeqNum(byte[] encodedRegionName);
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALActionsListener.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALActionsListener.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALActionsListener.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALActionsListener.java Wed Jul 10 17:53:33 2013
@@ -38,7 +38,7 @@ public interface WALActionsListener {
* @param oldPath the path to the old hlog
* @param newPath the path to the new hlog
*/
- public void preLogRoll(Path oldPath, Path newPath) throws IOException;
+ void preLogRoll(Path oldPath, Path newPath) throws IOException;
/**
* The WAL has been rolled. The oldPath can be null if this is
@@ -46,31 +46,31 @@ public interface WALActionsListener {
* @param oldPath the path to the old hlog
* @param newPath the path to the new hlog
*/
- public void postLogRoll(Path oldPath, Path newPath) throws IOException;
+ void postLogRoll(Path oldPath, Path newPath) throws IOException;
/**
* The WAL is going to be archived.
* @param oldPath the path to the old hlog
* @param newPath the path to the new hlog
*/
- public void preLogArchive(Path oldPath, Path newPath) throws IOException;
+ void preLogArchive(Path oldPath, Path newPath) throws IOException;
/**
* The WAL has been archived.
* @param oldPath the path to the old hlog
* @param newPath the path to the new hlog
*/
- public void postLogArchive(Path oldPath, Path newPath) throws IOException;
+ void postLogArchive(Path oldPath, Path newPath) throws IOException;
/**
* A request was made that the WAL be rolled.
*/
- public void logRollRequested();
+ void logRollRequested();
/**
* The WAL is about to close.
*/
- public void logCloseRequested();
+ void logCloseRequested();
/**
* Called before each write.
@@ -78,8 +78,9 @@ public interface WALActionsListener {
* @param logKey
* @param logEdit
*/
- public void visitLogEntryBeforeWrite(HRegionInfo info, HLogKey logKey,
- WALEdit logEdit);
+ void visitLogEntryBeforeWrite(
+ HRegionInfo info, HLogKey logKey, WALEdit logEdit
+ );
/**
*
@@ -87,7 +88,8 @@ public interface WALActionsListener {
* @param logKey
* @param logEdit
*/
- public void visitLogEntryBeforeWrite(HTableDescriptor htd, HLogKey logKey,
- WALEdit logEdit);
+ void visitLogEntryBeforeWrite(
+ HTableDescriptor htd, HLogKey logKey, WALEdit logEdit
+ );
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceInterface.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceInterface.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceInterface.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceInterface.java Wed Jul 10 17:53:33 2013
@@ -41,61 +41,63 @@ public interface ReplicationSourceInterf
* @param peerClusterId the id of the peer cluster
* @throws IOException
*/
- public void init(final Configuration conf,
- final FileSystem fs,
- final ReplicationSourceManager manager,
- final Stoppable stopper,
- final String peerClusterId) throws IOException;
+ void init(
+ final Configuration conf,
+ final FileSystem fs,
+ final ReplicationSourceManager manager,
+ final Stoppable stopper,
+ final String peerClusterId
+ ) throws IOException;
/**
* Add a log to the list of logs to replicate
* @param log path to the log to replicate
*/
- public void enqueueLog(Path log);
+ void enqueueLog(Path log);
/**
* Get the current log that's replicated
* @return the current log
*/
- public Path getCurrentPath();
+ Path getCurrentPath();
/**
* Start the replication
*/
- public void startup();
+ void startup();
/**
* End the replication
* @param reason why it's terminating
*/
- public void terminate(String reason);
+ void terminate(String reason);
/**
* End the replication
* @param reason why it's terminating
* @param cause the error that's causing it
*/
- public void terminate(String reason, Exception cause);
+ void terminate(String reason, Exception cause);
/**
* Get the id that the source is replicating to
*
* @return peer cluster id
*/
- public String getPeerClusterZnode();
+ String getPeerClusterZnode();
/**
* Get the id that the source is replicating to.
*
* @return peer cluster id
*/
- public String getPeerClusterId();
+ String getPeerClusterId();
/**
* Get a string representation of the current statistics
* for this source
* @return printable stats
*/
- public String getStats();
+ String getStats();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/Constants.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/Constants.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/Constants.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/Constants.java Wed Jul 10 17:53:33 2013
@@ -28,19 +28,19 @@ import org.apache.hadoop.classification.
@InterfaceAudience.Public
@InterfaceStability.Stable
public interface Constants {
- public static final String VERSION_STRING = "0.0.2";
+ String VERSION_STRING = "0.0.2";
- public static final int DEFAULT_MAX_AGE = 60 * 60 * 4; // 4 hours
+ int DEFAULT_MAX_AGE = 60 * 60 * 4; // 4 hours
- public static final int DEFAULT_LISTEN_PORT = 8080;
+ int DEFAULT_LISTEN_PORT = 8080;
- public static final String MIMETYPE_TEXT = "text/plain";
- public static final String MIMETYPE_HTML = "text/html";
- public static final String MIMETYPE_XML = "text/xml";
- public static final String MIMETYPE_BINARY = "application/octet-stream";
- public static final String MIMETYPE_PROTOBUF = "application/x-protobuf";
- public static final String MIMETYPE_PROTOBUF_IETF = "application/protobuf";
- public static final String MIMETYPE_JSON = "application/json";
+ String MIMETYPE_TEXT = "text/plain";
+ String MIMETYPE_HTML = "text/html";
+ String MIMETYPE_XML = "text/xml";
+ String MIMETYPE_BINARY = "application/octet-stream";
+ String MIMETYPE_PROTOBUF = "application/x-protobuf";
+ String MIMETYPE_PROTOBUF_IETF = "application/protobuf";
+ String MIMETYPE_JSON = "application/json";
- public static final String CRLF = "\r\n";
+ String CRLF = "\r\n";
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/ProtobufMessageHandler.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/ProtobufMessageHandler.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/ProtobufMessageHandler.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/ProtobufMessageHandler.java Wed Jul 10 17:53:33 2013
@@ -29,11 +29,11 @@ import org.apache.hadoop.classification.
* ProtobufMessageBodyProducer adapters.
*/
@InterfaceAudience.Private
-public abstract interface ProtobufMessageHandler {
+public interface ProtobufMessageHandler {
/**
* @return the protobuf represention of the model
*/
- public byte[] createProtobufOutput();
+ byte[] createProtobufOutput();
/**
* Initialize the model from a protobuf representation.
@@ -41,6 +41,6 @@ public abstract interface ProtobufMessag
* @return reference to self for convenience
* @throws IOException
*/
- public ProtobufMessageHandler getObjectFromMessage(byte[] message)
+ ProtobufMessageHandler getObjectFromMessage(byte[] message)
throws IOException;
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/CodeToClassAndBackFor96Migration.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/CodeToClassAndBackFor96Migration.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/CodeToClassAndBackFor96Migration.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/CodeToClassAndBackFor96Migration.java Wed Jul 10 17:53:33 2013
@@ -39,13 +39,13 @@ interface CodeToClassAndBackFor96Migrati
/**
* Static map that contains mapping from code to class
*/
- static final Map<Byte, Class<?>> CODE_TO_CLASS =
+ Map<Byte, Class<?>> CODE_TO_CLASS =
new HashMap<Byte, Class<?>>();
/**
* Static map that contains mapping from class to code
*/
- static final Map<Class<?>, Byte> CLASS_TO_CODE =
+ Map<Class<?>, Byte> CLASS_TO_CODE =
new HashMap<Class<?>, Byte>();
/**
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/thrift/IncrementCoalescerMBean.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/thrift/IncrementCoalescerMBean.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/thrift/IncrementCoalescerMBean.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/thrift/IncrementCoalescerMBean.java Wed Jul 10 17:53:33 2013
@@ -19,31 +19,31 @@
package org.apache.hadoop.hbase.thrift;
public interface IncrementCoalescerMBean {
- public int getQueueSize();
+ int getQueueSize();
- public int getMaxQueueSize();
+ int getMaxQueueSize();
- public void setMaxQueueSize(int newSize);
+ void setMaxQueueSize(int newSize);
- public long getPoolCompletedTaskCount();
+ long getPoolCompletedTaskCount();
- public long getPoolTaskCount();
+ long getPoolTaskCount();
- public int getPoolLargestPoolSize();
+ int getPoolLargestPoolSize();
- public int getCorePoolSize();
+ int getCorePoolSize();
- public void setCorePoolSize(int newCoreSize);
+ void setCorePoolSize(int newCoreSize);
- public int getMaxPoolSize();
+ int getMaxPoolSize();
- public void setMaxPoolSize(int newMaxSize);
+ void setMaxPoolSize(int newMaxSize);
- public long getFailedIncrements();
+ long getFailedIncrements();
- public long getSuccessfulCoalescings();
+ long getSuccessfulCoalescings();
- public long getTotalIncrements();
+ long getTotalIncrements();
- public long getCountersMapSize();
+ long getCountersMapSize();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java Wed Jul 10 17:53:33 2013
@@ -47,9 +47,9 @@ import org.apache.hadoop.hbase.client.HB
public final class Canary implements Tool {
// Sink interface used by the canary to outputs information
public interface Sink {
- public void publishReadFailure(HRegionInfo region);
- public void publishReadFailure(HRegionInfo region, HColumnDescriptor column);
- public void publishReadTiming(HRegionInfo region, HColumnDescriptor column, long msTime);
+ void publishReadFailure(HRegionInfo region);
+ void publishReadFailure(HRegionInfo region, HColumnDescriptor column);
+ void publishReadTiming(HRegionInfo region, HColumnDescriptor column, long msTime);
}
// Simple implementation of canary sink that allows to plot on
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/CancelableProgressable.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/CancelableProgressable.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/CancelableProgressable.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/CancelableProgressable.java Wed Jul 10 17:53:33 2013
@@ -34,6 +34,6 @@ public interface CancelableProgressable
* operation should be canceled and rolled back.
* @return whether to continue (true) or cancel (false) the operation
*/
- public boolean progress();
+ boolean progress();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java Wed Jul 10 17:53:33 2013
@@ -2903,7 +2903,7 @@ public class HBaseFsck extends Configure
}
public interface ErrorReporter {
- public static enum ERROR_CODE {
+ enum ERROR_CODE {
UNKNOWN, NO_META_REGION, NULL_META_REGION, NO_VERSION_FILE, NOT_IN_META_HDFS, NOT_IN_META,
NOT_IN_META_OR_DEPLOYED, NOT_IN_HDFS_OR_DEPLOYED, NOT_IN_HDFS, SERVER_DOES_NOT_MATCH_META, NOT_DEPLOYED,
MULTI_DEPLOYED, SHOULD_NOT_BE_DEPLOYED, MULTI_META_REGION, RS_CONNECT_FAILURE,
@@ -2912,20 +2912,26 @@ public class HBaseFsck extends Configure
ORPHAN_HDFS_REGION, LINGERING_SPLIT_PARENT, NO_TABLEINFO_FILE, LINGERING_REFERENCE_HFILE,
WRONG_USAGE, EMPTY_META_CELL, EXPIRED_TABLE_LOCK
}
- public void clear();
- public void report(String message);
- public void reportError(String message);
- public void reportError(ERROR_CODE errorCode, String message);
- public void reportError(ERROR_CODE errorCode, String message, TableInfo table);
- public void reportError(ERROR_CODE errorCode, String message, TableInfo table, HbckInfo info);
- public void reportError(ERROR_CODE errorCode, String message, TableInfo table, HbckInfo info1, HbckInfo info2);
- public int summarize();
- public void detail(String details);
- public ArrayList<ERROR_CODE> getErrorList();
- public void progress();
- public void print(String message);
- public void resetErrors();
- public boolean tableHasErrors(TableInfo table);
+ void clear();
+ void report(String message);
+ void reportError(String message);
+ void reportError(ERROR_CODE errorCode, String message);
+ void reportError(ERROR_CODE errorCode, String message, TableInfo table);
+ void reportError(ERROR_CODE errorCode, String message, TableInfo table, HbckInfo info);
+ void reportError(
+ ERROR_CODE errorCode,
+ String message,
+ TableInfo table,
+ HbckInfo info1,
+ HbckInfo info2
+ );
+ int summarize();
+ void detail(String details);
+ ArrayList<ERROR_CODE> getErrorList();
+ void progress();
+ void print(String message);
+ void resetErrors();
+ boolean tableHasErrors(TableInfo table);
}
static class PrintingErrorReporter implements ErrorReporter {
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/KeyRange.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/KeyRange.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/KeyRange.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/KeyRange.java Wed Jul 10 17:53:33 2013
@@ -25,7 +25,7 @@ import org.apache.hadoop.classification.
*/
@InterfaceAudience.Private
public interface KeyRange {
- abstract byte[] getStartKey();
+ byte[] getStartKey();
- abstract byte[] getEndKey();
+ byte[] getEndKey();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ModifyRegionUtils.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ModifyRegionUtils.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ModifyRegionUtils.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ModifyRegionUtils.java Wed Jul 10 17:53:33 2013
@@ -60,7 +60,7 @@ public abstract class ModifyRegionUtils
}
public interface RegionFillTask {
- public void fillRegion(final HRegion region) throws IOException;
+ void fillRegion(final HRegion region) throws IOException;
}
/**
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/RegionSplitter.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/RegionSplitter.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/RegionSplitter.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/RegionSplitter.java Wed Jul 10 17:53:33 2013
@@ -147,7 +147,7 @@ public class RegionSplitter {
* {@link RegionSplitter#rollingSplit(String, SplitAlgorithm, Configuration)} with the
* argument splitClassName giving the name of your class.
*/
- public static interface SplitAlgorithm {
+ public interface SplitAlgorithm {
/**
* Split a pre-existing region into 2 regions.
*
@@ -1046,4 +1046,4 @@ public class RegionSplitter {
+ "," + rowToStr(lastRow()) + "]";
}
}
-}
\ No newline at end of file
+}
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestCase.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestCase.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestCase.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestCase.java Wed Jul 10 17:53:33 2013
@@ -382,11 +382,11 @@ public abstract class HBaseTestCase exte
/**
* Implementors can flushcache.
*/
- public static interface FlushCache {
+ public interface FlushCache {
/**
* @throws IOException
*/
- public void flushcache() throws IOException;
+ void flushcache() throws IOException;
}
/**
@@ -395,23 +395,23 @@ public abstract class HBaseTestCase exte
*
* TOOD: Come up w/ a better name for this interface.
*/
- public static interface Incommon {
+ public interface Incommon {
/**
*
* @param delete
* @param writeToWAL
* @throws IOException
*/
- public void delete(Delete delete, boolean writeToWAL)
+ void delete(Delete delete, boolean writeToWAL)
throws IOException;
/**
* @param put
* @throws IOException
*/
- public void put(Put put) throws IOException;
+ void put(Put put) throws IOException;
- public Result get(Get get) throws IOException;
+ Result get(Get get) throws IOException;
/**
* @param family
@@ -421,8 +421,9 @@ public abstract class HBaseTestCase exte
* @return scanner for specified columns, first row and timestamp
* @throws IOException
*/
- public ScannerIncommon getScanner(byte [] family, byte [][] qualifiers,
- byte [] firstRow, long ts)
+ ScannerIncommon getScanner(
+ byte[] family, byte[][] qualifiers, byte[] firstRow, long ts
+ )
throws IOException;
}
@@ -520,10 +521,10 @@ public abstract class HBaseTestCase exte
public interface ScannerIncommon
extends Iterable<Result> {
- public boolean next(List<KeyValue> values)
+ boolean next(List<KeyValue> values)
throws IOException;
- public void close() throws IOException;
+ void close() throws IOException;
}
public static class ClientScannerIncommon implements ScannerIncommon {
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java Wed Jul 10 17:53:33 2013
@@ -198,7 +198,7 @@ public class PerformanceEvaluation exten
/**
* Implementations can have their status set.
*/
- static interface Status {
+ interface Status {
/**
* Sets status
* @param msg status message