You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2013/07/10 19:53:35 UTC

svn commit: r1501880 [3/4] - in /hbase/trunk: hbase-client/src/main/java/org/apache/hadoop/hbase/ hbase-client/src/main/java/org/apache/hadoop/hbase/catalog/ hbase-client/src/main/java/org/apache/hadoop/hbase/client/ hbase-client/src/main/java/org/apac...

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/handler/TotesHRegionInfo.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/handler/TotesHRegionInfo.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/handler/TotesHRegionInfo.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/handler/TotesHRegionInfo.java Wed Jul 10 17:53:33 2013
@@ -33,5 +33,5 @@ public interface TotesHRegionInfo {
   /**
    * @return HRegionInfo instance.
    */
-  public HRegionInfo getHRegionInfo();
+  HRegionInfo getHRegionInfo();
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandler.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandler.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandler.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandler.java Wed Jul 10 17:53:33 2013
@@ -29,16 +29,16 @@ import com.google.protobuf.Message;
  */
 @InterfaceAudience.Private
 public interface MonitoredRPCHandler extends MonitoredTask {
-  public abstract String getRPC();
-  public abstract String getRPC(boolean withParams);
-  public abstract long getRPCPacketLength();
-  public abstract String getClient();
-  public abstract long getRPCStartTime();
-  public abstract long getRPCQueueTime();
-  public abstract boolean isRPCRunning();
-  public abstract boolean isOperationRunning();
+  String getRPC();
+  String getRPC(boolean withParams);
+  long getRPCPacketLength();
+  String getClient();
+  long getRPCStartTime();
+  long getRPCQueueTime();
+  boolean isRPCRunning();
+  boolean isOperationRunning();
 
-  public abstract void setRPC(String methodName, Object [] params, long queueTime);
-  public abstract void setRPCPacket(Message param);
-  public abstract void setConnection(String clientAddress, int remotePort);
+  void setRPC(String methodName, Object[] params, long queueTime);
+  void setRPCPacket(Message param);
+  void setConnection(String clientAddress, int remotePort);
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java Wed Jul 10 17:53:33 2013
@@ -32,47 +32,47 @@ public interface MonitoredTask extends C
     ABORTED;
   }
 
-  public abstract long getStartTime();
-  public abstract String getDescription();
-  public abstract String getStatus();
-  public abstract long getStatusTime();
-  public abstract State getState();
-  public abstract long getStateTime();
-  public abstract long getCompletionTimestamp();
-
-  public abstract void markComplete(String msg);
-  public abstract void pause(String msg);
-  public abstract void resume(String msg);
-  public abstract void abort(String msg);
-  public abstract void expireNow();
+  long getStartTime();
+  String getDescription();
+  String getStatus();
+  long getStatusTime();
+  State getState();
+  long getStateTime();
+  long getCompletionTimestamp();
+
+  void markComplete(String msg);
+  void pause(String msg);
+  void resume(String msg);
+  void abort(String msg);
+  void expireNow();
 
-  public abstract void setStatus(String status);
-  public abstract void setDescription(String description);
+  void setStatus(String status);
+  void setDescription(String description);
 
   /**
    * Explicitly mark this status as able to be cleaned up,
    * even though it might not be complete.
    */
-  public abstract void cleanup();
+  void cleanup();
 
   /**
    * Public exposure of Object.clone() in order to allow clients to easily 
    * capture current state.
    * @return a copy of the object whose references will not change
    */
-  public abstract MonitoredTask clone();
+  MonitoredTask clone();
 
   /**
    * Creates a string map of internal details for extensible exposure of 
    * monitored tasks.
    * @return A Map containing information for this task.
    */
-  public abstract Map<String, Object> toMap() throws IOException;
+  Map<String, Object> toMap() throws IOException;
 
   /**
    * Creates a JSON object for parseable exposure of monitored tasks.
    * @return An encoded JSON object containing information for this task.
    */
-  public abstract String toJSON() throws IOException;
+  String toJSON() throws IOException;
 
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinatorRpcs.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinatorRpcs.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinatorRpcs.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinatorRpcs.java Wed Jul 10 17:53:33 2013
@@ -39,7 +39,7 @@ public interface ProcedureCoordinatorRpc
    * @param listener
    * @return true if succeed, false if encountered initialization errors.
    */
-  public boolean start(final ProcedureCoordinator listener);
+  boolean start(final ProcedureCoordinator listener);
 
   /**
    * Notify the members that the coordinator has aborted the procedure and that it should release
@@ -50,7 +50,7 @@ public interface ProcedureCoordinatorRpc
    * @throws IOException if the rpcs can't reach the other members of the procedure (and can't
    *           recover).
    */
-  public void sendAbortToMembers(Procedure procName, ForeignException cause) throws IOException;
+  void sendAbortToMembers(Procedure procName, ForeignException cause) throws IOException;
 
   /**
    * Notify the members to acquire barrier for the procedure
@@ -61,7 +61,7 @@ public interface ProcedureCoordinatorRpc
    * @throws IllegalArgumentException if the procedure was already marked as failed
    * @throws IOException if we can't reach the remote notification mechanism
    */
-  public void sendGlobalBarrierAcquire(Procedure procName, byte[] info, List<String> members)
+  void sendGlobalBarrierAcquire(Procedure procName, byte[] info, List<String> members)
       throws IOException, IllegalArgumentException;
 
   /**
@@ -74,12 +74,12 @@ public interface ProcedureCoordinatorRpc
    * @param members members to tell we have reached in-barrier phase
    * @throws IOException if we can't reach the remote notification mechanism
    */
-  public void sendGlobalBarrierReached(Procedure procName, List<String> members) throws IOException;
+  void sendGlobalBarrierReached(Procedure procName, List<String> members) throws IOException;
 
   /**
    * Notify Members to reset the distributed state for procedure
    * @param procName name of the procedure to reset
    * @throws IOException if the remote notification mechanism cannot be reached
    */
-  public void resetMembers(Procedure procName) throws IOException;
+  void resetMembers(Procedure procName) throws IOException;
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureMemberRpcs.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureMemberRpcs.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureMemberRpcs.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureMemberRpcs.java Wed Jul 10 17:53:33 2013
@@ -35,13 +35,13 @@ public interface ProcedureMemberRpcs ext
   /**
    * Initialize and start any threads or connections the member needs.
    */
-  public void start(final String memberName, final ProcedureMember member);
+  void start(final String memberName, final ProcedureMember member);
 
   /**
    * Each subprocedure is being executed on a member.  This is the identifier for the member.
    * @return the member name
    */
-  public String getMemberName();
+  String getMemberName();
 
   /**
    * Notify the coordinator that we aborted the specified {@link Subprocedure}
@@ -51,7 +51,7 @@ public interface ProcedureMemberRpcs ext
    * @throws IOException thrown when the rpcs can't reach the other members of the procedure (and
    *  thus can't recover).
    */
-  public void sendMemberAborted(Subprocedure sub, ForeignException cause) throws IOException;
+  void sendMemberAborted(Subprocedure sub, ForeignException cause) throws IOException;
 
   /**
    * Notify the coordinator that the specified {@link Subprocedure} has acquired the locally required
@@ -60,7 +60,7 @@ public interface ProcedureMemberRpcs ext
    * @param sub the specified {@link Subprocedure}
    * @throws IOException if we can't reach the coordinator
    */
-  public void sendMemberAcquired(Subprocedure sub) throws IOException;
+  void sendMemberAcquired(Subprocedure sub) throws IOException;
 
   /**
    * Notify the coordinator that the specified {@link Subprocedure} has completed the work that
@@ -69,5 +69,5 @@ public interface ProcedureMemberRpcs ext
    * @param sub the specified {@link Subprocedure}
    * @throws IOException if we can't reach the coordinator
    */
-  public void sendMemberCompleted(Subprocedure sub) throws IOException;
-}
\ No newline at end of file
+  void sendMemberCompleted(Subprocedure sub) throws IOException;
+}

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/SubprocedureFactory.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/SubprocedureFactory.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/SubprocedureFactory.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/SubprocedureFactory.java Wed Jul 10 17:53:33 2013
@@ -36,5 +36,5 @@ public interface SubprocedureFactory {
    *           request
    * @throws IllegalStateException if the current runner cannot accept any more new requests
    */
-  public Subprocedure buildSubprocedure(String procName, byte[] procArgs);
+  Subprocedure buildSubprocedure(String procName, byte[] procArgs);
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ColumnTracker.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ColumnTracker.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ColumnTracker.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ColumnTracker.java Wed Jul 10 17:53:33 2013
@@ -56,25 +56,26 @@ public interface ColumnTracker {
    * @throws IOException in case there is an internal consistency problem
    *      caused by a data corruption.
    */
-  public ScanQueryMatcher.MatchCode checkColumn(byte[] bytes, int offset,
-      int length, long ttl, byte type, boolean ignoreCount)
+  ScanQueryMatcher.MatchCode checkColumn(
+    byte[] bytes, int offset, int length, long ttl, byte type, boolean ignoreCount
+  )
       throws IOException;
 
   /**
    * Updates internal variables in between files
    */
-  public void update();
+  void update();
 
   /**
    * Resets the Matcher
    */
-  public void reset();
+  void reset();
 
   /**
    *
    * @return <code>true</code> when done.
    */
-  public boolean done();
+  boolean done();
 
   /**
    * Used by matcher and scan/get to get a hint of the next column
@@ -87,13 +88,14 @@ public interface ColumnTracker {
    *
    * @return null, or a ColumnCount that we should seek to
    */
-  public ColumnCount getColumnHint();
+  ColumnCount getColumnHint();
 
   /**
    * Retrieve the MatchCode for the next row or column
    */
-  public MatchCode getNextRowOrNextColumn(byte[] bytes, int offset,
-      int qualLength);
+  MatchCode getNextRowOrNextColumn(
+    byte[] bytes, int offset, int qualLength
+  );
 
   /**
    * Give the tracker a chance to declare it's done based on only the timestamp
@@ -102,5 +104,5 @@ public interface ColumnTracker {
    * @param timestamp
    * @return <code>true</code> to early out based on timestamp.
    */
-  public boolean isDone(long timestamp);
+  boolean isDone(long timestamp);
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactionRequestor.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactionRequestor.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactionRequestor.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactionRequestor.java Wed Jul 10 17:53:33 2013
@@ -34,7 +34,7 @@ public interface CompactionRequestor {
    *         compactions were started
    * @throws IOException
    */
-  public List<CompactionRequest> requestCompaction(final HRegion r, final String why)
+  List<CompactionRequest> requestCompaction(final HRegion r, final String why)
       throws IOException;
 
   /**
@@ -47,8 +47,9 @@ public interface CompactionRequestor {
    *         compactions were started
    * @throws IOException
    */
-  public List<CompactionRequest> requestCompaction(final HRegion r, final String why,
-      List<Pair<CompactionRequest, Store>> requests)
+  List<CompactionRequest> requestCompaction(
+    final HRegion r, final String why, List<Pair<CompactionRequest, Store>> requests
+  )
       throws IOException;
 
   /**
@@ -60,8 +61,9 @@ public interface CompactionRequestor {
    * @return The created {@link CompactionRequest} or <tt>null</tt> if no compaction was started.
    * @throws IOException
    */
-  public CompactionRequest requestCompaction(final HRegion r, final Store s, final String why,
-      CompactionRequest request) throws IOException;
+  CompactionRequest requestCompaction(
+    final HRegion r, final Store s, final String why, CompactionRequest request
+  ) throws IOException;
 
   /**
    * @param r Region to compact
@@ -74,8 +76,9 @@ public interface CompactionRequestor {
    *         compactions were started.
    * @throws IOException
    */
-  public List<CompactionRequest> requestCompaction(final HRegion r, final String why, int pri,
-      List<Pair<CompactionRequest, Store>> requests) throws IOException;
+  List<CompactionRequest> requestCompaction(
+    final HRegion r, final String why, int pri, List<Pair<CompactionRequest, Store>> requests
+  ) throws IOException;
 
   /**
    * @param r Region to compact
@@ -87,6 +90,7 @@ public interface CompactionRequestor {
    * @return The created {@link CompactionRequest} or <tt>null</tt> if no compaction was started
    * @throws IOException
    */
-  public CompactionRequest requestCompaction(final HRegion r, final Store s, final String why,
-      int pri, CompactionRequest request) throws IOException;
+  CompactionRequest requestCompaction(
+    final HRegion r, final Store s, final String why, int pri, CompactionRequest request
+  ) throws IOException;
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DeleteTracker.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DeleteTracker.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DeleteTracker.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DeleteTracker.java Wed Jul 10 17:53:33 2013
@@ -43,8 +43,9 @@ public interface DeleteTracker {
    * @param timestamp timestamp
    * @param type delete type as byte
    */
-  public void add(byte [] buffer, int qualifierOffset, int qualifierLength,
-      long timestamp, byte type);
+  void add(
+    byte[] buffer, int qualifierOffset, int qualifierLength, long timestamp, byte type
+  );
 
   /**
    * Check if the specified KeyValue buffer has been deleted by a previously
@@ -55,13 +56,14 @@ public interface DeleteTracker {
    * @param timestamp timestamp
    * @return deleteResult The result tells whether the KeyValue is deleted and why
    */
-  public DeleteResult isDeleted(byte [] buffer, int qualifierOffset,
-      int qualifierLength, long timestamp);
+  DeleteResult isDeleted(
+    byte[] buffer, int qualifierOffset, int qualifierLength, long timestamp
+  );
 
   /**
    * @return true if there are no current delete, false otherwise
    */
-  public boolean isEmpty();
+  boolean isEmpty();
 
   /**
    * Called at the end of every StoreFile.
@@ -69,14 +71,14 @@ public interface DeleteTracker {
    * Many optimized implementations of Trackers will require an update at
    * when the end of each StoreFile is reached.
    */
-  public void update();
+  void update();
 
   /**
    * Called between rows.
    * <p>
    * This clears everything as if a new DeleteTracker was instantiated.
    */
-  public void reset();
+  void reset();
 
 
   /**
@@ -102,7 +104,7 @@ public interface DeleteTracker {
    * Based on the delete result, the ScanQueryMatcher will decide the next
    * operation
    */
-  public static enum DeleteResult {
+  enum DeleteResult {
     FAMILY_DELETED, // The KeyValue is deleted by a delete family.
     FAMILY_VERSION_DELETED, // The KeyValue is deleted by a delete family version.
     COLUMN_DELETED, // The KeyValue is deleted by a delete column.

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java Wed Jul 10 17:53:33 2013
@@ -5602,7 +5602,7 @@ public class HRegion implements HeapSize
    * bulkLoadHFile() to perform any necessary
    * pre/post processing of a given bulkload call
    */
-  public static interface BulkLoadListener {
+  public interface BulkLoadListener {
 
     /**
      * Called before an HFile is actually loaded

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/InternalScanner.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/InternalScanner.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/InternalScanner.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/InternalScanner.java Wed Jul 10 17:53:33 2013
@@ -47,7 +47,7 @@ public interface InternalScanner extends
    * @return true if more rows exist after this one, false if scanner is done
    * @throws IOException e
    */
-  public boolean next(List<KeyValue> results) throws IOException;
+  boolean next(List<KeyValue> results) throws IOException;
 
   /**
    * Grab the next row's worth of values with a limit on the number of values
@@ -57,11 +57,11 @@ public interface InternalScanner extends
    * @return true if more rows exist after this one, false if scanner is done
    * @throws IOException e
    */
-  public boolean next(List<KeyValue> result, int limit) throws IOException;
+  boolean next(List<KeyValue> result, int limit) throws IOException;
 
   /**
    * Closes the scanner and releases any resources it has allocated
    * @throws IOException
    */
-  public void close() throws IOException;
+  void close() throws IOException;
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/KeyValueScanner.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/KeyValueScanner.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/KeyValueScanner.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/KeyValueScanner.java Wed Jul 10 17:53:33 2013
@@ -34,20 +34,20 @@ public interface KeyValueScanner {
    * Look at the next KeyValue in this scanner, but do not iterate scanner.
    * @return the next KeyValue
    */
-  public KeyValue peek();
+  KeyValue peek();
 
   /**
    * Return the next KeyValue in this scanner, iterating the scanner
    * @return the next KeyValue
    */
-  public KeyValue next() throws IOException;
+  KeyValue next() throws IOException;
 
   /**
    * Seek the scanner at or after the specified KeyValue.
    * @param key seek value
    * @return true if scanner has values left, false if end of scanner
    */
-  public boolean seek(KeyValue key) throws IOException;
+  boolean seek(KeyValue key) throws IOException;
 
   /**
    * Reseek the scanner at or after the specified KeyValue.
@@ -57,7 +57,7 @@ public interface KeyValueScanner {
    * @param key seek value (should be non-null)
    * @return true if scanner has values left, false if end of scanner
    */
-  public boolean reseek(KeyValue key) throws IOException;
+  boolean reseek(KeyValue key) throws IOException;
 
   /**
    * Get the sequence id associated with this KeyValueScanner. This is required
@@ -65,12 +65,12 @@ public interface KeyValueScanner {
    * The default implementation for this would be to return 0. A file having
    * lower sequence id will be considered to be the older one.
    */
-  public long getSequenceID();
+  long getSequenceID();
 
   /**
    * Close the KeyValue scanner.
    */
-  public void close();
+  void close();
 
   /**
    * Allows to filter out scanners (both StoreFile and memstore) that we don't
@@ -82,8 +82,9 @@ public interface KeyValueScanner {
    *          this query, based on TTL
    * @return true if the scanner should be included in the query
    */
-  public boolean shouldUseScanner(Scan scan, SortedSet<byte[]> columns,
-      long oldestUnexpiredTS);
+  boolean shouldUseScanner(
+    Scan scan, SortedSet<byte[]> columns, long oldestUnexpiredTS
+  );
 
   // "Lazy scanner" optimizations
 
@@ -97,7 +98,7 @@ public interface KeyValueScanner {
    * @param forward do a forward-only "reseek" instead of a random-access seek
    * @param useBloom whether to enable multi-column Bloom filter optimization
    */
-  public boolean requestSeek(KeyValue kv, boolean forward, boolean useBloom)
+  boolean requestSeek(KeyValue kv, boolean forward, boolean useBloom)
       throws IOException;
 
   /**
@@ -106,7 +107,7 @@ public interface KeyValueScanner {
    * store scanner bubbles up to the top of the key-value heap. This method is
    * then used to ensure the top store file scanner has done a seek operation.
    */
-  public boolean realSeekDone();
+  boolean realSeekDone();
 
   /**
    * Does the real seek operation in case it was skipped by
@@ -115,11 +116,11 @@ public interface KeyValueScanner {
    * of the scanners). The easiest way to achieve this is to call
    * {@link #realSeekDone()} first.
    */
-  public void enforceSeek() throws IOException;
+  void enforceSeek() throws IOException;
 
   /**
    * @return true if this is a file scanner. Otherwise a memory scanner is
    *         assumed.
    */
-  public boolean isFileScanner();  
+  boolean isFileScanner();
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/LastSequenceId.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/LastSequenceId.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/LastSequenceId.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/LastSequenceId.java Wed Jul 10 17:53:33 2013
@@ -29,5 +29,5 @@ public interface LastSequenceId {
    * @param regionName Encoded region name
    * @return Last flushed sequence Id for regionName or -1 if it can't be determined
    */
-  public long getLastSequenceId(byte[] regionName);
+  long getLastSequenceId(byte[] regionName);
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/LeaseListener.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/LeaseListener.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/LeaseListener.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/LeaseListener.java Wed Jul 10 17:53:33 2013
@@ -32,5 +32,5 @@ import org.apache.hadoop.classification.
 @InterfaceAudience.Private
 public interface LeaseListener {
   /** When a lease expires, this method is called. */
-  public void leaseExpired();
+  void leaseExpired();
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/OnlineRegions.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/OnlineRegions.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/OnlineRegions.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/OnlineRegions.java Wed Jul 10 17:53:33 2013
@@ -35,7 +35,7 @@ interface OnlineRegions extends Server {
    * Add to online regions.
    * @param r
    */
-  public void addToOnlineRegions(final HRegion r);
+  void addToOnlineRegions(final HRegion r);
 
   /**
    * This method removes HRegion corresponding to hri from the Map of onlineRegions.
@@ -44,7 +44,7 @@ interface OnlineRegions extends Server {
    * @param destination Destination, if any, null otherwise.
    * @return True if we removed a region from online list.
    */
-  public boolean removeFromOnlineRegions(final HRegion r, ServerName destination);
+  boolean removeFromOnlineRegions(final HRegion r, ServerName destination);
 
   /**
    * Return {@link HRegion} instance.
@@ -54,7 +54,7 @@ interface OnlineRegions extends Server {
    * @return HRegion for the passed encoded <code>encodedRegionName</code> or
    * null if named region is not member of the online regions.
    */
-  public HRegion getFromOnlineRegions(String encodedRegionName);
+  HRegion getFromOnlineRegions(String encodedRegionName);
 
    /**
     * Get all online regions of a table in this RS.
@@ -62,5 +62,5 @@ interface OnlineRegions extends Server {
     * @return List of HRegion
     * @throws java.io.IOException
     */
-   public List<HRegion> getOnlineRegions(byte[] tableName) throws IOException;
+   List<HRegion> getOnlineRegions(byte[] tableName) throws IOException;
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionScanner.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionScanner.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionScanner.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionScanner.java Wed Jul 10 17:53:33 2013
@@ -34,13 +34,13 @@ public interface RegionScanner extends I
   /**
    * @return The RegionInfo for this scanner.
    */
-  public HRegionInfo getRegionInfo();
+  HRegionInfo getRegionInfo();
 
   /**
    * @return True if a filter indicates that this scanner will return no further rows.
    * @throws IOException in case of I/O failure on a filter.
    */
-  public boolean isFilterDone() throws IOException;
+  boolean isFilterDone() throws IOException;
 
   /**
    * Do a reseek to the required row. Should not be used to seek to a key which
@@ -52,17 +52,17 @@ public interface RegionScanner extends I
    *           if row is null
    *
    */
-  public boolean reseek(byte[] row) throws IOException;
+  boolean reseek(byte[] row) throws IOException;
 
   /**
    * @return The preferred max buffersize. See {@link Scan#setMaxResultSize(long)}
    */
-  public long getMaxResultSize();
+  long getMaxResultSize();
 
   /**
    * @return The Scanner's MVCC readPt see {@link MultiVersionConsistencyControl}
    */
-  public long getMvccReadPoint();
+  long getMvccReadPoint();
 
   /**
    * Grab the next row's worth of values with the default limit on the number of values
@@ -74,7 +74,7 @@ public interface RegionScanner extends I
    * @return true if more rows exist after this one, false if scanner is done
    * @throws IOException e
    */
-  public boolean nextRaw(List<KeyValue> result) throws IOException;
+  boolean nextRaw(List<KeyValue> result) throws IOException;
 
   /**
    * Grab the next row's worth of values with a limit on the number of values
@@ -102,5 +102,5 @@ public interface RegionScanner extends I
    * @return true if more rows exist after this one, false if scanner is done
    * @throws IOException e
    */
-  public boolean nextRaw(List<KeyValue> result, int limit) throws IOException;
+  boolean nextRaw(List<KeyValue> result, int limit) throws IOException;
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerServices.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerServices.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerServices.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerServices.java Wed Jul 10 17:53:33 2013
@@ -40,31 +40,31 @@ public interface RegionServerServices ex
   /**
    * @return True if this regionserver is stopping.
    */
-  public boolean isStopping();
+  boolean isStopping();
 
   /** @return the HLog for a particular region. Pass null for getting the
    * default (common) WAL */
-  public HLog getWAL(HRegionInfo regionInfo) throws IOException;
+  HLog getWAL(HRegionInfo regionInfo) throws IOException;
 
   /**
    * @return Implementation of {@link CompactionRequestor} or null.
    */
-  public CompactionRequestor getCompactionRequester();
+  CompactionRequestor getCompactionRequester();
 
   /**
    * @return Implementation of {@link FlushRequester} or null.
    */
-  public FlushRequester getFlushRequester();
+  FlushRequester getFlushRequester();
 
   /**
    * @return the RegionServerAccounting for this Region Server
    */
-  public RegionServerAccounting getRegionServerAccounting();
+  RegionServerAccounting getRegionServerAccounting();
 
   /**
    * @return RegionServer's instance of {@link TableLockManager}
    */
-  public TableLockManager getTableLockManager();
+  TableLockManager getTableLockManager();
 
   /**
    * Tasks to perform after region open to complete deploy of region on
@@ -75,42 +75,42 @@ public interface RegionServerServices ex
    * @throws KeeperException
    * @throws IOException
    */
-  public void postOpenDeployTasks(final HRegion r, final CatalogTracker ct)
+  void postOpenDeployTasks(final HRegion r, final CatalogTracker ct)
   throws KeeperException, IOException;
 
   /**
    * Returns a reference to the region server's RPC server
    */
-  public RpcServerInterface getRpcServer();
+  RpcServerInterface getRpcServer();
 
   /**
    * Get the regions that are currently being opened or closed in the RS
    * @return map of regions in transition in this RS
    */
-  public ConcurrentMap<byte[], Boolean> getRegionsInTransitionInRS();
+  ConcurrentMap<byte[], Boolean> getRegionsInTransitionInRS();
 
   /**
    * @return Return the FileSystem object used by the regionserver
    */
-  public FileSystem getFileSystem();
+  FileSystem getFileSystem();
 
   /**
    * @return The RegionServer's "Leases" service
    */
-  public Leases getLeases();
+  Leases getLeases();
 
   /**
    * @return hbase executor service
    */
-  public ExecutorService getExecutorService();
+  ExecutorService getExecutorService();
 
   /**
    * @return The RegionServer's CatalogTracker
    */
-  public CatalogTracker getCatalogTracker();
+  CatalogTracker getCatalogTracker();
 
   /**
    * @return set of recovering regions on the hosting region server
    */
-  public Map<String, HRegion> getRecoveringRegions();
+  Map<String, HRegion> getRecoveringRegions();
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationService.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationService.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationService.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationService.java Wed Jul 10 17:53:33 2013
@@ -38,17 +38,18 @@ public interface ReplicationService {
    * Initializes the replication service object.
    * @throws IOException
    */
-  public void initialize(Server rs, FileSystem fs, Path logdir,
-                         Path oldLogDir) throws IOException;
+  void initialize(
+    Server rs, FileSystem fs, Path logdir, Path oldLogDir
+  ) throws IOException;
 
   /**
    * Start replication services.
    * @throws IOException
    */
-  public void startReplicationService() throws IOException;
+  void startReplicationService() throws IOException;
 
   /**
    * Stops replication service.
    */
-  public void stopReplicationService();
+  void stopReplicationService();
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSinkService.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSinkService.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSinkService.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSinkService.java Wed Jul 10 17:53:33 2013
@@ -38,5 +38,5 @@ public interface ReplicationSinkService 
    * @param cells Cells that the WALEntries refer to (if cells is non-null)
    * @throws IOException
    */
-  public void replicateLogEntries(List<WALEntry> entries, CellScanner cells) throws IOException;
-}
\ No newline at end of file
+  void replicateLogEntries(List<WALEntry> entries, CellScanner cells) throws IOException;
+}

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.java Wed Jul 10 17:53:33 2013
@@ -32,5 +32,5 @@ public interface ReplicationSourceServic
    * Returns a WALObserver for the service. This is needed to 
    * observe log rolls and log archival events.
    */
-  public WALActionsListener getWALActionsListener();
-}
\ No newline at end of file
+  WALActionsListener getWALActionsListener();
+}

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java Wed Jul 10 17:53:33 2013
@@ -640,13 +640,13 @@ public class SplitLogWorker extends ZooK
    * is better to have workers prepare the task and then have the
    * {@link SplitLogManager} commit the work in SplitLogManager.TaskFinisher
    */
-  static public interface TaskExecutor {
-    static public enum Status {
+  public interface TaskExecutor {
+    enum Status {
       DONE(),
       ERR(),
       RESIGNED(),
       PREEMPTED()
     }
-    public Status exec(String name, CancelableProgressable p);
+    Status exec(String name, CancelableProgressable p);
   }
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java Wed Jul 10 17:53:33 2013
@@ -50,14 +50,13 @@ public interface Store extends HeapSize,
 
   /* The default priority for user-specified compaction requests.
    * The user gets top priority unless we have blocking compactions. (Pri <= 0)
-   */
-  public static final int PRIORITY_USER = 1;
-  public static final int NO_PRIORITY = Integer.MIN_VALUE;
+   */ int PRIORITY_USER = 1;
+  int NO_PRIORITY = Integer.MIN_VALUE;
 
   // General Accessors
-  public KeyValue.KVComparator getComparator();
+  KeyValue.KVComparator getComparator();
 
-  public Collection<StoreFile> getStorefiles();
+  Collection<StoreFile> getStorefiles();
 
   /**
    * Close all the readers We don't need to worry about subsequent requests because the HRegion
@@ -65,7 +64,7 @@ public interface Store extends HeapSize,
    * @return the {@link StoreFile StoreFiles} that were previously being used.
    * @throws IOException on failure
    */
-  public Collection<StoreFile> close() throws IOException;
+  Collection<StoreFile> close() throws IOException;
 
   /**
    * Return a scanner for both the memstore and the HStore files. Assumes we are not in a
@@ -75,7 +74,7 @@ public interface Store extends HeapSize,
    * @return a scanner over the current key values
    * @throws IOException on failure
    */
-  public KeyValueScanner getScanner(Scan scan, final NavigableSet<byte[]> targetCols)
+  KeyValueScanner getScanner(Scan scan, final NavigableSet<byte[]> targetCols)
       throws IOException;
 
   /**
@@ -89,11 +88,16 @@ public interface Store extends HeapSize,
    * @param stopRow
    * @return all scanners for this store
    */
-  public List<KeyValueScanner> getScanners(boolean cacheBlocks,
-      boolean isGet, boolean isCompaction, ScanQueryMatcher matcher, byte[] startRow,
-      byte[] stopRow) throws IOException;
+  List<KeyValueScanner> getScanners(
+    boolean cacheBlocks,
+    boolean isGet,
+    boolean isCompaction,
+    ScanQueryMatcher matcher,
+    byte[] startRow,
+    byte[] stopRow
+  ) throws IOException;
 
-  public ScanInfo getScanInfo();
+  ScanInfo getScanInfo();
 
   /**
    * Adds or replaces the specified KeyValues.
@@ -108,14 +112,14 @@ public interface Store extends HeapSize,
    * @return memstore size delta
    * @throws IOException
    */
-  public long upsert(Iterable<? extends Cell> cells, long readpoint) throws IOException;
+  long upsert(Iterable<? extends Cell> cells, long readpoint) throws IOException;
 
   /**
    * Adds a value to the memstore
    * @param kv
    * @return memstore size delta
    */
-  public long add(KeyValue kv);
+  long add(KeyValue kv);
 
   /**
    * When was the last edit done in the memstore
@@ -127,7 +131,7 @@ public interface Store extends HeapSize,
    * key & memstoreTS value of the kv parameter.
    * @param kv
    */
-  public void rollback(final KeyValue kv);
+  void rollback(final KeyValue kv);
 
   /**
    * Find the key that matches <i>row</i> exactly, or the one that immediately precedes it. WARNING:
@@ -141,9 +145,9 @@ public interface Store extends HeapSize,
    * @return Found keyvalue or null if none found.
    * @throws IOException
    */
-  public KeyValue getRowKeyAtOrBefore(final byte[] row) throws IOException;
+  KeyValue getRowKeyAtOrBefore(final byte[] row) throws IOException;
 
-  public FileSystem getFileSystem();
+  FileSystem getFileSystem();
 
   /*
    * @param maxKeyCount
@@ -152,44 +156,48 @@ public interface Store extends HeapSize,
    * @param includeMVCCReadpoint whether we should out the MVCC readpoint
    * @return Writer for a new StoreFile in the tmp dir.
    */
-  public StoreFile.Writer createWriterInTmp(long maxKeyCount, Compression.Algorithm compression,
-      boolean isCompaction, boolean includeMVCCReadpoint) throws IOException;
+  StoreFile.Writer createWriterInTmp(
+    long maxKeyCount,
+    Compression.Algorithm compression,
+    boolean isCompaction,
+    boolean includeMVCCReadpoint
+  ) throws IOException;
 
   // Compaction oriented methods
 
-  public boolean throttleCompaction(long compactionSize);
+  boolean throttleCompaction(long compactionSize);
 
   /**
    * getter for CompactionProgress object
    * @return CompactionProgress object; can be null
    */
-  public CompactionProgress getCompactionProgress();
+  CompactionProgress getCompactionProgress();
 
-  public CompactionContext requestCompaction() throws IOException;
+  CompactionContext requestCompaction() throws IOException;
 
-  public CompactionContext requestCompaction(int priority, CompactionRequest baseRequest)
+  CompactionContext requestCompaction(int priority, CompactionRequest baseRequest)
       throws IOException;
 
-  public void cancelRequestedCompaction(CompactionContext compaction);
+  void cancelRequestedCompaction(CompactionContext compaction);
 
-  public List<StoreFile> compact(CompactionContext compaction) throws IOException;
+  List<StoreFile> compact(CompactionContext compaction) throws IOException;
 
   /**
    * @return true if we should run a major compaction.
    */
-  public boolean isMajorCompaction() throws IOException;
+  boolean isMajorCompaction() throws IOException;
 
-  public void triggerMajorCompaction();
+  void triggerMajorCompaction();
 
   /**
    * See if there's too much store files in this store
    * @return true if number of store files is greater than the number defined in minFilesToCompact
    */
-  public boolean needsCompaction();
+  boolean needsCompaction();
 
-  public int getCompactPriority();
+  int getCompactPriority();
 
-  public StoreFlushContext createFlushContext(long cacheFlushId);
+  StoreFlushContext createFlushContext(long cacheFlushId);
 
   /**
    * Call to complete a compaction. Its for the case where we find in the WAL a compaction
@@ -197,18 +205,18 @@ public interface Store extends HeapSize,
    * See HBASE-2331.
    * @param compaction
    */
-  public void completeCompactionMarker(CompactionDescriptor compaction)
+  void completeCompactionMarker(CompactionDescriptor compaction)
       throws IOException;
 
   // Split oriented methods
 
-  public boolean canSplit();
+  boolean canSplit();
 
   /**
    * Determines if Store should be split
    * @return byte[] if store should be split, null otherwise.
    */
-  public byte[] getSplitPoint();
+  byte[] getSplitPoint();
 
   // Bulk Load methods
 
@@ -216,7 +224,7 @@ public interface Store extends HeapSize,
    * This throws a WrongRegionException if the HFile does not fit in this region, or an
    * InvalidHFileException if the HFile is not valid.
    */
-  public void assertBulkLoadHFileOk(Path srcPath) throws IOException;
+  void assertBulkLoadHFileOk(Path srcPath) throws IOException;
 
   /**
    * This method should only be called from HRegion. It is assumed that the ranges of values in the
@@ -225,7 +233,7 @@ public interface Store extends HeapSize,
    * @param srcPathStr
    * @param sequenceId sequence Id associated with the HFile
    */
-  public void bulkLoadHFile(String srcPathStr, long sequenceId) throws IOException;
+  void bulkLoadHFile(String srcPathStr, long sequenceId) throws IOException;
 
   // General accessors into the state of the store
   // TODO abstract some of this out into a metrics class
@@ -233,50 +241,50 @@ public interface Store extends HeapSize,
   /**
    * @return <tt>true</tt> if the store has any underlying reference files to older HFiles
    */
-  public boolean hasReferences();
+  boolean hasReferences();
 
   /**
    * @return The size of this store's memstore, in bytes
    */
-  public long getMemStoreSize();
+  long getMemStoreSize();
 
-  public HColumnDescriptor getFamily();
+  HColumnDescriptor getFamily();
 
   /**
    * @return The maximum memstoreTS in all store files.
    */
-  public long getMaxMemstoreTS();
+  long getMaxMemstoreTS();
 
   /**
    * @return the data block encoder
    */
-  public HFileDataBlockEncoder getDataBlockEncoder();
+  HFileDataBlockEncoder getDataBlockEncoder();
 
   /** @return aggregate size of all HStores used in the last compaction */
-  public long getLastCompactSize();
+  long getLastCompactSize();
 
   /** @return aggregate size of HStore */
-  public long getSize();
+  long getSize();
 
   /**
    * @return Count of store files
    */
-  public int getStorefilesCount();
+  int getStorefilesCount();
 
   /**
    * @return The size of the store files, in bytes, uncompressed.
    */
-  public long getStoreSizeUncompressed();
+  long getStoreSizeUncompressed();
 
   /**
    * @return The size of the store files, in bytes.
    */
-  public long getStorefilesSize();
+  long getStorefilesSize();
 
   /**
    * @return The size of the store file indexes, in bytes.
    */
-  public long getStorefilesIndexSize();
+  long getStorefilesIndexSize();
 
   /**
    * Returns the total size of all index blocks in the data block indexes, including the root level,
@@ -284,14 +292,14 @@ public interface Store extends HeapSize,
    * single-level indexes.
    * @return the total size of block indexes in the store
    */
-  public long getTotalStaticIndexSize();
+  long getTotalStaticIndexSize();
 
   /**
    * Returns the total byte size of all Bloom filter bit arrays. For compound Bloom filters even the
    * Bloom blocks currently not loaded into the block cache are counted.
    * @return the total size of all Bloom filters in the store
    */
-  public long getTotalStaticBloomSize();
+  long getTotalStaticBloomSize();
 
   // Test-helper methods
 
@@ -299,40 +307,40 @@ public interface Store extends HeapSize,
    * Used for tests.
    * @return cache configuration for this Store.
    */
-  public CacheConfig getCacheConfig();
+  CacheConfig getCacheConfig();
 
   /**
    * @return the parent region info hosting this store
    */
-  public HRegionInfo getRegionInfo();
+  HRegionInfo getRegionInfo();
 
-  public RegionCoprocessorHost getCoprocessorHost();
+  RegionCoprocessorHost getCoprocessorHost();
 
-  public boolean areWritesEnabled();
+  boolean areWritesEnabled();
 
   /**
    * @return The smallest mvcc readPoint across all the scanners in this
    * region. Writes older than this readPoint, are included  in every
    * read operation.
    */
-  public long getSmallestReadPoint();
+  long getSmallestReadPoint();
 
-  public String getColumnFamilyName();
+  String getColumnFamilyName();
 
-  public String getTableName();
+  String getTableName();
 
   /*
    * @param o Observer who wants to know about changes in set of Readers
    */
-  public void addChangedReaderObserver(ChangedReadersObserver o);
+  void addChangedReaderObserver(ChangedReadersObserver o);
 
   /*
    * @param o Observer no longer interested in changes in set of Readers.
    */
-  public void deleteChangedReaderObserver(ChangedReadersObserver o);
+  void deleteChangedReaderObserver(ChangedReadersObserver o);
 
   /**
    * @return Whether this store has too many store files.
    */
-  public boolean hasTooManyStoreFiles();
+  boolean hasTooManyStoreFiles();
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreConfigInformation.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreConfigInformation.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreConfigInformation.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreConfigInformation.java Wed Jul 10 17:53:33 2013
@@ -34,22 +34,22 @@ public interface StoreConfigInformation 
    * TODO: remove after HBASE-7252 is fixed.
    * @return Gets the Memstore flush size for the region that this store works with.
    */
-  public long getMemstoreFlushSize();
+  long getMemstoreFlushSize();
 
   /**
    * @return Gets the cf-specific time-to-live for store files.
    */
-  public long getStoreFileTtl();
+  long getStoreFileTtl();
 
   /**
    * @return Gets the cf-specific compaction check frequency multiplier.
    *         The need for compaction (outside of normal checks during flush, open, etc.) will
    *         be ascertained every multiplier * HConstants.THREAD_WAKE_FREQUENCY milliseconds.
    */
-  public long getCompactionCheckMultiplier();
+  long getCompactionCheckMultiplier();
 
   /**
    * The number of files required before flushes for this store will be blocked.
    */
-  public long getBlockingFileCount();
+  long getBlockingFileCount();
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFileManager.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFileManager.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFileManager.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFileManager.java Wed Jul 10 17:53:33 2013
@@ -44,40 +44,41 @@ public interface StoreFileManager {
    * Loads the initial store files into empty StoreFileManager.
    * @param storeFiles The files to load.
    */
-  public abstract void loadFiles(List<StoreFile> storeFiles);
+  void loadFiles(List<StoreFile> storeFiles);
 
   /**
    * Adds new file, either for from MemStore flush or bulk insert, into the structure.
    * @param sf New store file.
    */
-  public abstract void insertNewFile(StoreFile sf);
+  void insertNewFile(StoreFile sf);
 
   /**
    * Adds compaction results into the structure.
    * @param compactedFiles The input files for the compaction.
    * @param results The resulting files for the compaction.
    */
-  public abstract void addCompactionResults(
-      Collection<StoreFile> compactedFiles, Collection<StoreFile> results);
+  void addCompactionResults(
+    Collection<StoreFile> compactedFiles, Collection<StoreFile> results
+  );
 
   /**
    * Clears all the files currently in use and returns them.
    * @return The files previously in use.
    */
-  public abstract ImmutableCollection<StoreFile> clearFiles();
+  ImmutableCollection<StoreFile> clearFiles();
 
   /**
    * Gets the snapshot of the store files currently in use. Can be used for things like metrics
    * and checks; should not assume anything about relations between store files in the list.
    * @return The list of StoreFiles.
    */
-  public abstract Collection<StoreFile> getStorefiles();
+  Collection<StoreFile> getStorefiles();
 
   /**
    * Returns the number of files currently in use.
    * @return The number of files.
    */
-  public abstract int getStorefileCount();
+  int getStorefileCount();
 
   /**
    * Gets the store files to scan for a Scan or Get request.
@@ -86,8 +87,9 @@ public interface StoreFileManager {
    * @param stopRow Stop row of the request.
    * @return The list of files that are to be read for this request.
    */
-  public abstract Collection<StoreFile> getFilesForScanOrGet(boolean isGet,
-      byte[] startRow, byte[] stopRow);
+  Collection<StoreFile> getFilesForScanOrGet(
+    boolean isGet, byte[] startRow, byte[] stopRow
+  );
 
   /**
    * Gets initial, full list of candidate store files to check for row-key-before.
@@ -95,8 +97,9 @@ public interface StoreFileManager {
    * @return The files that may have the key less than or equal to targetKey, in reverse
    *         order of new-ness, and preference for target key.
    */
-  public abstract Iterator<StoreFile> getCandidateFilesForRowKeyBefore(
-      KeyValue targetKey);
+  Iterator<StoreFile> getCandidateFilesForRowKeyBefore(
+    KeyValue targetKey
+  );
 
   /**
    * Updates the candidate list for finding row key before. Based on the list of candidates
@@ -109,8 +112,9 @@ public interface StoreFileManager {
    * @param candidate The current best candidate found.
    * @return The list to replace candidateFiles.
    */
-  public abstract Iterator<StoreFile> updateCandidateFilesForRowKeyBefore(
-      Iterator<StoreFile> candidateFiles, KeyValue targetKey, KeyValue candidate);
+  Iterator<StoreFile> updateCandidateFilesForRowKeyBefore(
+    Iterator<StoreFile> candidateFiles, KeyValue targetKey, KeyValue candidate
+  );
 
 
   /**
@@ -118,10 +122,10 @@ public interface StoreFileManager {
    * @return The mid-point, or null if no split is possible.
    * @throws IOException
    */
-  public abstract byte[] getSplitPoint() throws IOException;
+  byte[] getSplitPoint() throws IOException;
 
   /**
    * @return The store compaction priority.
    */
-  public abstract int getStoreCompactionPriority();
+  int getStoreCompactionPriority();
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/Dictionary.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/Dictionary.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/Dictionary.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/Dictionary.java Wed Jul 10 17:53:33 2013
@@ -28,7 +28,7 @@ import org.apache.hadoop.classification.
  */
 @InterfaceAudience.Private
 interface Dictionary {
-  static final byte NOT_IN_DICTIONARY = -1;
+  byte NOT_IN_DICTIONARY = -1;
 
   /**
    * Gets an entry from the dictionary.
@@ -36,7 +36,7 @@ interface Dictionary {
    * @param idx index of the entry
    * @return the entry, or null if non existent
    */
-  public byte[] getEntry(short idx);
+  byte[] getEntry(short idx);
 
   /**
    * Finds the index of an entry.
@@ -47,7 +47,7 @@ interface Dictionary {
    * @param length Length beyond <code>offset</code> that comprises entry; must be > 0.
    * @return the index of the entry, or {@link #NOT_IN_DICTIONARY} if not found
    */
-  public short findEntry(byte[] data, int offset, int length);
+  short findEntry(byte[] data, int offset, int length);
 
   /**
    * Adds an entry to the dictionary.
@@ -62,10 +62,10 @@ interface Dictionary {
    * @return the index of the entry
    */
 
-  public short addEntry(byte[] data, int offset, int length);
+  short addEntry(byte[] data, int offset, int length);
 
   /**
    * Flushes the dictionary, empties all values.
    */
-  public void clear();
+  void clear();
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java Wed Jul 10 17:53:33 2013
@@ -41,26 +41,26 @@ import org.apache.hadoop.io.Writable;
 
 @InterfaceAudience.Private
 public interface HLog {
-  public static final Log LOG = LogFactory.getLog(HLog.class);
+  Log LOG = LogFactory.getLog(HLog.class);
 
   /** File Extension used while splitting an HLog into regions (HBASE-2312) */
-  public static final String SPLITTING_EXT = "-splitting";
-  public static final boolean SPLIT_SKIP_ERRORS_DEFAULT = false;
+  String SPLITTING_EXT = "-splitting";
+  boolean SPLIT_SKIP_ERRORS_DEFAULT = false;
   /** The META region's HLog filename extension */
-  public static final String META_HLOG_FILE_EXTN = ".meta";
+  String META_HLOG_FILE_EXTN = ".meta";
 
   /**
    * Configuration name of HLog Trailer's warning size. If a waltrailer's size is greater than the
    * configured size, a warning is logged. This is used with Protobuf reader/writer.
    */
-  public static final String WAL_TRAILER_WARN_SIZE =
+  String WAL_TRAILER_WARN_SIZE =
     "hbase.regionserver.waltrailer.warn.size";
-  public static final int DEFAULT_WAL_TRAILER_WARN_SIZE = 1024*1024; // 1MB
+  int DEFAULT_WAL_TRAILER_WARN_SIZE = 1024*1024; // 1MB
 
-  static final Pattern EDITFILES_NAME_PATTERN = Pattern.compile("-?[0-9]+");
-  public static final String RECOVERED_LOG_TMPFILE_SUFFIX = ".temp";
+  Pattern EDITFILES_NAME_PATTERN = Pattern.compile("-?[0-9]+");
+  String RECOVERED_LOG_TMPFILE_SUFFIX = ".temp";
 
-  public interface Reader {
+  interface Reader {
 
     /**
      * @param fs File system.
@@ -88,7 +88,7 @@ public interface HLog {
     WALTrailer getWALTrailer();
   }
 
-  public interface Writer {
+  interface Writer {
     void init(FileSystem fs, Path path, Configuration c) throws IOException;
 
     void close() throws IOException;
@@ -110,7 +110,7 @@ public interface HLog {
    * Utility class that lets us keep track of the edit with it's key Only used
    * when splitting logs
    */
-  public static class Entry implements Writable {
+  class Entry implements Writable {
     private WALEdit edit;
     private HLogKey key;
 
@@ -185,19 +185,19 @@ public interface HLog {
    *
    * @param listener
    */
-  public void registerWALActionsListener(final WALActionsListener listener);
+  void registerWALActionsListener(final WALActionsListener listener);
 
   /**
    * unregisters WALActionsListener
    *
    * @param listener
    */
-  public boolean unregisterWALActionsListener(final WALActionsListener listener);
+  boolean unregisterWALActionsListener(final WALActionsListener listener);
 
   /**
    * @return Current state of the monotonically increasing file id.
    */
-  public long getFilenum();
+  long getFilenum();
 
   /**
    * Called by HRegionServer when it opens a new region to ensure that log
@@ -208,12 +208,12 @@ public interface HLog {
    *          We'll set log edit/sequence number to this value if it is greater
    *          than the current value.
    */
-  public void setSequenceNumber(final long newvalue);
+  void setSequenceNumber(final long newvalue);
 
   /**
    * @return log sequence number
    */
-  public long getSequenceNumber();
+  long getSequenceNumber();
 
   /**
    * Roll the log writer. That is, start writing log messages to a new file.
@@ -228,7 +228,7 @@ public interface HLog {
    * @throws org.apache.hadoop.hbase.exceptions.FailedLogCloseException
    * @throws IOException
    */
-  public byte[][] rollWriter() throws FailedLogCloseException, IOException;
+  byte[][] rollWriter() throws FailedLogCloseException, IOException;
 
   /**
    * Roll the log writer. That is, start writing log messages to a new file.
@@ -246,7 +246,7 @@ public interface HLog {
    * @throws org.apache.hadoop.hbase.exceptions.FailedLogCloseException
    * @throws IOException
    */
-  public byte[][] rollWriter(boolean force) throws FailedLogCloseException,
+  byte[][] rollWriter(boolean force) throws FailedLogCloseException,
       IOException;
 
   /**
@@ -254,21 +254,22 @@ public interface HLog {
    *
    * @throws IOException
    */
-  public void close() throws IOException;
+  void close() throws IOException;
 
   /**
    * Shut down the log and delete the log directory
    *
    * @throws IOException
    */
-  public void closeAndDelete() throws IOException;
+  void closeAndDelete() throws IOException;
 
   /**
    * Same as {@link #appendNoSync(HRegionInfo, byte[], WALEdit, UUID, long, HTableDescriptor)},
    * except it causes a sync on the log
    */
-  public void append(HRegionInfo info, byte[] tableName, WALEdit edits,
-      final long now, HTableDescriptor htd) throws IOException;
+  void append(
+    HRegionInfo info, byte[] tableName, WALEdit edits, final long now, HTableDescriptor htd
+  ) throws IOException;
 
   /**
    * Append a set of edits to the log. Log edits are keyed by (encoded)
@@ -281,8 +282,14 @@ public interface HLog {
    * @param htd
    * @param isInMemstore Whether the record is in memstore. False for system records.
    */
-  public void append(HRegionInfo info, byte[] tableName, WALEdit edits,
-      final long now, HTableDescriptor htd, boolean isInMemstore) throws IOException;
+  void append(
+    HRegionInfo info,
+    byte[] tableName,
+    WALEdit edits,
+    final long now,
+    HTableDescriptor htd,
+    boolean isInMemstore
+  ) throws IOException;
 
   /**
    * Append a set of edits to the log. Log edits are keyed by (encoded)
@@ -299,21 +306,27 @@ public interface HLog {
    * @return txid of this transaction
    * @throws IOException
    */
-  public long appendNoSync(HRegionInfo info, byte[] tableName, WALEdit edits,
-      UUID clusterId, final long now, HTableDescriptor htd) throws IOException;
+  long appendNoSync(
+    HRegionInfo info,
+    byte[] tableName,
+    WALEdit edits,
+    UUID clusterId,
+    final long now,
+    HTableDescriptor htd
+  ) throws IOException;
 
-  public void hsync() throws IOException;
+  void hsync() throws IOException;
 
-  public void hflush() throws IOException;
+  void hflush() throws IOException;
 
-  public void sync() throws IOException;
+  void sync() throws IOException;
 
-  public void sync(long txid) throws IOException;
+  void sync(long txid) throws IOException;
 
   /**
    * Obtain a log sequence number.
    */
-  public long obtainSeqNum();
+  long obtainSeqNum();
 
   /**
    * WAL keeps track of the sequence numbers that were not yet flushed from memstores
@@ -330,13 +343,13 @@ public interface HLog {
    *         the resulting file as an upper-bound seqNum for that file), or NULL if flush
    *         should not be started.
    */
-  public Long startCacheFlush(final byte[] encodedRegionName);
+  Long startCacheFlush(final byte[] encodedRegionName);
 
   /**
    * Complete the cache flush.
    * @param encodedRegionName Encoded region name.
    */
-  public void completeCacheFlush(final byte[] encodedRegionName);
+  void completeCacheFlush(final byte[] encodedRegionName);
 
   /**
    * Abort a cache flush. Call if the flush fails. Note that the only recovery
@@ -344,24 +357,24 @@ public interface HLog {
    * snapshot content dropped by the failure gets restored to the memstore.v
    * @param encodedRegionName Encoded region name.
    */
-  public void abortCacheFlush(byte[] encodedRegionName);
+  void abortCacheFlush(byte[] encodedRegionName);
 
   /**
    * @return Coprocessor host.
    */
-  public WALCoprocessorHost getCoprocessorHost();
+  WALCoprocessorHost getCoprocessorHost();
 
   /**
    * Get LowReplication-Roller status
    *
    * @return lowReplicationRollEnabled
    */
-  public boolean isLowReplicationRollEnabled();
+  boolean isLowReplicationRollEnabled();
 
   /** Gets the earliest sequence number in the memstore for this particular region.
    * This can serve as best-effort "recent" WAL number for this region.
    * @param encodedRegionName The region to get the number for.
    * @return The number if present, HConstants.NO_SEQNUM if absent.
    */
-  public long getEarliestMemstoreSeqNum(byte[] encodedRegionName);
+  long getEarliestMemstoreSeqNum(byte[] encodedRegionName);
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALActionsListener.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALActionsListener.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALActionsListener.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALActionsListener.java Wed Jul 10 17:53:33 2013
@@ -38,7 +38,7 @@ public interface WALActionsListener {
    * @param oldPath the path to the old hlog
    * @param newPath the path to the new hlog
    */
-  public void preLogRoll(Path oldPath, Path newPath) throws IOException;
+  void preLogRoll(Path oldPath, Path newPath) throws IOException;
 
   /**
    * The WAL has been rolled. The oldPath can be null if this is
@@ -46,31 +46,31 @@ public interface WALActionsListener {
    * @param oldPath the path to the old hlog
    * @param newPath the path to the new hlog
    */
-  public void postLogRoll(Path oldPath, Path newPath) throws IOException;
+  void postLogRoll(Path oldPath, Path newPath) throws IOException;
 
   /**
    * The WAL is going to be archived.
    * @param oldPath the path to the old hlog
    * @param newPath the path to the new hlog
    */
-  public void preLogArchive(Path oldPath, Path newPath) throws IOException;
+  void preLogArchive(Path oldPath, Path newPath) throws IOException;
 
   /**
    * The WAL has been archived.
    * @param oldPath the path to the old hlog
    * @param newPath the path to the new hlog
    */
-  public void postLogArchive(Path oldPath, Path newPath) throws IOException;
+  void postLogArchive(Path oldPath, Path newPath) throws IOException;
 
   /**
    * A request was made that the WAL be rolled.
    */
-  public void logRollRequested();
+  void logRollRequested();
 
   /**
    * The WAL is about to close.
    */
-  public void logCloseRequested();
+  void logCloseRequested();
 
   /**
   * Called before each write.
@@ -78,8 +78,9 @@ public interface WALActionsListener {
   * @param logKey
   * @param logEdit
   */
- public void visitLogEntryBeforeWrite(HRegionInfo info, HLogKey logKey,
-   WALEdit logEdit);
+  void visitLogEntryBeforeWrite(
+    HRegionInfo info, HLogKey logKey, WALEdit logEdit
+  );
 
   /**
    *
@@ -87,7 +88,8 @@ public interface WALActionsListener {
    * @param logKey
    * @param logEdit
    */
- public void visitLogEntryBeforeWrite(HTableDescriptor htd, HLogKey logKey,
-   WALEdit logEdit);
+  void visitLogEntryBeforeWrite(
+    HTableDescriptor htd, HLogKey logKey, WALEdit logEdit
+  );
 
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceInterface.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceInterface.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceInterface.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceInterface.java Wed Jul 10 17:53:33 2013
@@ -41,61 +41,63 @@ public interface ReplicationSourceInterf
    * @param peerClusterId the id of the peer cluster
    * @throws IOException
    */
-  public void init(final Configuration conf,
-                   final FileSystem fs,
-                   final ReplicationSourceManager manager,
-                   final Stoppable stopper,
-                   final String peerClusterId) throws IOException;
+  void init(
+    final Configuration conf,
+    final FileSystem fs,
+    final ReplicationSourceManager manager,
+    final Stoppable stopper,
+    final String peerClusterId
+  ) throws IOException;
 
   /**
    * Add a log to the list of logs to replicate
    * @param log path to the log to replicate
    */
-  public void enqueueLog(Path log);
+  void enqueueLog(Path log);
 
   /**
    * Get the current log that's replicated
    * @return the current log
    */
-  public Path getCurrentPath();
+  Path getCurrentPath();
 
   /**
    * Start the replication
    */
-  public void startup();
+  void startup();
 
   /**
    * End the replication
    * @param reason why it's terminating
    */
-  public void terminate(String reason);
+  void terminate(String reason);
 
   /**
    * End the replication
    * @param reason why it's terminating
    * @param cause the error that's causing it
    */
-  public void terminate(String reason, Exception cause);
+  void terminate(String reason, Exception cause);
 
   /**
    * Get the id that the source is replicating to
    *
    * @return peer cluster id
    */
-  public String getPeerClusterZnode();
+  String getPeerClusterZnode();
 
   /**
    * Get the id that the source is replicating to.
    *
    * @return peer cluster id
    */
-  public String getPeerClusterId();
+  String getPeerClusterId();
 
   /**
    * Get a string representation of the current statistics
    * for this source
    * @return printable stats
    */
-  public String getStats();
+  String getStats();
 
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/Constants.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/Constants.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/Constants.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/Constants.java Wed Jul 10 17:53:33 2013
@@ -28,19 +28,19 @@ import org.apache.hadoop.classification.
 @InterfaceAudience.Public
 @InterfaceStability.Stable
 public interface Constants {
-  public static final String VERSION_STRING = "0.0.2";
+  String VERSION_STRING = "0.0.2";
 
-  public static final int DEFAULT_MAX_AGE = 60 * 60 * 4;  // 4 hours
+  int DEFAULT_MAX_AGE = 60 * 60 * 4;  // 4 hours
 
-  public static final int DEFAULT_LISTEN_PORT = 8080;
+  int DEFAULT_LISTEN_PORT = 8080;
 
-  public static final String MIMETYPE_TEXT = "text/plain";
-  public static final String MIMETYPE_HTML = "text/html";
-  public static final String MIMETYPE_XML = "text/xml";
-  public static final String MIMETYPE_BINARY = "application/octet-stream";
-  public static final String MIMETYPE_PROTOBUF = "application/x-protobuf";
-  public static final String MIMETYPE_PROTOBUF_IETF = "application/protobuf";
-  public static final String MIMETYPE_JSON = "application/json";
+  String MIMETYPE_TEXT = "text/plain";
+  String MIMETYPE_HTML = "text/html";
+  String MIMETYPE_XML = "text/xml";
+  String MIMETYPE_BINARY = "application/octet-stream";
+  String MIMETYPE_PROTOBUF = "application/x-protobuf";
+  String MIMETYPE_PROTOBUF_IETF = "application/protobuf";
+  String MIMETYPE_JSON = "application/json";
 
-  public static final String CRLF = "\r\n";
+  String CRLF = "\r\n";
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/ProtobufMessageHandler.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/ProtobufMessageHandler.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/ProtobufMessageHandler.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/ProtobufMessageHandler.java Wed Jul 10 17:53:33 2013
@@ -29,11 +29,11 @@ import org.apache.hadoop.classification.
  * ProtobufMessageBodyProducer adapters. 
  */
 @InterfaceAudience.Private
-public abstract interface ProtobufMessageHandler {
+public interface ProtobufMessageHandler {
   /**
    * @return the protobuf represention of the model
    */
-  public byte[] createProtobufOutput();
+  byte[] createProtobufOutput();
 
   /**
    * Initialize the model from a protobuf representation.
@@ -41,6 +41,6 @@ public abstract interface ProtobufMessag
    * @return reference to self for convenience
    * @throws IOException
    */
-  public ProtobufMessageHandler getObjectFromMessage(byte[] message)
+  ProtobufMessageHandler getObjectFromMessage(byte[] message)
     throws IOException;
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/CodeToClassAndBackFor96Migration.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/CodeToClassAndBackFor96Migration.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/CodeToClassAndBackFor96Migration.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/CodeToClassAndBackFor96Migration.java Wed Jul 10 17:53:33 2013
@@ -39,13 +39,13 @@ interface CodeToClassAndBackFor96Migrati
   /**
    * Static map that contains mapping from code to class
    */
-  static final Map<Byte, Class<?>> CODE_TO_CLASS =
+  Map<Byte, Class<?>> CODE_TO_CLASS =
     new HashMap<Byte, Class<?>>();
 
   /**
    * Static map that contains mapping from class to code
    */
-  static final Map<Class<?>, Byte> CLASS_TO_CODE =
+  Map<Class<?>, Byte> CLASS_TO_CODE =
     new HashMap<Class<?>, Byte>();
 
   /**

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/thrift/IncrementCoalescerMBean.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/thrift/IncrementCoalescerMBean.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/thrift/IncrementCoalescerMBean.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/thrift/IncrementCoalescerMBean.java Wed Jul 10 17:53:33 2013
@@ -19,31 +19,31 @@
 package org.apache.hadoop.hbase.thrift;
 
 public interface IncrementCoalescerMBean {
-  public int getQueueSize();
+  int getQueueSize();
 
-  public int getMaxQueueSize();
+  int getMaxQueueSize();
 
-  public void setMaxQueueSize(int newSize);
+  void setMaxQueueSize(int newSize);
 
-  public long getPoolCompletedTaskCount();
+  long getPoolCompletedTaskCount();
 
-  public long getPoolTaskCount();
+  long getPoolTaskCount();
 
-  public int getPoolLargestPoolSize();
+  int getPoolLargestPoolSize();
 
-  public int getCorePoolSize();
+  int getCorePoolSize();
 
-  public void setCorePoolSize(int newCoreSize);
+  void setCorePoolSize(int newCoreSize);
 
-  public int getMaxPoolSize();
+  int getMaxPoolSize();
 
-  public void setMaxPoolSize(int newMaxSize);
+  void setMaxPoolSize(int newMaxSize);
 
-  public long getFailedIncrements();
+  long getFailedIncrements();
 
-  public long getSuccessfulCoalescings();
+  long getSuccessfulCoalescings();
 
-  public long getTotalIncrements();
+  long getTotalIncrements();
 
-  public long getCountersMapSize();
+  long getCountersMapSize();
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java Wed Jul 10 17:53:33 2013
@@ -47,9 +47,9 @@ import org.apache.hadoop.hbase.client.HB
 public final class Canary implements Tool {
   // Sink interface used by the canary to outputs information
   public interface Sink {
-    public void publishReadFailure(HRegionInfo region);
-    public void publishReadFailure(HRegionInfo region, HColumnDescriptor column);
-    public void publishReadTiming(HRegionInfo region, HColumnDescriptor column, long msTime);
+    void publishReadFailure(HRegionInfo region);
+    void publishReadFailure(HRegionInfo region, HColumnDescriptor column);
+    void publishReadTiming(HRegionInfo region, HColumnDescriptor column, long msTime);
   }
 
   // Simple implementation of canary sink that allows to plot on

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/CancelableProgressable.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/CancelableProgressable.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/CancelableProgressable.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/CancelableProgressable.java Wed Jul 10 17:53:33 2013
@@ -34,6 +34,6 @@ public interface CancelableProgressable 
    * operation should be canceled and rolled back.
    * @return whether to continue (true) or cancel (false) the operation
    */
-  public boolean progress();
+  boolean progress();
 
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java Wed Jul 10 17:53:33 2013
@@ -2903,7 +2903,7 @@ public class HBaseFsck extends Configure
   }
 
   public interface ErrorReporter {
-    public static enum ERROR_CODE {
+    enum ERROR_CODE {
       UNKNOWN, NO_META_REGION, NULL_META_REGION, NO_VERSION_FILE, NOT_IN_META_HDFS, NOT_IN_META,
       NOT_IN_META_OR_DEPLOYED, NOT_IN_HDFS_OR_DEPLOYED, NOT_IN_HDFS, SERVER_DOES_NOT_MATCH_META, NOT_DEPLOYED,
       MULTI_DEPLOYED, SHOULD_NOT_BE_DEPLOYED, MULTI_META_REGION, RS_CONNECT_FAILURE,
@@ -2912,20 +2912,26 @@ public class HBaseFsck extends Configure
       ORPHAN_HDFS_REGION, LINGERING_SPLIT_PARENT, NO_TABLEINFO_FILE, LINGERING_REFERENCE_HFILE,
       WRONG_USAGE, EMPTY_META_CELL, EXPIRED_TABLE_LOCK
     }
-    public void clear();
-    public void report(String message);
-    public void reportError(String message);
-    public void reportError(ERROR_CODE errorCode, String message);
-    public void reportError(ERROR_CODE errorCode, String message, TableInfo table);
-    public void reportError(ERROR_CODE errorCode, String message, TableInfo table, HbckInfo info);
-    public void reportError(ERROR_CODE errorCode, String message, TableInfo table, HbckInfo info1, HbckInfo info2);
-    public int summarize();
-    public void detail(String details);
-    public ArrayList<ERROR_CODE> getErrorList();
-    public void progress();
-    public void print(String message);
-    public void resetErrors();
-    public boolean tableHasErrors(TableInfo table);
+    void clear();
+    void report(String message);
+    void reportError(String message);
+    void reportError(ERROR_CODE errorCode, String message);
+    void reportError(ERROR_CODE errorCode, String message, TableInfo table);
+    void reportError(ERROR_CODE errorCode, String message, TableInfo table, HbckInfo info);
+    void reportError(
+      ERROR_CODE errorCode,
+      String message,
+      TableInfo table,
+      HbckInfo info1,
+      HbckInfo info2
+    );
+    int summarize();
+    void detail(String details);
+    ArrayList<ERROR_CODE> getErrorList();
+    void progress();
+    void print(String message);
+    void resetErrors();
+    boolean tableHasErrors(TableInfo table);
   }
 
   static class PrintingErrorReporter implements ErrorReporter {

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/KeyRange.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/KeyRange.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/KeyRange.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/KeyRange.java Wed Jul 10 17:53:33 2013
@@ -25,7 +25,7 @@ import org.apache.hadoop.classification.
  */
 @InterfaceAudience.Private
 public interface KeyRange {
-  abstract byte[] getStartKey();
+  byte[] getStartKey();
 
-  abstract byte[] getEndKey();
+  byte[] getEndKey();
 }

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ModifyRegionUtils.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ModifyRegionUtils.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ModifyRegionUtils.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ModifyRegionUtils.java Wed Jul 10 17:53:33 2013
@@ -60,7 +60,7 @@ public abstract class ModifyRegionUtils 
   }
 
   public interface RegionFillTask {
-    public void fillRegion(final HRegion region) throws IOException;
+    void fillRegion(final HRegion region) throws IOException;
   }
 
   /**

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/RegionSplitter.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/RegionSplitter.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/RegionSplitter.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/util/RegionSplitter.java Wed Jul 10 17:53:33 2013
@@ -147,7 +147,7 @@ public class RegionSplitter {
    * {@link RegionSplitter#rollingSplit(String, SplitAlgorithm, Configuration)} with the
    * argument splitClassName giving the name of your class.
    */
-  public static interface SplitAlgorithm {
+  public interface SplitAlgorithm {
     /**
      * Split a pre-existing region into 2 regions.
      *
@@ -1046,4 +1046,4 @@ public class RegionSplitter {
           + "," + rowToStr(lastRow()) + "]";
     }
   }
-}
\ No newline at end of file
+}

Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestCase.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestCase.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestCase.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestCase.java Wed Jul 10 17:53:33 2013
@@ -382,11 +382,11 @@ public abstract class HBaseTestCase exte
   /**
    * Implementors can flushcache.
    */
-  public static interface FlushCache {
+  public interface FlushCache {
     /**
      * @throws IOException
      */
-    public void flushcache() throws IOException;
+    void flushcache() throws IOException;
   }
 
   /**
@@ -395,23 +395,23 @@ public abstract class HBaseTestCase exte
    *
    * TOOD: Come up w/ a better name for this interface.
    */
-  public static interface Incommon {
+  public interface Incommon {
     /**
      *
      * @param delete
      * @param writeToWAL
      * @throws IOException
      */
-    public void delete(Delete delete,  boolean writeToWAL)
+    void delete(Delete delete, boolean writeToWAL)
     throws IOException;
 
     /**
      * @param put
      * @throws IOException
      */
-    public void put(Put put) throws IOException;
+    void put(Put put) throws IOException;
 
-    public Result get(Get get) throws IOException;
+    Result get(Get get) throws IOException;
 
     /**
      * @param family
@@ -421,8 +421,9 @@ public abstract class HBaseTestCase exte
      * @return scanner for specified columns, first row and timestamp
      * @throws IOException
      */
-    public ScannerIncommon getScanner(byte [] family, byte [][] qualifiers,
-        byte [] firstRow, long ts)
+    ScannerIncommon getScanner(
+      byte[] family, byte[][] qualifiers, byte[] firstRow, long ts
+    )
     throws IOException;
   }
 
@@ -520,10 +521,10 @@ public abstract class HBaseTestCase exte
 
   public interface ScannerIncommon
   extends Iterable<Result> {
-    public boolean next(List<KeyValue> values)
+    boolean next(List<KeyValue> values)
     throws IOException;
 
-    public void close() throws IOException;
+    void close() throws IOException;
   }
 
   public static class ClientScannerIncommon implements ScannerIncommon {

Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java?rev=1501880&r1=1501879&r2=1501880&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java Wed Jul 10 17:53:33 2013
@@ -198,7 +198,7 @@ public class PerformanceEvaluation exten
   /**
    * Implementations can have their status set.
    */
-  static interface Status {
+  interface Status {
     /**
      * Sets status
      * @param msg status message