You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2013/07/10 20:50:23 UTC
svn commit: r1501909 [2/4] - in /hbase/trunk:
hbase-client/src/main/java/org/apache/hadoop/hbase/
hbase-client/src/main/java/org/apache/hadoop/hbase/catalog/
hbase-client/src/main/java/org/apache/hadoop/hbase/client/
hbase-client/src/main/java/org/apac...
Modified: hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionServerSource.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionServerSource.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionServerSource.java (original)
+++ hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionServerSource.java Wed Jul 10 18:50:21 2013
@@ -28,22 +28,22 @@ public interface MetricsRegionServerSour
/**
* The name of the metrics
*/
- String METRICS_NAME = "Server";
+ static final String METRICS_NAME = "Server";
/**
* The name of the metrics context that metrics will be under.
*/
- String METRICS_CONTEXT = "regionserver";
+ static final String METRICS_CONTEXT = "regionserver";
/**
* Description
*/
- String METRICS_DESCRIPTION = "Metrics about HBase RegionServer";
+ static final String METRICS_DESCRIPTION = "Metrics about HBase RegionServer";
/**
* The name of the metrics context that metrics will be under in jmx
*/
- String METRICS_JMX_CONTEXT = "RegionServer,sub=" + METRICS_NAME;
+ static final String METRICS_JMX_CONTEXT = "RegionServer,sub=" + METRICS_NAME;
/**
* Update the Put time histogram
@@ -113,103 +113,103 @@ public interface MetricsRegionServerSour
void incrSlowAppend();
// Strings used for exporting to metrics system.
- String REGION_COUNT = "regionCount";
- String REGION_COUNT_DESC = "Number of regions";
- String STORE_COUNT = "storeCount";
- String STORE_COUNT_DESC = "Number of Stores";
- String STOREFILE_COUNT = "storeFileCount";
- String STOREFILE_COUNT_DESC = "Number of Store Files";
- String MEMSTORE_SIZE = "memStoreSize";
- String MEMSTORE_SIZE_DESC = "Size of the memstore";
- String STOREFILE_SIZE = "storeFileSize";
- String STOREFILE_SIZE_DESC = "Size of storefiles being served.";
- String TOTAL_REQUEST_COUNT = "totalRequestCount";
- String TOTAL_REQUEST_COUNT_DESC =
+ static final String REGION_COUNT = "regionCount";
+ static final String REGION_COUNT_DESC = "Number of regions";
+ static final String STORE_COUNT = "storeCount";
+ static final String STORE_COUNT_DESC = "Number of Stores";
+ static final String STOREFILE_COUNT = "storeFileCount";
+ static final String STOREFILE_COUNT_DESC = "Number of Store Files";
+ static final String MEMSTORE_SIZE = "memStoreSize";
+ static final String MEMSTORE_SIZE_DESC = "Size of the memstore";
+ static final String STOREFILE_SIZE = "storeFileSize";
+ static final String STOREFILE_SIZE_DESC = "Size of storefiles being served.";
+ static final String TOTAL_REQUEST_COUNT = "totalRequestCount";
+ static final String TOTAL_REQUEST_COUNT_DESC =
"Total number of requests this RegionServer has answered.";
- String READ_REQUEST_COUNT = "readRequestCount";
- String READ_REQUEST_COUNT_DESC =
+ static final String READ_REQUEST_COUNT = "readRequestCount";
+ static final String READ_REQUEST_COUNT_DESC =
"Number of read requests this region server has answered.";
- String WRITE_REQUEST_COUNT = "writeRequestCount";
- String WRITE_REQUEST_COUNT_DESC =
+ static final String WRITE_REQUEST_COUNT = "writeRequestCount";
+ static final String WRITE_REQUEST_COUNT_DESC =
"Number of mutation requests this region server has answered.";
- String CHECK_MUTATE_FAILED_COUNT = "checkMutateFailedCount";
- String CHECK_MUTATE_FAILED_COUNT_DESC =
+ static final String CHECK_MUTATE_FAILED_COUNT = "checkMutateFailedCount";
+ static final String CHECK_MUTATE_FAILED_COUNT_DESC =
"Number of Check and Mutate calls that failed the checks.";
- String CHECK_MUTATE_PASSED_COUNT = "checkMutatePassedCount";
- String CHECK_MUTATE_PASSED_COUNT_DESC =
+ static final String CHECK_MUTATE_PASSED_COUNT = "checkMutatePassedCount";
+ static final String CHECK_MUTATE_PASSED_COUNT_DESC =
"Number of Check and Mutate calls that passed the checks.";
- String STOREFILE_INDEX_SIZE = "storeFileIndexSize";
- String STOREFILE_INDEX_SIZE_DESC = "Size of indexes in storefiles on disk.";
- String STATIC_INDEX_SIZE = "staticIndexSize";
- String STATIC_INDEX_SIZE_DESC = "Uncompressed size of the static indexes.";
- String STATIC_BLOOM_SIZE = "staticBloomSize";
- String STATIC_BLOOM_SIZE_DESC =
+ static final String STOREFILE_INDEX_SIZE = "storeFileIndexSize";
+ static final String STOREFILE_INDEX_SIZE_DESC = "Size of indexes in storefiles on disk.";
+ static final String STATIC_INDEX_SIZE = "staticIndexSize";
+ static final String STATIC_INDEX_SIZE_DESC = "Uncompressed size of the static indexes.";
+ static final String STATIC_BLOOM_SIZE = "staticBloomSize";
+ static final String STATIC_BLOOM_SIZE_DESC =
"Uncompressed size of the static bloom filters.";
- String NUMBER_OF_MUTATIONS_WITHOUT_WAL = "mutationsWithoutWALCount";
- String NUMBER_OF_MUTATIONS_WITHOUT_WAL_DESC =
+ static final String NUMBER_OF_MUTATIONS_WITHOUT_WAL = "mutationsWithoutWALCount";
+ static final String NUMBER_OF_MUTATIONS_WITHOUT_WAL_DESC =
"Number of mutations that have been sent by clients with the write ahead logging turned off.";
- String DATA_SIZE_WITHOUT_WAL = "mutationsWithoutWALSize";
- String DATA_SIZE_WITHOUT_WAL_DESC =
+ static final String DATA_SIZE_WITHOUT_WAL = "mutationsWithoutWALSize";
+ static final String DATA_SIZE_WITHOUT_WAL_DESC =
"Size of data that has been sent by clients with the write ahead logging turned off.";
- String PERCENT_FILES_LOCAL = "percentFilesLocal";
- String PERCENT_FILES_LOCAL_DESC =
+ static final String PERCENT_FILES_LOCAL = "percentFilesLocal";
+ static final String PERCENT_FILES_LOCAL_DESC =
"The percent of HFiles that are stored on the local hdfs data node.";
- String COMPACTION_QUEUE_LENGTH = "compactionQueueLength";
- String COMPACTION_QUEUE_LENGTH_DESC = "Length of the queue for compactions.";
- String FLUSH_QUEUE_LENGTH = "flushQueueLength";
- String FLUSH_QUEUE_LENGTH_DESC = "Length of the queue for region flushes";
- String BLOCK_CACHE_FREE_SIZE = "blockCacheFreeSize";
- String BLOCK_CACHE_FREE_DESC =
+ static final String COMPACTION_QUEUE_LENGTH = "compactionQueueLength";
+ static final String COMPACTION_QUEUE_LENGTH_DESC = "Length of the queue for compactions.";
+ static final String FLUSH_QUEUE_LENGTH = "flushQueueLength";
+ static final String FLUSH_QUEUE_LENGTH_DESC = "Length of the queue for region flushes";
+ static final String BLOCK_CACHE_FREE_SIZE = "blockCacheFreeSize";
+ static final String BLOCK_CACHE_FREE_DESC =
"Size of the block cache that is not occupied.";
- String BLOCK_CACHE_COUNT = "blockCacheCount";
- String BLOCK_CACHE_COUNT_DESC = "Number of block in the block cache.";
- String BLOCK_CACHE_SIZE = "blockCacheSize";
- String BLOCK_CACHE_SIZE_DESC = "Size of the block cache.";
- String BLOCK_CACHE_HIT_COUNT = "blockCacheHitCount";
- String BLOCK_CACHE_HIT_COUNT_DESC = "Count of the hit on the block cache.";
- String BLOCK_CACHE_MISS_COUNT = "blockCacheMissCount";
- String BLOCK_COUNT_MISS_COUNT_DESC =
+ static final String BLOCK_CACHE_COUNT = "blockCacheCount";
+ static final String BLOCK_CACHE_COUNT_DESC = "Number of block in the block cache.";
+ static final String BLOCK_CACHE_SIZE = "blockCacheSize";
+ static final String BLOCK_CACHE_SIZE_DESC = "Size of the block cache.";
+ static final String BLOCK_CACHE_HIT_COUNT = "blockCacheHitCount";
+ static final String BLOCK_CACHE_HIT_COUNT_DESC = "Count of the hit on the block cache.";
+ static final String BLOCK_CACHE_MISS_COUNT = "blockCacheMissCount";
+ static final String BLOCK_COUNT_MISS_COUNT_DESC =
"Number of requests for a block that missed the block cache.";
- String BLOCK_CACHE_EVICTION_COUNT = "blockCacheEvictionCount";
- String BLOCK_CACHE_EVICTION_COUNT_DESC =
+ static final String BLOCK_CACHE_EVICTION_COUNT = "blockCacheEvictionCount";
+ static final String BLOCK_CACHE_EVICTION_COUNT_DESC =
"Count of the number of blocks evicted from the block cache.";
- String BLOCK_CACHE_HIT_PERCENT = "blockCountHitPercent";
- String BLOCK_CACHE_HIT_PERCENT_DESC =
+ static final String BLOCK_CACHE_HIT_PERCENT = "blockCountHitPercent";
+ static final String BLOCK_CACHE_HIT_PERCENT_DESC =
"Percent of block cache requests that are hits";
- String BLOCK_CACHE_EXPRESS_HIT_PERCENT = "blockCacheExpressHitPercent";
- String BLOCK_CACHE_EXPRESS_HIT_PERCENT_DESC =
+ static final String BLOCK_CACHE_EXPRESS_HIT_PERCENT = "blockCacheExpressHitPercent";
+ static final String BLOCK_CACHE_EXPRESS_HIT_PERCENT_DESC =
"The percent of the time that requests with the cache turned on hit the cache.";
- String RS_START_TIME_NAME = "regionServerStartTime";
- String ZOOKEEPER_QUORUM_NAME = "zookeeperQuorum";
- String SERVER_NAME_NAME = "serverName";
- String CLUSTER_ID_NAME = "clusterId";
- String RS_START_TIME_DESC = "RegionServer Start Time";
- String ZOOKEEPER_QUORUM_DESC = "Zookeeper Quorum";
- String SERVER_NAME_DESC = "Server Name";
- String CLUSTER_ID_DESC = "Cluster Id";
- String UPDATES_BLOCKED_TIME = "updatesBlockedTime";
- String UPDATES_BLOCKED_DESC =
+ static final String RS_START_TIME_NAME = "regionServerStartTime";
+ static final String ZOOKEEPER_QUORUM_NAME = "zookeeperQuorum";
+ static final String SERVER_NAME_NAME = "serverName";
+ static final String CLUSTER_ID_NAME = "clusterId";
+ static final String RS_START_TIME_DESC = "RegionServer Start Time";
+ static final String ZOOKEEPER_QUORUM_DESC = "Zookeeper Quorum";
+ static final String SERVER_NAME_DESC = "Server Name";
+ static final String CLUSTER_ID_DESC = "Cluster Id";
+ static final String UPDATES_BLOCKED_TIME = "updatesBlockedTime";
+ static final String UPDATES_BLOCKED_DESC =
"Number of MS updates have been blocked so that the memstore can be flushed.";
- String DELETE_KEY = "delete";
- String GET_KEY = "get";
- String INCREMENT_KEY = "increment";
- String MUTATE_KEY = "mutate";
- String APPEND_KEY = "append";
- String REPLAY_KEY = "replay";
- String SCAN_NEXT_KEY = "scanNext";
- String SLOW_MUTATE_KEY = "slowPutCount";
- String SLOW_GET_KEY = "slowGetCount";
- String SLOW_DELETE_KEY = "slowDeleteCount";
- String SLOW_INCREMENT_KEY = "slowIncrementCount";
- String SLOW_APPEND_KEY = "slowAppendCount";
- String SLOW_MUTATE_DESC =
+ static final String DELETE_KEY = "delete";
+ static final String GET_KEY = "get";
+ static final String INCREMENT_KEY = "increment";
+ static final String MUTATE_KEY = "mutate";
+ static final String APPEND_KEY = "append";
+ static final String REPLAY_KEY = "replay";
+ static final String SCAN_NEXT_KEY = "scanNext";
+ static final String SLOW_MUTATE_KEY = "slowPutCount";
+ static final String SLOW_GET_KEY = "slowGetCount";
+ static final String SLOW_DELETE_KEY = "slowDeleteCount";
+ static final String SLOW_INCREMENT_KEY = "slowIncrementCount";
+ static final String SLOW_APPEND_KEY = "slowAppendCount";
+ static final String SLOW_MUTATE_DESC =
"The number of Multis that took over 1000ms to complete";
- String SLOW_DELETE_DESC =
+ static final String SLOW_DELETE_DESC =
"The number of Deletes that took over 1000ms to complete";
- String SLOW_GET_DESC = "The number of Gets that took over 1000ms to complete";
- String SLOW_INCREMENT_DESC =
+ static final String SLOW_GET_DESC = "The number of Gets that took over 1000ms to complete";
+ static final String SLOW_INCREMENT_DESC =
"The number of Increments that took over 1000ms to complete";
- String SLOW_APPEND_DESC =
+ static final String SLOW_APPEND_DESC =
"The number of Appends that took over 1000ms to complete";
Modified: hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionServerWrapper.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionServerWrapper.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionServerWrapper.java (original)
+++ hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionServerWrapper.java Wed Jul 10 18:50:21 2013
@@ -27,35 +27,35 @@ public interface MetricsRegionServerWrap
/**
* Get ServerName
*/
- String getServerName();
+ public String getServerName();
/**
* Get the Cluster ID
*
* @return Cluster ID
*/
- String getClusterId();
+ public String getClusterId();
/**
* Get the Zookeeper Quorum Info
*
* @return Zookeeper Quorum Info
*/
- String getZookeeperQuorum();
+ public String getZookeeperQuorum();
/**
* Get the co-processors
*
* @return Co-processors
*/
- String getCoprocessors();
+ public String getCoprocessors();
/**
* Get HRegionServer start time
*
* @return Start time of RegionServer in milliseconds
*/
- long getStartCode();
+ public long getStartCode();
/**
* The number of online regions
Modified: hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionSource.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionSource.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionSource.java (original)
+++ hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionSource.java Wed Jul 10 18:50:21 2013
@@ -25,8 +25,8 @@ package org.apache.hadoop.hbase.regionse
*/
public interface MetricsRegionSource extends Comparable<MetricsRegionSource> {
- String OPS_SAMPLE_NAME = "ops";
- String SIZE_VALUE_NAME = "size";
+ public static final String OPS_SAMPLE_NAME = "ops";
+ public static final String SIZE_VALUE_NAME = "size";
/**
* Close the region's metrics as this region is closing.
Modified: hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/wal/MetricsEditsReplaySource.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/wal/MetricsEditsReplaySource.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/wal/MetricsEditsReplaySource.java (original)
+++ hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/wal/MetricsEditsReplaySource.java Wed Jul 10 18:50:21 2013
@@ -29,30 +29,30 @@ public interface MetricsEditsReplaySourc
/**
* The name of the metrics
*/
- String METRICS_NAME = "replay";
+ static final String METRICS_NAME = "replay";
/**
* The name of the metrics context that metrics will be under.
*/
- String METRICS_CONTEXT = "regionserver";
+ static final String METRICS_CONTEXT = "regionserver";
/**
* Description
*/
- String METRICS_DESCRIPTION = "Metrics about HBase RegionServer HLog Edits Replay";
+ static final String METRICS_DESCRIPTION = "Metrics about HBase RegionServer HLog Edits Replay";
/**
* The name of the metrics context that metrics will be under in jmx
*/
- String METRICS_JMX_CONTEXT = "RegionServer,sub=" + METRICS_NAME;
+ static final String METRICS_JMX_CONTEXT = "RegionServer,sub=" + METRICS_NAME;
- String REPLAY_TIME_NAME = "replayTime";
- String REPLAY_TIME_DESC = "Time an replay operation took.";
- String REPLAY_BATCH_SIZE_NAME = "replayBatchSize";
- String REPLAY_BATCH_SIZE_DESC = "Number of changes in each replay batch.";
- String REPLAY_DATA_SIZE_NAME = "replayDataSize";
- String REPLAY_DATA_SIZE_DESC = "Size (in bytes) of the data of each replay.";
+ static final String REPLAY_TIME_NAME = "replayTime";
+ static final String REPLAY_TIME_DESC = "Time an replay operation took.";
+ static final String REPLAY_BATCH_SIZE_NAME = "replayBatchSize";
+ static final String REPLAY_BATCH_SIZE_DESC = "Number of changes in each replay batch.";
+ static final String REPLAY_DATA_SIZE_NAME = "replayDataSize";
+ static final String REPLAY_DATA_SIZE_DESC = "Size (in bytes) of the data of each replay.";
/**
* Add the time a replay command took
Modified: hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/wal/MetricsWALSource.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/wal/MetricsWALSource.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/wal/MetricsWALSource.java (original)
+++ hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/wal/MetricsWALSource.java Wed Jul 10 18:50:21 2013
@@ -29,34 +29,34 @@ public interface MetricsWALSource extend
/**
* The name of the metrics
*/
- String METRICS_NAME = "WAL";
+ static final String METRICS_NAME = "WAL";
/**
* The name of the metrics context that metrics will be under.
*/
- String METRICS_CONTEXT = "regionserver";
+ static final String METRICS_CONTEXT = "regionserver";
/**
* Description
*/
- String METRICS_DESCRIPTION = "Metrics about HBase RegionServer HLog";
+ static final String METRICS_DESCRIPTION = "Metrics about HBase RegionServer HLog";
/**
* The name of the metrics context that metrics will be under in jmx
*/
- String METRICS_JMX_CONTEXT = "RegionServer,sub=" + METRICS_NAME;
+ static final String METRICS_JMX_CONTEXT = "RegionServer,sub=" + METRICS_NAME;
- String APPEND_TIME = "appendTime";
- String APPEND_TIME_DESC = "Time an append to the log took.";
- String APPEND_COUNT = "appendCount";
- String APPEND_COUNT_DESC = "Number of appends to the write ahead log.";
- String APPEND_SIZE = "appendSize";
- String APPEND_SIZE_DESC = "Size (in bytes) of the data appended to the HLog.";
- String SLOW_APPEND_COUNT = "slowAppendCount";
- String SLOW_APPEND_COUNT_DESC = "Number of appends that were slow.";
- String SYNC_TIME = "syncTime";
- String SYNC_TIME_DESC = "The time it took to sync the HLog to HDFS.";
+ static final String APPEND_TIME = "appendTime";
+ static final String APPEND_TIME_DESC = "Time an append to the log took.";
+ static final String APPEND_COUNT = "appendCount";
+ static final String APPEND_COUNT_DESC = "Number of appends to the write ahead log.";
+ static final String APPEND_SIZE = "appendSize";
+ static final String APPEND_SIZE_DESC = "Size (in bytes) of the data appended to the HLog.";
+ static final String SLOW_APPEND_COUNT = "slowAppendCount";
+ static final String SLOW_APPEND_COUNT_DESC = "Number of appends that were slow.";
+ static final String SYNC_TIME = "syncTime";
+ static final String SYNC_TIME_DESC = "The time it took to sync the HLog to HDFS.";
/**
* Add the append size.
Modified: hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/replication/regionserver/MetricsReplicationSource.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/replication/regionserver/MetricsReplicationSource.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/replication/regionserver/MetricsReplicationSource.java (original)
+++ hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/replication/regionserver/MetricsReplicationSource.java Wed Jul 10 18:50:21 2013
@@ -28,21 +28,21 @@ public interface MetricsReplicationSourc
/**
* The name of the metrics
*/
- String METRICS_NAME = "Replication";
+ static final String METRICS_NAME = "Replication";
/**
* The name of the metrics context that metrics will be under.
*/
- String METRICS_CONTEXT = "regionserver";
+ static final String METRICS_CONTEXT = "regionserver";
/**
* The name of the metrics context that metrics will be under.
*/
- String METRICS_JMX_CONTEXT = "RegionServer,sub=" + METRICS_NAME;
+ static final String METRICS_JMX_CONTEXT = "RegionServer,sub=" + METRICS_NAME;
/**
* A description.
*/
- String METRICS_DESCRIPTION = "Metrics about HBase replication";
+ static final String METRICS_DESCRIPTION = "Metrics about HBase replication";
}
Modified: hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/rest/MetricsRESTSource.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/rest/MetricsRESTSource.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/rest/MetricsRESTSource.java (original)
+++ hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/rest/MetricsRESTSource.java Wed Jul 10 18:50:21 2013
@@ -25,27 +25,27 @@ import org.apache.hadoop.hbase.metrics.B
*/
public interface MetricsRESTSource extends BaseSource {
- String METRICS_NAME = "REST";
+ public static String METRICS_NAME = "REST";
- String CONTEXT = "rest";
+ public static String CONTEXT = "rest";
- String JMX_CONTEXT = "REST";
+ public static String JMX_CONTEXT = "REST";
- String METRICS_DESCRIPTION = "Metrics about the HBase REST server";
+ public static String METRICS_DESCRIPTION = "Metrics about the HBase REST server";
- String REQUEST_KEY = "requests";
+ static String REQUEST_KEY = "requests";
- String SUCCESSFUL_GET_KEY = "successfulGet";
+ static String SUCCESSFUL_GET_KEY = "successfulGet";
- String SUCCESSFUL_PUT_KEY = "successfulPut";
+ static String SUCCESSFUL_PUT_KEY = "successfulPut";
- String SUCCESSFUL_DELETE_KEY = "successfulDelete";
+ static String SUCCESSFUL_DELETE_KEY = "successfulDelete";
- String FAILED_GET_KEY = "failedGet";
+ static String FAILED_GET_KEY = "failedGet";
- String FAILED_PUT_KEY = "failedPut";
+ static String FAILED_PUT_KEY = "failedPut";
- String FAILED_DELETE_KEY = "failedDelete";
+ static String FAILED_DELETE_KEY = "failedDelete";
/**
* Increment the number of requests
Modified: hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/thrift/MetricsThriftServerSource.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/thrift/MetricsThriftServerSource.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/thrift/MetricsThriftServerSource.java (original)
+++ hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/thrift/MetricsThriftServerSource.java Wed Jul 10 18:50:21 2013
@@ -25,12 +25,12 @@ import org.apache.hadoop.hbase.metrics.B
*/
public interface MetricsThriftServerSource extends BaseSource {
- String BATCH_GET_KEY = "batchGet";
- String BATCH_MUTATE_KEY = "batchMutate";
- String TIME_IN_QUEUE_KEY = "timeInQueue";
- String THRIFT_CALL_KEY = "thriftCall";
- String SLOW_THRIFT_CALL_KEY = "slowThriftCall";
- String CALL_QUEUE_LEN_KEY = "callQueueLen";
+ static final String BATCH_GET_KEY = "batchGet";
+ static final String BATCH_MUTATE_KEY = "batchMutate";
+ static final String TIME_IN_QUEUE_KEY = "timeInQueue";
+ static final String THRIFT_CALL_KEY = "thriftCall";
+ static final String SLOW_THRIFT_CALL_KEY = "slowThriftCall";
+ static final String CALL_QUEUE_LEN_KEY = "callQueueLen";
/**
* Add how long an operation was in the queue.
Modified: hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/thrift/MetricsThriftServerSourceFactory.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/thrift/MetricsThriftServerSourceFactory.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/thrift/MetricsThriftServerSourceFactory.java (original)
+++ hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/thrift/MetricsThriftServerSourceFactory.java Wed Jul 10 18:50:21 2013
@@ -21,12 +21,12 @@ package org.apache.hadoop.hbase.thrift;
/** Factory that will be used to create metrics sources for the two diffent types of thrift servers. */
public interface MetricsThriftServerSourceFactory {
- String METRICS_NAME = "Thrift";
- String METRICS_DESCRIPTION = "Thrift Server Metrics";
- String THRIFT_ONE_METRICS_CONTEXT = "thrift-one";
- String THRIFT_ONE_JMX_CONTEXT = "Thrift,sub=ThriftOne";
- String THRIFT_TWO_METRICS_CONTEXT = "thrift-two";
- String THRIFT_TWO_JMX_CONTEXT = "Thrift,sub=ThriftTwo";
+ static final String METRICS_NAME = "Thrift";
+ static final String METRICS_DESCRIPTION = "Thrift Server Metrics";
+ static final String THRIFT_ONE_METRICS_CONTEXT = "thrift-one";
+ static final String THRIFT_ONE_JMX_CONTEXT = "Thrift,sub=ThriftOne";
+ static final String THRIFT_TWO_METRICS_CONTEXT = "thrift-two";
+ static final String THRIFT_TWO_JMX_CONTEXT = "Thrift,sub=ThriftTwo";
/** Create a Source for a thrift one server */
MetricsThriftServerSource createThriftOneSource();
Modified: hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/metrics2/MetricHistogram.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/metrics2/MetricHistogram.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/metrics2/MetricHistogram.java (original)
+++ hbase/trunk/hbase-hadoop-compat/src/main/java/org/apache/hadoop/metrics2/MetricHistogram.java Wed Jul 10 18:50:21 2013
@@ -25,14 +25,14 @@ package org.apache.hadoop.metrics2;
public interface MetricHistogram {
//Strings used to create metrics names.
- String NUM_OPS_METRIC_NAME = "_num_ops";
- String MIN_METRIC_NAME = "_min";
- String MAX_METRIC_NAME = "_max";
- String MEAN_METRIC_NAME = "_mean";
- String MEDIAN_METRIC_NAME = "_median";
- String SEVENTY_FIFTH_PERCENTILE_METRIC_NAME = "_75th_percentile";
- String NINETY_FIFTH_PERCENTILE_METRIC_NAME = "_95th_percentile";
- String NINETY_NINETH_PERCENTILE_METRIC_NAME = "_99th_percentile";
+ static final String NUM_OPS_METRIC_NAME = "_num_ops";
+ static final String MIN_METRIC_NAME = "_min";
+ static final String MAX_METRIC_NAME = "_max";
+ static final String MEAN_METRIC_NAME = "_mean";
+ static final String MEDIAN_METRIC_NAME = "_median";
+ static final String SEVENTY_FIFTH_PERCENTILE_METRIC_NAME = "_75th_percentile";
+ static final String NINETY_FIFTH_PERCENTILE_METRIC_NAME = "_95th_percentile";
+ static final String NINETY_NINETH_PERCENTILE_METRIC_NAME = "_99th_percentile";
/**
* Add a single value to a histogram's stream of values.
Modified: hbase/trunk/hbase-hadoop-compat/src/test/java/org/apache/hadoop/hbase/HadoopShims.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-hadoop-compat/src/test/java/org/apache/hadoop/hbase/HadoopShims.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-hadoop-compat/src/test/java/org/apache/hadoop/hbase/HadoopShims.java (original)
+++ hbase/trunk/hbase-hadoop-compat/src/test/java/org/apache/hadoop/hbase/HadoopShims.java Wed Jul 10 18:50:21 2013
@@ -32,6 +32,6 @@ public interface HadoopShims {
* TaskAttemptId.forName()
* @return a concrete TaskAttemptContext instance of o.a.h.mapreduce.TaskAttemptContext
*/
- <T,J> T createTestTaskAttemptContext(final J job, final String taskId);
+ public <T,J> T createTestTaskAttemptContext(final J job, final String taskId);
}
Modified: hbase/trunk/hbase-hadoop-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelper.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-hadoop-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelper.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-hadoop-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelper.java (original)
+++ hbase/trunk/hbase-hadoop-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelper.java Wed Jul 10 18:50:21 2013
@@ -27,7 +27,7 @@ public interface MetricsAssertHelper {
* Init helper. This method will make sure that the metrics system is set
* up for tests.
*/
- void init();
+ public void init();
/**
* Assert that a tag exists and has a given value.
@@ -37,7 +37,7 @@ public interface MetricsAssertHelper {
* @param source The BaseSource{@link BaseSource} that will provide the tags,
* gauges, and counters.
*/
- void assertTag(String name, String expected, BaseSource source);
+ public void assertTag(String name, String expected, BaseSource source);
/**
* Assert that a gauge exists and that it's value is equal to the expected value.
@@ -47,7 +47,7 @@ public interface MetricsAssertHelper {
* @param source The BaseSource{@link BaseSource} that will provide the tags,
* gauges, and counters.
*/
- void assertGauge(String name, long expected, BaseSource source);
+ public void assertGauge(String name, long expected, BaseSource source);
/**
* Assert that a gauge exists and it's value is greater than a given value
@@ -57,7 +57,7 @@ public interface MetricsAssertHelper {
* @param source The BaseSource{@link BaseSource} that will provide the tags,
* gauges, and counters.
*/
- void assertGaugeGt(String name, long expected, BaseSource source);
+ public void assertGaugeGt(String name, long expected, BaseSource source);
/**
* Assert that a gauge exists and it's value is less than a given value
@@ -67,7 +67,7 @@ public interface MetricsAssertHelper {
* @param source The BaseSource{@link BaseSource} that will provide the tags,
* gauges, and counters.
*/
- void assertGaugeLt(String name, long expected, BaseSource source);
+ public void assertGaugeLt(String name, long expected, BaseSource source);
/**
* Assert that a gauge exists and that it's value is equal to the expected value.
@@ -77,7 +77,7 @@ public interface MetricsAssertHelper {
* @param source The BaseSource{@link BaseSource} that will provide the tags,
* gauges, and counters.
*/
- void assertGauge(String name, double expected, BaseSource source);
+ public void assertGauge(String name, double expected, BaseSource source);
/**
* Assert that a gauge exists and it's value is greater than a given value
@@ -87,7 +87,7 @@ public interface MetricsAssertHelper {
* @param source The BaseSource{@link BaseSource} that will provide the tags,
* gauges, and counters.
*/
- void assertGaugeGt(String name, double expected, BaseSource source);
+ public void assertGaugeGt(String name, double expected, BaseSource source);
/**
* Assert that a gauge exists and it's value is less than a given value
@@ -97,7 +97,7 @@ public interface MetricsAssertHelper {
* @param source The BaseSource{@link BaseSource} that will provide the tags,
* gauges, and counters.
*/
- void assertGaugeLt(String name, double expected, BaseSource source);
+ public void assertGaugeLt(String name, double expected, BaseSource source);
/**
* Assert that a counter exists and that it's value is equal to the expected value.
@@ -107,7 +107,7 @@ public interface MetricsAssertHelper {
* @param source The BaseSource{@link BaseSource} that will provide the tags,
* gauges, and counters.
*/
- void assertCounter(String name, long expected, BaseSource source);
+ public void assertCounter(String name, long expected, BaseSource source);
/**
* Assert that a counter exists and that it's value is greater than the given value.
@@ -117,7 +117,7 @@ public interface MetricsAssertHelper {
* @param source The BaseSource{@link BaseSource} that will provide the tags,
* gauges, and counters.
*/
- void assertCounterGt(String name, long expected, BaseSource source);
+ public void assertCounterGt(String name, long expected, BaseSource source);
/**
* Assert that a counter exists and that it's value is less than the given value.
@@ -127,7 +127,7 @@ public interface MetricsAssertHelper {
* @param source The BaseSource{@link BaseSource} that will provide the tags,
* gauges, and counters.
*/
- void assertCounterLt(String name, long expected, BaseSource source);
+ public void assertCounterLt(String name, long expected, BaseSource source);
/**
* Get the value of a counter.
@@ -137,7 +137,7 @@ public interface MetricsAssertHelper {
* gauges, and counters.
* @return long value of the counter.
*/
- long getCounter(String name, BaseSource source);
+ public long getCounter(String name, BaseSource source);
/**
* Get the value of a gauge as a double.
@@ -147,7 +147,7 @@ public interface MetricsAssertHelper {
* gauges, and counters.
* @return double value of the gauge.
*/
- double getGaugeDouble(String name, BaseSource source);
+ public double getGaugeDouble(String name, BaseSource source);
/**
* Get the value of a gauge as a long.
@@ -157,5 +157,5 @@ public interface MetricsAssertHelper {
* gauges, and counters.
* @return long value of the gauge.
*/
- long getGaugeLong(String name, BaseSource source);
+ public long getGaugeLong(String name, BaseSource source);
}
Modified: hbase/trunk/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/builder/TestTokenizerData.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/builder/TestTokenizerData.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/builder/TestTokenizerData.java (original)
+++ hbase/trunk/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/builder/TestTokenizerData.java Wed Jul 10 18:50:21 2013
@@ -31,7 +31,7 @@ public interface TestTokenizerData {
List<byte[]> getInputs();
List<byte[]> getOutputs();
- class InMemory {
+ public static class InMemory {
public Collection<Object[]> getAllAsObjectArray() {
List<Object[]> all = Lists.newArrayList();
all.add(new Object[] { new TestTokenizerDataBasic() });
Modified: hbase/trunk/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/column/TestColumnData.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/column/TestColumnData.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/column/TestColumnData.java (original)
+++ hbase/trunk/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/column/TestColumnData.java Wed Jul 10 18:50:21 2013
@@ -32,7 +32,7 @@ public interface TestColumnData {
List<ByteRange> getInputs();
List<ByteRange> getOutputs();
- class InMemory {
+ public static class InMemory {
public Collection<Object[]> getAllAsObjectArray() {
List<Object[]> all = Lists.newArrayList();
all.add(new Object[] { new TestColumnDataSimple() });
Modified: hbase/trunk/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/TestRowData.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/TestRowData.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/TestRowData.java (original)
+++ hbase/trunk/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/TestRowData.java Wed Jul 10 18:50:21 2013
@@ -54,7 +54,7 @@ public interface TestRowData {
void individualSearcherAssertions(CellSearcher searcher);
- class InMemory {
+ public static class InMemory {
/*
* The following are different styles of data that the codec may encounter. Having these small
Modified: hbase/trunk/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/timestamp/TestTimestampData.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/timestamp/TestTimestampData.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/timestamp/TestTimestampData.java (original)
+++ hbase/trunk/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/timestamp/TestTimestampData.java Wed Jul 10 18:50:21 2013
@@ -33,7 +33,7 @@ public interface TestTimestampData {
long getMinimum();
List<Long> getOutputs();
- class InMemory {
+ public static class InMemory {
public Collection<Object[]> getAllAsObjectArray() {
List<Object[]> all = Lists.newArrayList();
all.add(new Object[] { new TestTimestampDataBasic() });
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/InterProcessLock.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/InterProcessLock.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/InterProcessLock.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/InterProcessLock.java Wed Jul 10 18:50:21 2013
@@ -36,7 +36,7 @@ public interface InterProcessLock {
* @throws InterruptedException If current thread is interrupted while
* waiting for the lock
*/
- void acquire() throws IOException, InterruptedException;
+ public void acquire() throws IOException, InterruptedException;
/**
* Acquire the lock within a wait time.
@@ -50,7 +50,7 @@ public interface InterProcessLock {
* @throws InterruptedException If the thread is interrupted while waiting to
* acquire the lock
*/
- boolean tryAcquire(long timeoutMs)
+ public boolean tryAcquire(long timeoutMs)
throws IOException, InterruptedException;
/**
@@ -59,7 +59,7 @@ public interface InterProcessLock {
* @throws InterruptedException If the thread is interrupted while releasing
* the lock
*/
- void release() throws IOException, InterruptedException;
+ public void release() throws IOException, InterruptedException;
/**
* If supported, attempts to reap all the locks of this type by forcefully
@@ -69,7 +69,7 @@ public interface InterProcessLock {
* lock holder is still alive.
* @throws IOException If there is an unrecoverable error reaping the locks
*/
- void reapExpiredLocks(long expireTimeoutMs) throws IOException;
+ public void reapExpiredLocks(long expireTimeoutMs) throws IOException;
/**
* If supported, attempts to reap all the locks of this type by forcefully
@@ -80,12 +80,12 @@ public interface InterProcessLock {
* with timeout=0.
* @throws IOException If there is an unrecoverable error reaping the locks
*/
- void reapAllLocks() throws IOException;
+ public void reapAllLocks() throws IOException;
/**
* An interface for objects that process lock metadata.
*/
- interface MetadataHandler {
+ public static interface MetadataHandler {
/**
* Called after lock metadata is successfully read from a distributed
@@ -93,7 +93,7 @@ public interface InterProcessLock {
* printing the metadata in a humanly-readable format.
* @param metadata The metadata
*/
- void handleMetadata(byte[] metadata);
+ public void handleMetadata(byte[] metadata);
}
/**
@@ -101,5 +101,5 @@ public interface InterProcessLock {
* {@link MetadataHandler}.
* @throws InterruptedException If there is an unrecoverable error
*/
- void visitLocks(MetadataHandler handler) throws IOException;
+ public void visitLocks(MetadataHandler handler) throws IOException;
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/InterProcessReadWriteLock.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/InterProcessReadWriteLock.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/InterProcessReadWriteLock.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/InterProcessReadWriteLock.java Wed Jul 10 18:50:21 2013
@@ -34,7 +34,7 @@ public interface InterProcessReadWriteLo
* which the lock was acquired).
* @return An instantiated InterProcessLock instance
*/
- InterProcessLock readLock(byte[] metadata);
+ public InterProcessLock readLock(byte[] metadata);
/**
* Obtain a write lock containing given metadata.
@@ -43,5 +43,5 @@ public interface InterProcessReadWriteLo
* which the lock was acquired).
* @return An instantiated InterProcessLock instance
*/
- InterProcessLock writeLock(byte[] metadata);
+ public InterProcessLock writeLock(byte[] metadata);
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java Wed Jul 10 18:50:21 2013
@@ -35,7 +35,7 @@ public interface TableDescriptors {
* @return HTableDescriptor for tablename
* @throws IOException
*/
- HTableDescriptor get(final String tablename)
+ public HTableDescriptor get(final String tablename)
throws IOException;
/**
@@ -43,7 +43,7 @@ public interface TableDescriptors {
* @return HTableDescriptor for tablename
* @throws IOException
*/
- HTableDescriptor get(final byte[] tablename)
+ public HTableDescriptor get(final byte[] tablename)
throws IOException;
/**
@@ -52,7 +52,7 @@ public interface TableDescriptors {
* @return Map of all descriptors.
* @throws IOException
*/
- Map<String, HTableDescriptor> getAll()
+ public Map<String, HTableDescriptor> getAll()
throws IOException;
/**
@@ -60,7 +60,7 @@ public interface TableDescriptors {
* @param htd Descriptor to set into TableDescriptors
* @throws IOException
*/
- void add(final HTableDescriptor htd)
+ public void add(final HTableDescriptor htd)
throws IOException;
/**
@@ -68,6 +68,6 @@ public interface TableDescriptors {
* @return Instance of table descriptor or null if none found.
* @throws IOException
*/
- HTableDescriptor remove(final String tablename)
+ public HTableDescriptor remove(final String tablename)
throws IOException;
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/constraint/Constraint.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/constraint/Constraint.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/constraint/Constraint.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/constraint/Constraint.java Wed Jul 10 18:50:21 2013
@@ -76,6 +76,6 @@ public interface Constraint extends Conf
* @throws org.apache.hadoop.hbase.exceptions.ConstraintException when the {@link Put} does not match the
* constraint.
*/
- void check(Put p) throws ConstraintException;
+ public void check(Put p) throws ConstraintException;
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorService.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorService.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorService.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorService.java Wed Jul 10 18:50:21 2013
@@ -29,5 +29,5 @@ import org.apache.hadoop.classification.
@InterfaceAudience.Public
@InterfaceStability.Evolving
public interface CoprocessorService {
- Service getService();
+ public Service getService();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.java Wed Jul 10 18:50:21 2013
@@ -31,12 +31,12 @@ import org.apache.hadoop.hbase.regionser
@InterfaceStability.Evolving
public interface RegionCoprocessorEnvironment extends CoprocessorEnvironment {
/** @return the region associated with this coprocessor */
- HRegion getRegion();
+ public HRegion getRegion();
/** @return reference to the region server services */
- RegionServerServices getRegionServerServices();
+ public RegionServerServices getRegionServerServices();
/** @return shared data between all instances of this coprocessor */
- ConcurrentMap<String, Object> getSharedData();
+ public ConcurrentMap<String, Object> getSharedData();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/WALCoprocessorEnvironment.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/WALCoprocessorEnvironment.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/WALCoprocessorEnvironment.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/WALCoprocessorEnvironment.java Wed Jul 10 18:50:21 2013
@@ -28,5 +28,5 @@ import org.apache.hadoop.hbase.regionser
@InterfaceStability.Evolving
public interface WALCoprocessorEnvironment extends CoprocessorEnvironment {
/** @return reference to the region server services */
- HLog getWAL();
+ public HLog getWAL();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/errorhandling/ForeignExceptionListener.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/errorhandling/ForeignExceptionListener.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/errorhandling/ForeignExceptionListener.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/errorhandling/ForeignExceptionListener.java Wed Jul 10 18:50:21 2013
@@ -36,5 +36,5 @@ public interface ForeignExceptionListene
* Implementers must ensure that this method is thread-safe.
* @param e exception causing the error. Implementations must accept and handle null here.
*/
- void receive(ForeignException e);
-}
+ public void receive(ForeignException e);
+}
\ No newline at end of file
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/errorhandling/ForeignExceptionSnare.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/errorhandling/ForeignExceptionSnare.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/errorhandling/ForeignExceptionSnare.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/errorhandling/ForeignExceptionSnare.java Wed Jul 10 18:50:21 2013
@@ -47,7 +47,7 @@ public interface ForeignExceptionSnare {
* @throws ForeignException
* all exceptions from remote sources are procedure exceptions
*/
- void rethrowException() throws ForeignException;
+ public void rethrowException() throws ForeignException;
/**
* Non-exceptional form of {@link #rethrowException()}. Checks to see if any
@@ -56,12 +56,12 @@ public interface ForeignExceptionSnare {
*
* @return <tt>true</tt> if there has been an error,<tt>false</tt> otherwise
*/
- boolean hasException();
+ public boolean hasException();
/**
* Get the value of the captured exception.
*
* @return the captured foreign exception or null if no exception captured.
*/
- ForeignException getException();
+ public ForeignException getException();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/EventHandler.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/EventHandler.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/EventHandler.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/EventHandler.java Wed Jul 10 18:50:21 2013
@@ -87,12 +87,12 @@ public abstract class EventHandler imple
* Called before any event is processed
* @param event The event handler whose process method is about to be called.
*/
- void beforeProcess(EventHandler event);
+ public void beforeProcess(EventHandler event);
/**
* Called after any event is processed
* @param event The event handler whose process method is about to be called.
*/
- void afterProcess(EventHandler event);
+ public void afterProcess(EventHandler event);
}
/**
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/fs/HFileSystem.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/fs/HFileSystem.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/fs/HFileSystem.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/fs/HFileSystem.java Wed Jul 10 18:50:21 2013
@@ -296,7 +296,7 @@ public class HFileSystem extends FilterF
/**
* Interface to implement to add a specific reordering logic in hdfs.
*/
- interface ReorderBlocks {
+ static interface ReorderBlocks {
/**
*
* @param conf - the conf to use
@@ -304,7 +304,7 @@ public class HFileSystem extends FilterF
* @param src - the file name currently read
* @throws IOException - if something went wrong
*/
- void reorderBlocks(Configuration conf, LocatedBlocks lbs, String src) throws IOException;
+ public void reorderBlocks(Configuration conf, LocatedBlocks lbs, String src) throws IOException;
}
/**
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/io/WritableWithSize.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/io/WritableWithSize.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/io/WritableWithSize.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/io/WritableWithSize.java Wed Jul 10 18:50:21 2013
@@ -34,5 +34,5 @@ public interface WritableWithSize {
*
* @return the size of the writable
*/
- long getWritableSize();
+ public long getWritableSize();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java Wed Jul 10 18:50:21 2013
@@ -36,14 +36,14 @@ public interface BlockCache {
* @param buf The block contents wrapped in a ByteBuffer.
* @param inMemory Whether block should be treated as in-memory
*/
- void cacheBlock(BlockCacheKey cacheKey, Cacheable buf, boolean inMemory);
+ public void cacheBlock(BlockCacheKey cacheKey, Cacheable buf, boolean inMemory);
/**
* Add block to cache (defaults to not in-memory).
* @param cacheKey The block's cache key.
* @param buf The object to cache.
*/
- void cacheBlock(BlockCacheKey cacheKey, Cacheable buf);
+ public void cacheBlock(BlockCacheKey cacheKey, Cacheable buf);
/**
* Fetch block from cache.
@@ -54,62 +54,62 @@ public interface BlockCache {
* @return Block or null if block is not in 2 cache.
* @see HFileReaderV2#readBlock(long, long, boolean, boolean, boolean, BlockType)
*/
- Cacheable getBlock(BlockCacheKey cacheKey, boolean caching, boolean repeat);
+ public Cacheable getBlock(BlockCacheKey cacheKey, boolean caching, boolean repeat);
/**
* Evict block from cache.
* @param cacheKey Block to evict
* @return true if block existed and was evicted, false if not
*/
- boolean evictBlock(BlockCacheKey cacheKey);
+ public boolean evictBlock(BlockCacheKey cacheKey);
/**
* Evicts all blocks for the given HFile.
*
* @return the number of blocks evicted
*/
- int evictBlocksByHfileName(String hfileName);
+ public int evictBlocksByHfileName(String hfileName);
/**
* Get the statistics for this block cache.
* @return Stats
*/
- CacheStats getStats();
+ public CacheStats getStats();
/**
* Shutdown the cache.
*/
- void shutdown();
+ public void shutdown();
/**
* Returns the total size of the block cache, in bytes.
* @return size of cache, in bytes
*/
- long size();
+ public long size();
/**
* Returns the free size of the block cache, in bytes.
* @return free space in cache, in bytes
*/
- long getFreeSize();
+ public long getFreeSize();
/**
* Returns the occupied size of the block cache, in bytes.
* @return occupied space in cache, in bytes
*/
- long getCurrentSize();
+ public long getCurrentSize();
/**
* Returns the number of evictions that have occurred.
* @return number of evictions
*/
- long getEvictedCount();
+ public long getEvictedCount();
/**
* Returns the number of blocks currently cached in the block cache.
* @return number of blocks in the cache
*/
- long getBlockCount();
+ public long getBlockCount();
/**
* Performs a BlockCache summary and returns a List of BlockCacheColumnFamilySummary objects.
@@ -123,5 +123,5 @@ public interface BlockCache {
* @return List of BlockCacheColumnFamilySummary
* @throws IOException exception
*/
- List<BlockCacheColumnFamilySummary> getBlockCacheColumnFamilySummaries(Configuration conf) throws IOException;
+ public List<BlockCacheColumnFamilySummary> getBlockCacheColumnFamilySummaries(Configuration conf) throws IOException;
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/Cacheable.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/Cacheable.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/Cacheable.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/Cacheable.java Wed Jul 10 18:50:21 2013
@@ -42,23 +42,23 @@ public interface Cacheable extends HeapS
* @return int length in bytes of the serialized form.
*/
- int getSerializedLength();
+ public int getSerializedLength();
/**
* Serializes its data into destination.
*/
- void serialize(ByteBuffer destination);
+ public void serialize(ByteBuffer destination);
/**
* Returns CacheableDeserializer instance which reconstructs original object from ByteBuffer.
*
* @return CacheableDeserialzer instance.
*/
- CacheableDeserializer<Cacheable> getDeserializer();
+ public CacheableDeserializer<Cacheable> getDeserializer();
/**
* @return the block type of this cached HFile block
*/
- BlockType getBlockType();
+ public BlockType getBlockType();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheableDeserializer.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheableDeserializer.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheableDeserializer.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheableDeserializer.java Wed Jul 10 18:50:21 2013
@@ -33,7 +33,7 @@ public interface CacheableDeserializer<T
*
* @return T the deserialized object.
*/
- T deserialize(ByteBuffer b) throws IOException;
+ public T deserialize(ByteBuffer b) throws IOException;
/**
*
@@ -43,12 +43,12 @@ public interface CacheableDeserializer<T
* @return T the deserialized object.
* @throws IOException
*/
- T deserialize(ByteBuffer b, boolean reuse) throws IOException;
+ public T deserialize(ByteBuffer b, boolean reuse) throws IOException;
/**
* Get the identifier of this deserialiser. Identifier is unique for each
* deserializer and generated by {@link CacheableDeserializerIdManager}
* @return identifier number of this cacheable deserializer
*/
- int getDeserialiserIdentifier();
+ public int getDeserialiserIdentifier();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileDataBlockEncoder.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileDataBlockEncoder.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileDataBlockEncoder.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileDataBlockEncoder.java Wed Jul 10 18:50:21 2013
@@ -49,9 +49,8 @@ public interface HFileDataBlockEncoder {
* generated).
* @return non null block which is coded according to the settings.
*/
- HFileBlock diskToCacheFormat(
- HFileBlock block, boolean isCompaction
- );
+ public HFileBlock diskToCacheFormat(HFileBlock block,
+ boolean isCompaction);
/**
* Should be called before an encoded or unencoded data block is written to
@@ -61,39 +60,37 @@ public interface HFileDataBlockEncoder {
* @param blockType block type
* @throws IOException
*/
- void beforeWriteToDisk(
- ByteBuffer in,
- boolean includesMemstoreTS,
- HFileBlockEncodingContext encodingResult,
- BlockType blockType
- ) throws IOException;
+ public void beforeWriteToDisk(
+ ByteBuffer in, boolean includesMemstoreTS,
+ HFileBlockEncodingContext encodingResult,
+ BlockType blockType) throws IOException;
/**
* Decides whether we should use a scanner over encoded blocks.
* @param isCompaction whether we are in a compaction.
* @return Whether to use encoded scanner.
*/
- boolean useEncodedScanner(boolean isCompaction);
+ public boolean useEncodedScanner(boolean isCompaction);
/**
* Save metadata in HFile which will be written to disk
* @param writer writer for a given HFile
* @exception IOException on disk problems
*/
- void saveMetadata(HFile.Writer writer)
+ public void saveMetadata(HFile.Writer writer)
throws IOException;
/** @return the on-disk data block encoding */
- DataBlockEncoding getEncodingOnDisk();
+ public DataBlockEncoding getEncodingOnDisk();
/** @return the preferred in-cache data block encoding for normal reads */
- DataBlockEncoding getEncodingInCache();
+ public DataBlockEncoding getEncodingInCache();
/**
* @return the effective in-cache data block encoding, taking into account
* whether we are doing a compaction.
*/
- DataBlockEncoding getEffectiveEncodingInCache(boolean isCompaction);
+ public DataBlockEncoding getEffectiveEncodingInCache(boolean isCompaction);
/**
* Create an encoder specific encoding context object for writing. And the
@@ -104,9 +101,8 @@ public interface HFileDataBlockEncoder {
* @param headerBytes header bytes
* @return a new {@link HFileBlockEncodingContext} object
*/
- HFileBlockEncodingContext newOnDiskDataBlockEncodingContext(
- Algorithm compressionAlgorithm, byte[] headerBytes
- );
+ public HFileBlockEncodingContext newOnDiskDataBlockEncodingContext(
+ Algorithm compressionAlgorithm, byte[] headerBytes);
/**
* create a encoder specific decoding context for reading. And the
@@ -116,8 +112,7 @@ public interface HFileDataBlockEncoder {
* @param compressionAlgorithm
* @return a new {@link HFileBlockDecodingContext} object
*/
- HFileBlockDecodingContext newOnDiskDataBlockDecodingContext(
- Algorithm compressionAlgorithm
- );
+ public HFileBlockDecodingContext newOnDiskDataBlockDecodingContext(
+ Algorithm compressionAlgorithm);
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileScanner.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileScanner.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileScanner.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileScanner.java Wed Jul 10 18:50:21 2013
@@ -54,8 +54,8 @@ public interface HFileScanner {
* false when it is called.
* @throws IOException
*/
- int seekTo(byte[] key) throws IOException;
- int seekTo(byte[] key, int offset, int length) throws IOException;
+ public int seekTo(byte[] key) throws IOException;
+ public int seekTo(byte[] key, int offset, int length) throws IOException;
/**
* Reseek to or just before the passed <code>key</code>. Similar to seekTo
* except that this can be called even if the scanner is not at the beginning
@@ -76,8 +76,8 @@ public interface HFileScanner {
* 1, such that k[i] < key, and scanner is left in position i.
* @throws IOException
*/
- int reseekTo(byte[] key) throws IOException;
- int reseekTo(byte[] key, int offset, int length) throws IOException;
+ public int reseekTo(byte[] key) throws IOException;
+ public int reseekTo(byte[] key, int offset, int length) throws IOException;
/**
* Consider the key stream of all the keys in the file,
* <code>k[0] .. k[n]</code>, where there are n keys in the file.
@@ -88,28 +88,28 @@ public interface HFileScanner {
* return false (EOF).
* @throws IOException
*/
- boolean seekBefore(byte[] key) throws IOException;
- boolean seekBefore(byte[] key, int offset, int length) throws IOException;
+ public boolean seekBefore(byte [] key) throws IOException;
+ public boolean seekBefore(byte []key, int offset, int length) throws IOException;
/**
* Positions this scanner at the start of the file.
* @return False if empty file; i.e. a call to next would return false and
* the current key and value are undefined.
* @throws IOException
*/
- boolean seekTo() throws IOException;
+ public boolean seekTo() throws IOException;
/**
* Scans to the next entry in the file.
* @return Returns false if you are at the end otherwise true if more in file.
* @throws IOException
*/
- boolean next() throws IOException;
+ public boolean next() throws IOException;
/**
* Gets a buffer view to the current key. You must call
* {@link #seekTo(byte[])} before this method.
* @return byte buffer for the key. The limit is set to the key size, and the
* position is 0, the start of the buffer view.
*/
- ByteBuffer getKey();
+ public ByteBuffer getKey();
/**
* Gets a buffer view to the current value. You must call
* {@link #seekTo(byte[])} before this method.
@@ -117,31 +117,31 @@ public interface HFileScanner {
* @return byte buffer for the value. The limit is set to the value size, and
* the position is 0, the start of the buffer view.
*/
- ByteBuffer getValue();
+ public ByteBuffer getValue();
/**
* @return Instance of {@link KeyValue}.
*/
- KeyValue getKeyValue();
+ public KeyValue getKeyValue();
/**
* Convenience method to get a copy of the key as a string - interpreting the
* bytes as UTF8. You must call {@link #seekTo(byte[])} before this method.
* @return key as a string
*/
- String getKeyString();
+ public String getKeyString();
/**
* Convenience method to get a copy of the value as a string - interpreting
* the bytes as UTF8. You must call {@link #seekTo(byte[])} before this method.
* @return value as a string
*/
- String getValueString();
+ public String getValueString();
/**
* @return Reader that underlies this Scanner instance.
*/
- HFile.Reader getReader();
+ public HFile.Reader getReader();
/**
* @return True is scanner has had one of the seek calls invoked; i.e.
* {@link #seekBefore(byte[])} or {@link #seekTo()} or {@link #seekTo(byte[])}.
* Otherwise returns false.
*/
- boolean isSeeked();
+ public boolean isSeeked();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/Delayable.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/Delayable.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/Delayable.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/Delayable.java Wed Jul 10 18:50:21 2013
@@ -35,17 +35,17 @@ public interface Delayable {
* should be set when ending the delay or right away. There are cases when
* the return value can be set right away, even if the call is delayed.
*/
- void startDelay(boolean delayReturnValue);
+ public void startDelay(boolean delayReturnValue);
/**
* @return is the call delayed?
*/
- boolean isDelayed();
+ public boolean isDelayed();
/**
* @return is the return value delayed?
*/
- boolean isReturnValueDelayed();
+ public boolean isReturnValueDelayed();
/**
* Signal that the RPC server is now allowed to send the response.
@@ -54,14 +54,14 @@ public interface Delayable {
* not be delayed, this parameter must be null.
* @throws IOException
*/
- void endDelay(Object result) throws IOException;
+ public void endDelay(Object result) throws IOException;
/**
* Signal the end of a delayed RPC, without specifying the return value. Use
* this only if the return value was not delayed
* @throws IOException
*/
- void endDelay() throws IOException;
+ public void endDelay() throws IOException;
/**
* End the call, throwing and exception to the caller. This works regardless
@@ -69,5 +69,5 @@ public interface Delayable {
* @param t Object to throw to the client.
* @throws IOException
*/
- void endDelayThrowing(Throwable t) throws IOException;
-}
+ public void endDelayThrowing(Throwable t) throws IOException;
+}
\ No newline at end of file
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRPCErrorHandler.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRPCErrorHandler.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRPCErrorHandler.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRPCErrorHandler.java Wed Jul 10 18:50:21 2013
@@ -31,5 +31,5 @@ public interface HBaseRPCErrorHandler {
* @param e the throwable
* @return if the server should be shut down
*/
- boolean checkOOME(final Throwable e) ;
+ public boolean checkOOME(final Throwable e) ;
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServerInterface.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServerInterface.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServerInterface.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServerInterface.java Wed Jul 10 18:50:21 2013
@@ -65,7 +65,7 @@ public interface RpcServerInterface {
*/
MetricsHBaseServer getMetrics();
- void setQosFunction(Function<Pair<RequestHeader, Message>, Integer> newFunc);
+ public void setQosFunction(Function<Pair<RequestHeader, Message>, Integer> newFunc);
/**
* Refresh autentication manager policy.
@@ -73,4 +73,4 @@ public interface RpcServerInterface {
*/
@VisibleForTesting
void refreshAuthManager(PolicyProvider pp);
-}
+}
\ No newline at end of file
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/AssignmentManager.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/AssignmentManager.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/AssignmentManager.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/AssignmentManager.java Wed Jul 10 18:50:21 2013
@@ -1145,11 +1145,11 @@ public class AssignmentManager extends Z
/**
* A specific runnable that works only on a region.
*/
- private interface RegionRunnable extends Runnable{
+ private static interface RegionRunnable extends Runnable{
/**
* @return - the name of the region it works on.
*/
- String getRegionName();
+ public String getRegionName();
}
/**
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ClusterStatusPublisher.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ClusterStatusPublisher.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ClusterStatusPublisher.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ClusterStatusPublisher.java Wed Jul 10 18:50:21 2013
@@ -220,14 +220,14 @@ public class ClusterStatusPublisher exte
}
- public interface Publisher extends Closeable {
+ public static interface Publisher extends Closeable {
- void connect(Configuration conf) throws IOException;
+ public void connect(Configuration conf) throws IOException;
- void publish(ClusterStatus cs);
+ public void publish(ClusterStatus cs);
@Override
- void close();
+ public void close();
}
public static class MulticastPublisher implements Publisher {
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/LoadBalancer.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/LoadBalancer.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/LoadBalancer.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/LoadBalancer.java Wed Jul 10 18:50:21 2013
@@ -50,21 +50,21 @@ public interface LoadBalancer extends Co
* Set the current cluster status. This allows a LoadBalancer to map host name to a server
* @param st
*/
- void setClusterStatus(ClusterStatus st);
+ public void setClusterStatus(ClusterStatus st);
/**
* Set the master service.
* @param masterServices
*/
- void setMasterServices(MasterServices masterServices);
+ public void setMasterServices(MasterServices masterServices);
/**
* Perform the major balance operation
* @param clusterState
* @return List of plans
*/
- List<RegionPlan> balanceCluster(Map<ServerName, List<HRegionInfo>> clusterState);
+ public List<RegionPlan> balanceCluster(Map<ServerName, List<HRegionInfo>> clusterState);
/**
* Perform a Round Robin assignment of regions.
@@ -72,10 +72,7 @@ public interface LoadBalancer extends Co
* @param servers
* @return Map of servername to regioninfos
*/
- Map<ServerName, List<HRegionInfo>> roundRobinAssignment(
- List<HRegionInfo> regions,
- List<ServerName> servers
- );
+ public Map<ServerName, List<HRegionInfo>> roundRobinAssignment(List<HRegionInfo> regions, List<ServerName> servers);
/**
* Assign regions to the previously hosting region server
@@ -83,10 +80,7 @@ public interface LoadBalancer extends Co
* @param servers
* @return List of plans
*/
- Map<ServerName, List<HRegionInfo>> retainAssignment(
- Map<HRegionInfo, ServerName> regions,
- List<ServerName> servers
- );
+ public Map<ServerName, List<HRegionInfo>> retainAssignment(Map<HRegionInfo, ServerName> regions, List<ServerName> servers);
/**
* Sync assign a region
@@ -94,10 +88,7 @@ public interface LoadBalancer extends Co
* @param servers
* @return Map regioninfos to servernames
*/
- Map<HRegionInfo, ServerName> immediateAssignment(
- List<HRegionInfo> regions,
- List<ServerName> servers
- );
+ public Map<HRegionInfo, ServerName> immediateAssignment(List<HRegionInfo> regions, List<ServerName> servers);
/**
* Get a random region server from the list
@@ -105,7 +96,6 @@ public interface LoadBalancer extends Co
* @param servers
* @return Servername
*/
- ServerName randomAssignment(
- HRegionInfo regionInfo, List<ServerName> servers
- );
+ public ServerName randomAssignment(HRegionInfo regionInfo,
+ List<ServerName> servers);
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java Wed Jul 10 18:50:21 2013
@@ -40,32 +40,32 @@ public interface MasterServices extends
/**
* @return Master's instance of the {@link AssignmentManager}
*/
- AssignmentManager getAssignmentManager();
+ public AssignmentManager getAssignmentManager();
/**
* @return Master's filesystem {@link MasterFileSystem} utility class.
*/
- MasterFileSystem getMasterFileSystem();
+ public MasterFileSystem getMasterFileSystem();
/**
* @return Master's {@link ServerManager} instance.
*/
- ServerManager getServerManager();
+ public ServerManager getServerManager();
/**
* @return Master's instance of {@link ExecutorService}
*/
- ExecutorService getExecutorService();
+ public ExecutorService getExecutorService();
/**
* @return Master's instance of {@link TableLockManager}
*/
- TableLockManager getTableLockManager();
+ public TableLockManager getTableLockManager();
/**
* @return Master's instance of {@link MasterCoprocessorHost}
*/
- MasterCoprocessorHost getCoprocessorHost();
+ public MasterCoprocessorHost getCoprocessorHost();
/**
* Check table is modifiable; i.e. exists and is offline.
@@ -75,7 +75,7 @@ public interface MasterServices extends
* @throws IOException
*/
// We actually throw the exceptions mentioned in the
- void checkTableModifiable(final byte[] tableName)
+ public void checkTableModifiable(final byte [] tableName)
throws IOException, TableNotFoundException, TableNotDisabledException;
/**
@@ -84,7 +84,7 @@ public interface MasterServices extends
* @param splitKeys Starting row keys for the initial table regions. If null
* a single region is created.
*/
- void createTable(HTableDescriptor desc, byte[][] splitKeys)
+ public void createTable(HTableDescriptor desc, byte [][] splitKeys)
throws IOException;
/**
@@ -92,7 +92,7 @@ public interface MasterServices extends
* @param tableName The table name
* @throws IOException
*/
- void deleteTable(final byte[] tableName) throws IOException;
+ public void deleteTable(final byte[] tableName) throws IOException;
/**
* Modify the descriptor of an existing table
@@ -100,7 +100,7 @@ public interface MasterServices extends
* @param descriptor The updated table descriptor
* @throws IOException
*/
- void modifyTable(final byte[] tableName, final HTableDescriptor descriptor)
+ public void modifyTable(final byte[] tableName, final HTableDescriptor descriptor)
throws IOException;
/**
@@ -108,14 +108,14 @@ public interface MasterServices extends
* @param tableName The table name
* @throws IOException
*/
- void enableTable(final byte[] tableName) throws IOException;
+ public void enableTable(final byte[] tableName) throws IOException;
/**
* Disable an existing table
* @param tableName The table name
* @throws IOException
*/
- void disableTable(final byte[] tableName) throws IOException;
+ public void disableTable(final byte[] tableName) throws IOException;
/**
* Add a new column to an existing table
@@ -123,7 +123,7 @@ public interface MasterServices extends
* @param column The column definition
* @throws IOException
*/
- void addColumn(final byte[] tableName, final HColumnDescriptor column)
+ public void addColumn(final byte[] tableName, final HColumnDescriptor column)
throws IOException;
/**
@@ -132,7 +132,7 @@ public interface MasterServices extends
* @param descriptor The updated column definition
* @throws IOException
*/
- void modifyColumn(byte[] tableName, HColumnDescriptor descriptor)
+ public void modifyColumn(byte[] tableName, HColumnDescriptor descriptor)
throws IOException;
/**
@@ -141,18 +141,18 @@ public interface MasterServices extends
* @param columnName The column name
* @throws IOException
*/
- void deleteColumn(final byte[] tableName, final byte[] columnName)
+ public void deleteColumn(final byte[] tableName, final byte[] columnName)
throws IOException;
/**
* @return Return table descriptors implementation.
*/
- TableDescriptors getTableDescriptors();
+ public TableDescriptors getTableDescriptors();
/**
* @return true if master enables ServerShutdownHandler;
*/
- boolean isServerShutdownHandlerEnabled();
+ public boolean isServerShutdownHandlerEnabled();
/**
* Registers a new protocol buffer {@link Service} subclass as a master coprocessor endpoint.
@@ -167,7 +167,7 @@ public interface MasterServices extends
* @return {@code true} if the registration was successful, {@code false}
* otherwise
*/
- boolean registerService(Service instance);
+ public boolean registerService(Service instance);
/**
* Merge two regions. The real implementation is on the regionserver, master
@@ -178,13 +178,12 @@ public interface MasterServices extends
* two adjacent regions
* @throws IOException
*/
- void dispatchMergingRegions(
- final HRegionInfo region_a, final HRegionInfo region_b, final boolean forcible
- ) throws IOException;
+ public void dispatchMergingRegions(final HRegionInfo region_a,
+ final HRegionInfo region_b, final boolean forcible) throws IOException;
/**
* @return true if master is initialized
*/
- boolean isInitialized();
+ public boolean isInitialized();
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SnapshotSentinel.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SnapshotSentinel.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SnapshotSentinel.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SnapshotSentinel.java Wed Jul 10 18:50:21 2013
@@ -34,30 +34,30 @@ public interface SnapshotSentinel {
* @return <tt>false</tt> if the snapshot is still in progress, <tt>true</tt> if the snapshot has
* finished
*/
- boolean isFinished();
+ public boolean isFinished();
/**
* @return -1 if the snapshot is in progress, otherwise the completion timestamp.
*/
- long getCompletionTimestamp();
+ public long getCompletionTimestamp();
/**
* Actively cancel a running snapshot.
* @param why Reason for cancellation.
*/
- void cancel(String why);
+ public void cancel(String why);
/**
* @return the description of the snapshot being run
*/
- SnapshotDescription getSnapshot();
+ public SnapshotDescription getSnapshot();
/**
* Get the exception that caused the snapshot to fail, if the snapshot has failed.
* @return {@link ForeignException} that caused the snapshot to fail, or <tt>null</tt> if the
* snapshot is still in progress or has succeeded
*/
- ForeignException getExceptionIfFailed();
+ public ForeignException getExceptionIfFailed();
/**
* Rethrow the exception returned by {@link SnapshotSentinel#getExceptionIfFailed}.
@@ -65,5 +65,5 @@ public interface SnapshotSentinel {
*
* @throws ForeignException all exceptions from remote sources are procedure exceptions
*/
- void rethrowExceptionIfFailed() throws ForeignException;
+ public void rethrowExceptionIfFailed() throws ForeignException;
}
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java Wed Jul 10 18:50:21 2013
@@ -1593,11 +1593,11 @@ public class SplitLogManager extends Zoo
* a serialization point at the end of the task processing. Must be
* restartable and idempotent.
*/
- public interface TaskFinisher {
+ static public interface TaskFinisher {
/**
* status that can be returned finish()
*/
- enum Status {
+ static public enum Status {
/**
* task completed successfully
*/
@@ -1616,7 +1616,7 @@ public class SplitLogManager extends Zoo
* @param taskname
* @return DONE if task completed successfully, ERR otherwise
*/
- Status finish(ServerName workerName, String taskname);
+ public Status finish(ServerName workerName, String taskname);
}
enum ResubmitDirective {
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/TableLockManager.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/TableLockManager.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/TableLockManager.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/TableLockManager.java Wed Jul 10 18:50:21 2013
@@ -82,20 +82,20 @@ public abstract class TableLockManager {
* A distributed lock for a table.
*/
@InterfaceAudience.Private
- public interface TableLock {
+ public static interface TableLock {
/**
* Acquire the lock, with the configured lock timeout.
* @throws LockTimeoutException If unable to acquire a lock within a specified
* time period (if any)
* @throws IOException If unrecoverable error occurs
*/
- void acquire() throws IOException;
+ public void acquire() throws IOException;
/**
* Release the lock already held.
* @throws IOException If there is an unrecoverable error releasing the lock
*/
- void release() throws IOException;
+ public void release() throws IOException;
}
/**
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/FileCleanerDelegate.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/FileCleanerDelegate.java?rev=1501909&r1=1501908&r2=1501909&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/FileCleanerDelegate.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/FileCleanerDelegate.java Wed Jul 10 18:50:21 2013
@@ -35,5 +35,5 @@ public interface FileCleanerDelegate ext
* @param fStat file status of the file to check
* @return <tt>true</tt> if the file is deletable, <tt>false</tt> if not
*/
- boolean isFileDeletable(FileStatus fStat);
+ public boolean isFileDeletable(FileStatus fStat);
}