You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by bu...@apache.org on 2015/06/14 02:45:21 UTC
[3/3] hbase git commit: HBASE-13569 Correct Javadoc (for Java8)
HBASE-13569 Correct Javadoc (for Java8)
Signed-off-by: Sean Busbey <bu...@apache.org>
Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/682b8ab8
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/682b8ab8
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/682b8ab8
Branch: refs/heads/master
Commit: 682b8ab8a542a903e5807053282693e3a96bad2d
Parents: 293506c
Author: Gábor Lipták <gl...@gmail.com>
Authored: Sun Apr 26 21:07:45 2015 -0400
Committer: Sean Busbey <bu...@apache.org>
Committed: Sat Jun 13 01:10:37 2015 -0500
----------------------------------------------------------------------
.../apache/hadoop/hbase/HColumnDescriptor.java | 6 +--
.../org/apache/hadoop/hbase/HRegionInfo.java | 7 ++-
.../apache/hadoop/hbase/HTableDescriptor.java | 2 +-
.../apache/hadoop/hbase/MetaTableAccessor.java | 4 +-
.../hbase/client/AbstractClientScanner.java | 4 +-
.../org/apache/hadoop/hbase/client/Admin.java | 12 ++---
.../client/ClientSmallReversedScanner.java | 3 +-
.../apache/hadoop/hbase/client/Durability.java | 4 +-
.../apache/hadoop/hbase/client/HBaseAdmin.java | 3 +-
.../hadoop/hbase/client/HTableInterface.java | 2 +-
.../hadoop/hbase/client/HTableMultiplexer.java | 3 --
.../org/apache/hadoop/hbase/client/Put.java | 2 +-
.../org/apache/hadoop/hbase/client/Result.java | 10 ++---
.../hadoop/hbase/client/ResultScanner.java | 2 +-
.../hadoop/hbase/client/RetryingCallable.java | 4 +-
.../RpcRetryingCallerWithReadReplicas.java | 3 +-
.../org/apache/hadoop/hbase/client/Scan.java | 11 ++---
.../client/coprocessor/AggregationClient.java | 41 ++++++++---------
.../hadoop/hbase/client/package-info.java | 5 ---
.../hbase/coprocessor/ColumnInterpreter.java | 31 ++++++-------
.../hadoop/hbase/filter/FamilyFilter.java | 7 +--
.../org/apache/hadoop/hbase/filter/Filter.java | 1 -
.../apache/hadoop/hbase/filter/FilterBase.java | 18 ++++----
.../apache/hadoop/hbase/filter/FilterList.java | 8 ++--
.../hbase/filter/MultiRowRangeFilter.java | 2 +-
.../apache/hadoop/hbase/filter/PageFilter.java | 4 +-
.../apache/hadoop/hbase/filter/SkipFilter.java | 4 +-
.../apache/hadoop/hbase/ipc/AsyncRpcClient.java | 1 -
.../apache/hadoop/hbase/ipc/ConnectionId.java | 2 +-
.../hadoop/hbase/ipc/ServerRpcController.java | 3 +-
.../hadoop/hbase/protobuf/ProtobufUtil.java | 6 +--
.../hadoop/hbase/quotas/QuotaTableUtil.java | 10 ++---
.../hadoop/hbase/regionserver/BloomType.java | 2 +-
.../hbase/replication/ReplicationPeers.java | 2 +-
.../security/visibility/CellVisibility.java | 7 +--
.../apache/hadoop/hbase/zookeeper/ZKUtil.java | 3 --
.../hbase/zookeeper/ZooKeeperWatcher.java | 1 -
.../main/java/org/apache/hadoop/hbase/Cell.java | 47 +++++++++++---------
.../org/apache/hadoop/hbase/CellComparator.java | 4 +-
.../org/apache/hadoop/hbase/CellScannable.java | 2 +-
.../org/apache/hadoop/hbase/CellScanner.java | 2 +-
.../org/apache/hadoop/hbase/HConstants.java | 3 +-
.../java/org/apache/hadoop/hbase/KeyValue.java | 27 ++++++-----
.../org/apache/hadoop/hbase/KeyValueUtil.java | 2 +-
.../org/apache/hadoop/hbase/ServerName.java | 11 ++---
.../java/org/apache/hadoop/hbase/TableName.java | 4 +-
.../hadoop/hbase/io/CellOutputStream.java | 2 +-
.../org/apache/hadoop/hbase/io/TimeRange.java | 2 +-
.../hbase/io/crypto/KeyStoreKeyProvider.java | 2 +-
.../io/hadoopbackport/ThrottledInputStream.java | 8 ++--
.../apache/hadoop/hbase/io/util/Dictionary.java | 4 +-
.../org/apache/hadoop/hbase/security/User.java | 1 -
.../org/apache/hadoop/hbase/types/Struct.java | 6 +--
.../hadoop/hbase/util/AbstractByteRange.java | 4 +-
.../apache/hadoop/hbase/util/Addressing.java | 8 ++--
.../org/apache/hadoop/hbase/util/Base64.java | 2 +-
.../org/apache/hadoop/hbase/util/ByteRange.java | 4 +-
.../org/apache/hadoop/hbase/util/Bytes.java | 8 ++--
.../org/apache/hadoop/hbase/util/ClassSize.java | 4 +-
.../hbase/util/DefaultEnvironmentEdge.java | 3 +-
.../hbase/util/IncrementingEnvironmentEdge.java | 3 +-
.../apache/hadoop/hbase/util/JenkinsHash.java | 4 +-
.../org/apache/hadoop/hbase/util/KeyLocker.java | 5 ++-
.../apache/hadoop/hbase/util/OrderedBytes.java | 9 ++--
.../org/apache/hadoop/hbase/util/Sleeper.java | 4 +-
.../apache/hadoop/hbase/metrics/BaseSource.java | 2 +-
.../hbase/codec/prefixtree/PrefixTreeCodec.java | 3 +-
.../codec/prefixtree/PrefixTreeSeeker.java | 8 ++--
.../prefixtree/decode/ArraySearcherPool.java | 3 +-
.../decode/PrefixTreeArraySearcher.java | 11 ++---
.../prefixtree/encode/PrefixTreeEncoder.java | 15 ++++---
.../encode/column/ColumnNodeWriter.java | 7 ++-
.../encode/column/ColumnSectionWriter.java | 6 ++-
.../prefixtree/encode/row/RowNodeWriter.java | 3 +-
.../prefixtree/encode/tokenize/Tokenizer.java | 2 +
.../encode/tokenize/TokenizerNode.java | 32 ++++++-------
.../codec/prefixtree/scanner/CellSearcher.java | 34 ++++++++------
.../scanner/ReversibleCellScanner.java | 4 +-
.../hadoop/hbase/util/vint/UFIntTool.java | 8 ++--
.../hadoop/hbase/util/vint/UVIntTool.java | 4 +-
.../hadoop/hbase/util/vint/UVLongTool.java | 4 +-
.../hadoop/hbase/procedure2/Procedure.java | 6 +--
.../hbase/procedure2/SequentialProcedure.java | 4 +-
.../protobuf/HBaseZeroCopyByteString.java | 6 +++
.../apache/hadoop/hbase/InterProcessLock.java | 2 +-
.../example/HFileArchiveTableMonitor.java | 2 +-
.../hbase/constraint/ConstraintException.java | 2 +-
.../hadoop/hbase/constraint/package-info.java | 6 +--
.../SplitLogManagerCoordination.java | 2 +-
.../ZKSplitLogManagerCoordination.java | 5 +--
.../coprocessor/AggregateImplementation.java | 11 +++--
.../coprocessor/MultiRowMutationEndpoint.java | 10 ++---
.../hbase/coprocessor/RegionObserver.java | 8 ++--
.../hadoop/hbase/coprocessor/package-info.java | 36 +++++++--------
.../hbase/errorhandling/TimeoutException.java | 3 +-
.../apache/hadoop/hbase/http/HttpServer.java | 13 +++---
.../apache/hadoop/hbase/http/package-info.java | 3 +-
.../org/apache/hadoop/hbase/io/FileLink.java | 2 +-
.../org/apache/hadoop/hbase/io/HFileLink.java | 1 -
.../org/apache/hadoop/hbase/io/WALLink.java | 4 +-
.../org/apache/hadoop/hbase/io/hfile/HFile.java | 5 ++-
.../hadoop/hbase/io/hfile/HFileBlock.java | 12 ++---
.../hadoop/hbase/io/hfile/HFileReaderImpl.java | 2 +-
.../hadoop/hbase/io/hfile/HFileScanner.java | 18 ++++----
.../hadoop/hbase/io/hfile/LruBlockCache.java | 2 +-
.../hbase/io/hfile/LruCachedBlockQueue.java | 2 +-
.../hadoop/hbase/io/hfile/package-info.java | 2 +-
.../apache/hadoop/hbase/ipc/RpcCallContext.java | 2 +-
.../hbase/mapreduce/MultiTableInputFormat.java | 2 +-
.../hbase/mapreduce/TableMapReduceUtil.java | 9 ++--
.../hbase/mapreduce/TableRecordReaderImpl.java | 3 +-
.../org/apache/hadoop/hbase/master/HMaster.java | 2 +-
.../hadoop/hbase/master/MasterRpcServices.java | 2 +-
.../hadoop/hbase/master/RegionStateStore.java | 2 +-
.../master/balancer/SimpleLoadBalancer.java | 4 +-
.../master/balancer/StochasticLoadBalancer.java | 4 +-
.../hbase/regionserver/ColumnTracker.java | 14 +++---
.../hbase/regionserver/CompactionRequestor.java | 4 +-
.../hbase/regionserver/DefaultMemStore.java | 2 +-
.../hbase/regionserver/DeleteTracker.java | 7 +--
.../regionserver/ExplicitColumnTracker.java | 7 +--
.../hbase/regionserver/HeapMemoryManager.java | 2 +-
...IncreasingToUpperBoundRegionSplitPolicy.java | 4 +-
.../hadoop/hbase/regionserver/LruHashMap.java | 6 +--
.../hbase/regionserver/MemStoreChunkPool.java | 2 +-
.../hbase/regionserver/MetricsRegionServer.java | 3 +-
.../MiniBatchOperationInProgress.java | 2 +-
.../NonReversedNonLazyKeyValueScanner.java | 2 +-
.../hbase/regionserver/RSRpcServices.java | 8 ++--
.../hadoop/hbase/regionserver/Region.java | 3 +-
.../hbase/regionserver/RegionScanner.java | 4 +-
.../hbase/regionserver/ScanDeleteTracker.java | 7 +--
.../apache/hadoop/hbase/regionserver/Store.java | 7 +--
.../hadoop/hbase/regionserver/StoreScanner.java | 2 +-
.../hbase/regionserver/wal/ReplayHLogKey.java | 1 -
.../hadoop/hbase/regionserver/wal/WALEdit.java | 14 +++---
.../regionserver/wal/WALEditsReplaySink.java | 4 +-
.../replication/HBaseReplicationEndpoint.java | 1 -
.../HBaseInterClusterReplicationEndpoint.java | 3 +-
.../regionserver/ReplicationSink.java | 7 +--
.../regionserver/ReplicationSource.java | 4 +-
.../regionserver/ReplicationSourceManager.java | 2 +
.../regionserver/ReplicationThrottler.java | 2 +-
.../security/access/AccessControlLists.java | 1 +
.../hbase/security/access/AccessController.java | 2 +
.../security/visibility/VisibilityUtils.java | 2 +-
.../snapshot/SnapshotDescriptionUtils.java | 14 +++---
.../hadoop/hbase/util/BloomFilterWriter.java | 2 +-
.../org/apache/hadoop/hbase/util/FSUtils.java | 9 ++--
.../org/apache/hadoop/hbase/util/HBaseFsck.java | 2 +-
.../hadoop/hbase/util/HFileV1Detector.java | 6 ++-
.../hadoop/hbase/util/MultiHConnection.java | 1 -
.../hadoop/hbase/util/RegionSplitter.java | 14 +++---
.../hbase/util/ServerRegionReplicaUtil.java | 2 +-
.../org/apache/hadoop/hbase/wal/WALKey.java | 4 +-
.../hadoop/hbase/zookeeper/ZKSplitLog.java | 8 ++--
.../hadoop/hbase/thrift/ThriftServerRunner.java | 3 +-
.../hadoop/hbase/thrift/generated/Hbase.java | 4 --
158 files changed, 510 insertions(+), 458 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java
index 20cfbef..3da018a 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java
@@ -85,8 +85,8 @@ public class HColumnDescriptor implements Comparable<HColumnDescriptor> {
/**
* Key for cache data into L1 if cache is set up with more than one tier.
* To set in the shell, do something like this:
- * <code>hbase(main):003:0> create 't',
- * {NAME => 't', CONFIGURATION => {CACHE_DATA_IN_L1 => 'true'}}</code>
+ * <code>hbase(main):003:0> create 't',
+ * {NAME => 't', CONFIGURATION => {CACHE_DATA_IN_L1 => 'true'}}</code>
*/
public static final String CACHE_DATA_IN_L1 = "CACHE_DATA_IN_L1";
@@ -115,7 +115,7 @@ public class HColumnDescriptor implements Comparable<HColumnDescriptor> {
/**
* Retain all cells across flushes and compactions even if they fall behind
* a delete tombstone. To see all retained cells, do a 'raw' scan; see
- * Scan#setRaw or pass RAW => true attribute in the shell.
+ * Scan#setRaw or pass RAW => true attribute in the shell.
*/
public static final String KEEP_DELETED_CELLS = "KEEP_DELETED_CELLS";
public static final String COMPRESS_TAGS = "COMPRESS_TAGS";
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java
index adca3d7..c134063 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java
@@ -54,17 +54,21 @@ import com.google.protobuf.InvalidProtocolBufferException;
* about the region.
*
* The region has a unique name which consists of the following fields:
+ * <ul>
* <li> tableName : The name of the table </li>
* <li> startKey : The startKey for the region. </li>
* <li> regionId : A timestamp when the region is created. </li>
* <li> replicaId : An id starting from 0 to differentiate replicas of the same region range
* but hosted in separated servers. The same region range can be hosted in multiple locations.</li>
* <li> encodedName : An MD5 encoded string for the region name.</li>
+ * </ul>
*
* <br> Other than the fields in the region name, region info contains:
+ * <ul>
* <li> endKey : the endKey for the region (exclusive) </li>
* <li> split : Whether the region is split </li>
* <li> offline : Whether the region is offline </li>
+ * </ul>
*
* In 0.98 or before, a list of table's regions would fully cover the total keyspace, and at any
* point in time, a row key always belongs to a single region, which is hosted in a single server.
@@ -647,7 +651,7 @@ public class HRegionInfo implements Comparable<HRegionInfo> {
* by this region. For example, if the region is foo,a,g and this is
* passed ["b","c"] or ["a","c"] it will return true, but if this is passed
* ["b","z"] it will return false.
- * @throws IllegalArgumentException if the range passed is invalid (ie end < start)
+ * @throws IllegalArgumentException if the range passed is invalid (ie. end < start)
*/
public boolean containsRange(byte[] rangeStartKey, byte[] rangeEndKey) {
if (Bytes.compareTo(rangeStartKey, rangeEndKey) > 0) {
@@ -1098,7 +1102,6 @@ public class HRegionInfo implements Comparable<HRegionInfo> {
* @param r Result to pull from
* @return A pair of the {@link HRegionInfo} and the {@link ServerName}
* (or null for server address if no address set in hbase:meta).
- * @throws IOException
* @deprecated use MetaTableAccessor methods for interacting with meta layouts
*/
@Deprecated
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
index f2f0077..58067ea 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
@@ -179,7 +179,7 @@ public class HTableDescriptor implements Comparable<HTableDescriptor> {
/**
* <em>INTERNAL</em> flag to indicate whether or not the memstore should be replicated
- * for read-replicas (CONSISTENCY => TIMELINE).
+ * for read-replicas (CONSISTENCY => TIMELINE).
*/
public static final String REGION_MEMSTORE_REPLICATION = "REGION_MEMSTORE_REPLICATION";
private static final Bytes REGION_MEMSTORE_REPLICATION_KEY =
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java
index c53f998..86e8d46 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java
@@ -180,7 +180,7 @@ public class MetaTableAccessor {
}
}
- /** The delimiter for meta columns for replicaIds > 0 */
+ /** The delimiter for meta columns for replicaIds > 0 */
protected static final char META_REPLICA_ID_DELIMITER = '_';
/** A regex for parsing server columns from meta. See above javadoc for meta layout */
@@ -1080,7 +1080,7 @@ public class MetaTableAccessor {
/**
* Fetch table states from META table
* @param conn connection to use
- * @return map {tableName -> state}
+ * @return map {tableName -> state}
* @throws IOException
*/
public static Map<TableName, TableState> getTableStates(Connection conn)
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AbstractClientScanner.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AbstractClientScanner.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AbstractClientScanner.java
index dc325a3..7658faf 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AbstractClientScanner.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AbstractClientScanner.java
@@ -54,11 +54,11 @@ public abstract class AbstractClientScanner implements ResultScanner {
}
/**
- * Get <param>nbRows</param> rows.
+ * Get nbRows rows.
* How many RPCs are made is determined by the {@link Scan#setCaching(int)}
* setting (or hbase.client.scanner.caching in hbase-site.xml).
* @param nbRows number of rows to return
- * @return Between zero and <param>nbRows</param> RowResults. Scan is done
+ * @return Between zero and nbRows rowResults. Scan is done
* if returned array is of zero-length (We never return null).
* @throws IOException
*/
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java
index f00d6c9..fcc0cae 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java
@@ -199,7 +199,7 @@ public interface Admin extends Abortable, Closeable {
*
* @param desc table descriptor for table
* @throws IllegalArgumentException if the table name is reserved
- * @throws MasterNotRunningException if master is not running
+ * @throws org.apache.hadoop.hbase.MasterNotRunningException if master is not running
* @throws org.apache.hadoop.hbase.TableExistsException if table already exists (If concurrent
* threads, the table may have been created between test-for-existence and attempt-at-creation).
* @throws IOException if a remote or network exception occurs
@@ -218,7 +218,7 @@ public interface Admin extends Abortable, Closeable {
* @param endKey end of key range
* @param numRegions the total number of regions to create
* @throws IllegalArgumentException if the table name is reserved
- * @throws MasterNotRunningException if master is not running
+ * @throws org.apache.hadoop.hbase.MasterNotRunningException if master is not running
* @throws org.apache.hadoop.hbase.TableExistsException if table already exists (If concurrent
* threads, the table may have been created between test-for-existence and attempt-at-creation).
* @throws IOException
@@ -235,7 +235,7 @@ public interface Admin extends Abortable, Closeable {
* @param splitKeys array of split keys for the initial regions of the table
* @throws IllegalArgumentException if the table name is reserved, if the split keys are repeated
* and if the split key has empty byte array.
- * @throws MasterNotRunningException if master is not running
+ * @throws org.apache.hadoop.hbase.MasterNotRunningException if master is not running
* @throws org.apache.hadoop.hbase.TableExistsException if table already exists (If concurrent
* threads, the table may have been created between test-for-existence and attempt-at-creation).
* @throws IOException
@@ -248,11 +248,11 @@ public interface Admin extends Abortable, Closeable {
* It may throw ExecutionException if there was an error while executing the operation
* or TimeoutException in case the wait timeout was not long enough to allow the
* operation to complete.
+ * Throws IllegalArgumentException Bad table name, if the split keys
+ * are repeated and if the split key has empty byte array.
*
* @param desc table descriptor for table
* @param splitKeys keys to check if the table has been created with all split keys
- * @throws IllegalArgumentException Bad table name, if the split keys
- * are repeated and if the split key has empty byte array.
* @throws IOException if a remote or network exception occurs
* @return the result of the async creation. You can use Future.get(long, TimeUnit)
* to wait on the operation to complete.
@@ -727,7 +727,7 @@ public interface Admin extends Abortable, Closeable {
* @param destServerName The servername of the destination regionserver. If passed the empty byte
* array we'll assign to a random server. A server name is made of host, port and startcode.
* Here is an example: <code> host187.example.com,60020,1289493121758</code>
- * @throws UnknownRegionException Thrown if we can't find a region named
+ * @throws IOException if we can't find a region named
* <code>encodedRegionName</code>
*/
void move(final byte[] encodedRegionName, final byte[] destServerName)
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientSmallReversedScanner.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientSmallReversedScanner.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientSmallReversedScanner.java
index 28502dc..5fac93a 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientSmallReversedScanner.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientSmallReversedScanner.java
@@ -39,10 +39,11 @@ import org.apache.hadoop.hbase.util.Bytes;
import com.google.common.annotations.VisibleForTesting;
/**
+ * <p>
* Client scanner for small reversed scan. Generally, only one RPC is called to fetch the
* scan results, unless the results cross multiple regions or the row count of
* results exceed the caching.
- * <p/>
+ * </p>
* For small scan, it will get better performance than {@link ReversedClientScanner}
*/
@InterfaceAudience.Private
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Durability.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Durability.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Durability.java
index bc9a4ed..9b35e04 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Durability.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Durability.java
@@ -47,13 +47,13 @@ public enum Durability {
* Write the Mutation to the WAL synchronously.
* The data is flushed to the filesystem implementation, but not necessarily to disk.
* For HDFS this will flush the data to the designated number of DataNodes.
- * See <a href="https://issues.apache.org/jira/browse/HADOOP-6313">HADOOP-6313<a/>
+ * See <a href="https://issues.apache.org/jira/browse/HADOOP-6313">HADOOP-6313</a>
*/
SYNC_WAL,
/**
* Write the Mutation to the WAL synchronously and force the entries to disk.
* (Note: this is currently not supported and will behave identical to {@link #SYNC_WAL})
- * See <a href="https://issues.apache.org/jira/browse/HADOOP-6313">HADOOP-6313<a/>
+ * See <a href="https://issues.apache.org/jira/browse/HADOOP-6313">HADOOP-6313</a>
*/
FSYNC_WAL
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java
index 7047714..a06fb2c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java
@@ -231,7 +231,8 @@ public class HBaseAdmin implements Admin {
* The connection to master will be created when required by admin functions.
*
* @param connection The Connection instance to use
- * @throws MasterNotRunningException, ZooKeeperConnectionException are not
+ * @throws MasterNotRunningException
+ * @throws ZooKeeperConnectionException are not
* thrown anymore but kept into the interface for backward api compatibility
* @deprecated Constructing HBaseAdmin objects manually has been deprecated.
* Use {@link Connection#getAdmin()} to obtain an instance of {@link Admin} instead.
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableInterface.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableInterface.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableInterface.java
index 8436307..745c770 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableInterface.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableInterface.java
@@ -126,7 +126,7 @@ public interface HTableInterface extends Table {
* Executes all the buffered {@link Put} operations.
* <p>
* This method gets called once automatically for every {@link Put} or batch
- * of {@link Put}s (when <code>put(List<Put>)</code> is used) when
+ * of {@link Put}s (when <code>put(List<Put>)</code> is used) when
* {@link #isAutoFlush} is {@code true}.
* @throws IOException if a remote or network exception occurs.
* @deprecated as of 1.0.0. Replaced by {@link BufferedMutator#flush()}
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableMultiplexer.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableMultiplexer.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableMultiplexer.java
index 10308da..b1f5b9f 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableMultiplexer.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableMultiplexer.java
@@ -121,7 +121,6 @@ public class HTableMultiplexer {
* @param tableName
* @param put
* @return true if the request can be accepted by its corresponding buffer queue.
- * @throws IOException
*/
public boolean put(TableName tableName, final Put put) {
return put(tableName, put, this.retryNum);
@@ -133,7 +132,6 @@ public class HTableMultiplexer {
* @param tableName
* @param puts
* @return the list of puts which could not be queued
- * @throws IOException
*/
public List<Put> put(TableName tableName, final List<Put> puts) {
if (puts == null)
@@ -169,7 +167,6 @@ public class HTableMultiplexer {
* retried before dropping the request.
* Return false if the queue is already full.
* @return true if the request can be accepted by its corresponding buffer queue.
- * @throws IOException
*/
public boolean put(final TableName tableName, final Put put, int retry) {
if (retry <= 0) {
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Put.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Put.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Put.java
index c895eb4..717ea3f 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Put.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Put.java
@@ -328,7 +328,7 @@ public class Put extends Mutation implements HeapSize, Comparable<Row> {
/**
* A convenience method to determine if this object's familyMap contains
- * a value assigned to the given family & qualifier.
+ * a value assigned to the given family & qualifier.
* Both given arguments must match the KeyValue object to return true.
*
* @param family column family
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Result.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Result.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Result.java
index 666069c..702983b 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Result.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Result.java
@@ -204,7 +204,7 @@ public class Result implements CellScannable, CellScanner {
/**
* Return the array of Cells backing this Result instance.
*
- * The array is sorted from smallest -> largest using the
+ * The array is sorted from smallest -> largest using the
* {@link CellComparator#COMPARATOR}.
*
* The array only contains what your Get or Scan specifies and no more.
@@ -601,7 +601,7 @@ public class Result implements CellScannable, CellScanner {
* Map of families to all versions of its qualifiers and values.
* <p>
* Returns a three level Map of the form:
- * <code>Map&family,Map<qualifier,Map<timestamp,value>>></code>
+ * <code>Map&family,Map<qualifier,Map<timestamp,value>>></code>
* <p>
* Note: All other map returning methods make use of this map internally.
* @return map from families to qualifiers to versions
@@ -643,7 +643,7 @@ public class Result implements CellScannable, CellScanner {
/**
* Map of families to their most recent qualifiers and values.
* <p>
- * Returns a two level Map of the form: <code>Map&family,Map<qualifier,value>></code>
+ * Returns a two level Map of the form: <code>Map&family,Map<qualifier,value>></code>
* <p>
* The most recent version of each qualifier will be used.
* @return map from families to qualifiers and value
@@ -675,7 +675,7 @@ public class Result implements CellScannable, CellScanner {
/**
* Map of qualifiers to values.
* <p>
- * Returns a Map of the form: <code>Map<qualifier,value></code>
+ * Returns a Map of the form: <code>Map<qualifier,value></code>
* @param family column family to get
* @return map of qualifiers to values
*/
@@ -945,4 +945,4 @@ public class Result implements CellScannable, CellScanner {
throw new UnsupportedOperationException("Attempting to modify readonly EMPTY_RESULT!");
}
}
-}
\ No newline at end of file
+}
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ResultScanner.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ResultScanner.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ResultScanner.java
index 6b7f1dd..d3efbda 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ResultScanner.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ResultScanner.java
@@ -42,7 +42,7 @@ public interface ResultScanner extends Closeable, Iterable<Result> {
/**
* @param nbRows number of rows to return
- * @return Between zero and <param>nbRows</param> Results
+ * @return Between zero and nbRows results
* @throws IOException e
*/
Result [] next(int nbRows) throws IOException;
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RetryingCallable.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RetryingCallable.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RetryingCallable.java
index e468d3c..ea65fcf 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RetryingCallable.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RetryingCallable.java
@@ -24,7 +24,7 @@ import java.io.IOException;
import org.apache.hadoop.hbase.classification.InterfaceAudience;
/**
- * A Callable<T> that will be retried. If {@link #call(int)} invocation throws exceptions,
+ * A Callable<T> that will be retried. If {@link #call(int)} invocation throws exceptions,
* we will call {@link #throwable(Throwable, boolean)} with whatever the exception was.
* @param <T>
*/
@@ -42,7 +42,7 @@ public interface RetryingCallable<T> {
* make it so we succeed on next call (clear caches, do relookup of locations, etc.).
* @param t
* @param retrying True if we are in retrying mode (we are not in retrying mode when max
- * retries == 1; we ARE in retrying mode if retries > 1 even when we are the last attempt)
+ * retries == 1; we ARE in retrying mode if retries > 1 even when we are the last attempt)
*/
void throwable(final Throwable t, boolean retrying);
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerWithReadReplicas.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerWithReadReplicas.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerWithReadReplicas.java
index 8f28881..d610d8c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerWithReadReplicas.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerWithReadReplicas.java
@@ -174,6 +174,7 @@ public class RpcRetryingCallerWithReadReplicas {
}
/**
+ * <p>
* Algo:
* - we put the query into the execution pool.
* - after x ms, if we don't have a result, we add the queries for the secondary replicas
@@ -186,7 +187,7 @@ public class RpcRetryingCallerWithReadReplicas {
* - a call is a thread. Let's not multiply the number of thread by the number of replicas.
* Server side, if we can cancel when it's still in the handler pool, it's much better, as a call
* can take some i/o.
- * <p/>
+ * </p>
* Globally, the number of retries, timeout and so on still applies, but it's per replica,
* not global. We continue until all retries are done, or all timeouts are exceeded.
*/
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Scan.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Scan.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Scan.java
index a0193fb..14b721b 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Scan.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Scan.java
@@ -727,10 +727,10 @@ public class Scan extends Query {
* this can deliver huge perf gains when there's a cf with lots of data; however, it can
* also lead to some inconsistent results, as follows:
* - if someone does a concurrent update to both column families in question you may get a row
- * that never existed, e.g. for { rowKey = 5, { cat_videos => 1 }, { video => "my cat" } }
- * someone puts rowKey 5 with { cat_videos => 0 }, { video => "my dog" }, concurrent scan
- * filtering on "cat_videos == 1" can get { rowKey = 5, { cat_videos => 1 },
- * { video => "my dog" } }.
+ * that never existed, e.g. for { rowKey = 5, { cat_videos => 1 }, { video => "my cat" } }
+ * someone puts rowKey 5 with { cat_videos => 0 }, { video => "my dog" }, concurrent scan
+ * filtering on "cat_videos == 1" can get { rowKey = 5, { cat_videos => 1 },
+ * { video => "my dog" } }.
* - if there's a concurrent split and you have more than 2 column families, some rows may be
* missing some column families.
*/
@@ -982,7 +982,6 @@ public class Scan extends Query {
return ProtobufUtil.toScanMetrics(bytes);
}
-
public Boolean isAsyncPrefetch() {
return asyncPrefetch;
}
@@ -991,6 +990,4 @@ public class Scan extends Query {
this.asyncPrefetch = asyncPrefetch;
return this;
}
-
-
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java
index 5421e57..594a459 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java
@@ -64,16 +64,17 @@ import com.google.protobuf.Message;
* <p>
* This will serve as the client side handler for invoking the aggregate
* functions.
- * <ul>
* For all aggregate functions,
- * <li>start row < end row is an essential condition (if they are not
+ * <ul>
+ * <li>start row < end row is an essential condition (if they are not
* {@link HConstants#EMPTY_BYTE_ARRAY})
* <li>Column family can't be null. In case where multiple families are
* provided, an IOException will be thrown. An optional column qualifier can
- * also be defined.
+ * also be defined.</li>
* <li>For methods to find maximum, minimum, sum, rowcount, it returns the
* parameter type. For average and std, it returns a double value. For row
- * count, it returns a long value.
+ * count, it returns a long value.</li>
+ * </ul>
* <p>Call {@link #close()} when done.
*/
@InterfaceAudience.Private
@@ -109,10 +110,10 @@ public class AggregationClient implements Closeable {
* @param tableName
* @param ci
* @param scan
- * @return max val <R>
+ * @return max val <R>
* @throws Throwable
* The caller is supposed to handle the exception as they are thrown
- * & propagated to it.
+ * & propagated to it.
*/
public <R, S, P extends Message, Q extends Message, T extends Message> R max(
final TableName tableName, final ColumnInterpreter<R, S, P, Q, T> ci, final Scan scan)
@@ -129,10 +130,10 @@ public class AggregationClient implements Closeable {
* @param table
* @param ci
* @param scan
- * @return max val <R>
+ * @return max val <>
* @throws Throwable
* The caller is supposed to handle the exception as they are thrown
- * & propagated to it.
+ * & propagated to it.
*/
public <R, S, P extends Message, Q extends Message, T extends Message>
R max(final Table table, final ColumnInterpreter<R, S, P, Q, T> ci,
@@ -199,7 +200,7 @@ public class AggregationClient implements Closeable {
* @param tableName
* @param ci
* @param scan
- * @return min val <R>
+ * @return min val <R>
* @throws Throwable
*/
public <R, S, P extends Message, Q extends Message, T extends Message> R min(
@@ -217,7 +218,7 @@ public class AggregationClient implements Closeable {
* @param table
* @param ci
* @param scan
- * @return min val <R>
+ * @return min val <R>
* @throws Throwable
*/
public <R, S, P extends Message, Q extends Message, T extends Message>
@@ -269,11 +270,11 @@ public class AggregationClient implements Closeable {
* optimised the operation. In case qualifier is provided, I can't use the
* filter as it may set the flag to skip to next row, but the value read is
* not of the given filter: in this case, this particular row will not be
- * counted ==> an error.
+ * counted ==> an error.
* @param tableName
* @param ci
* @param scan
- * @return <R, S>
+ * @return <R, S>
* @throws Throwable
*/
public <R, S, P extends Message, Q extends Message, T extends Message> long rowCount(
@@ -290,11 +291,11 @@ public class AggregationClient implements Closeable {
* optimised the operation. In case qualifier is provided, I can't use the
* filter as it may set the flag to skip to next row, but the value read is
* not of the given filter: in this case, this particular row will not be
- * counted ==> an error.
+ * counted ==> an error.
* @param table
* @param ci
* @param scan
- * @return <R, S>
+ * @return <R, S>
* @throws Throwable
*/
public <R, S, P extends Message, Q extends Message, T extends Message>
@@ -341,7 +342,7 @@ public class AggregationClient implements Closeable {
* @param tableName
* @param ci
* @param scan
- * @return sum <S>
+ * @return sum <S>
* @throws Throwable
*/
public <R, S, P extends Message, Q extends Message, T extends Message> S sum(
@@ -358,7 +359,7 @@ public class AggregationClient implements Closeable {
* @param table
* @param ci
* @param scan
- * @return sum <S>
+ * @return sum <S>
* @throws Throwable
*/
public <R, S, P extends Message, Q extends Message, T extends Message>
@@ -485,7 +486,7 @@ public class AggregationClient implements Closeable {
* @param tableName
* @param ci
* @param scan
- * @return <R, S>
+ * @return <R, S>
* @throws Throwable
*/
public <R, S, P extends Message, Q extends Message, T extends Message>
@@ -504,7 +505,7 @@ public class AggregationClient implements Closeable {
* @param table
* @param ci
* @param scan
- * @return <R, S>
+ * @return <R, S>
* @throws Throwable
*/
public <R, S, P extends Message, Q extends Message, T extends Message> double avg(
@@ -593,7 +594,7 @@ public class AggregationClient implements Closeable {
* @param tableName
* @param ci
* @param scan
- * @return <R, S>
+ * @return <R, S>
* @throws Throwable
*/
public <R, S, P extends Message, Q extends Message, T extends Message>
@@ -613,7 +614,7 @@ public class AggregationClient implements Closeable {
* @param table
* @param ci
* @param scan
- * @return <R, S>
+ * @return <R, S>
* @throws Throwable
*/
public <R, S, P extends Message, Q extends Message, T extends Message> double std(
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/client/package-info.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/package-info.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/package-info.java
index ecf4595..cf28c91 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/package-info.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/package-info.java
@@ -203,14 +203,9 @@ public class MyLittleHBaseClient {
<li><a href="http://hbase.org">HBase Home Page</a>
<li><a href="http://hadoop.apache.org/">Hadoop Home Page</a>
</ul>
-</pre></code>
-</div>
-
<p>See also the section in the HBase Reference Guide where it discusses
<a href="http://hbase.apache.org/book.html#client">HBase Client</a>. It
has section on how to access HBase from inside your multithreaded environment
how to control resources consumed client-side, etc.</p>
-</body>
-</html>
*/
package org.apache.hadoop.hbase.client;
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java
index 43efb66..e247c08 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java
@@ -35,23 +35,23 @@ import com.google.protobuf.Message;
* for an example.
* <p>
* Takes two generic parameters and three Message parameters.
- * The cell value type of the interpreter is <T>.
+ * The cell value type of the interpreter is <T>.
* During some computations like sum, average, the return type can be different
* than the cell value data type, for eg, sum of int cell values might overflow
* in case of a int result, we should use Long for its result. Therefore, this
* class mandates to use a different (promoted) data type for result of these
- * computations <S>. All computations are performed on the promoted data type
- * <S>. There is a conversion method
- * {@link ColumnInterpreter#castToReturnType(Object)} which takes a <T> type and
- * returns a <S> type.
- * The AggregateImplementation uses PB messages to initialize the
+ * computations <S>. All computations are performed on the promoted data type
+ * <S>. There is a conversion method
+ * {@link ColumnInterpreter#castToReturnType(Object)} which takes a <T> type and
+ * returns a <S> type.
+ * The AggregateIm>lementation uses PB messages to initialize the
* user's ColumnInterpreter implementation, and for sending the responses
* back to AggregationClient.
- * @param <T> Cell value data type
- * @param <S> Promoted data type
- * @param <P> PB message that is used to transport initializer specific bytes
- * @param <Q> PB message that is used to transport Cell (<T>) instance
- * @param <R> PB message that is used to transport Promoted (<S>) instance
+ * @param T Cell value data type
+ * @param S Promoted data type
+ * @param P PB message that is used to transport initializer specific bytes
+ * @param Q PB message that is used to transport Cell (<T>) instance
+ * @param R PB message that is used to transport Promoted (<S>) instance
*/
@InterfaceAudience.Private
public abstract class ColumnInterpreter<T, S, P extends Message,
@@ -109,14 +109,15 @@ Q extends Message, R extends Message> {
* This takes care if either of arguments are null. returns 0 if they are
* equal or both are null;
* <ul>
- * <li>>0 if l1 > l2 or l1 is not null and l2 is null.
- * <li>< 0 if l1 < l2 or l1 is null and l2 is not null.
+ * <li>> 0 if l1 > l2 or l1 is not null and l2 is null.</li>
+ * <li>< 0 if l1 < l2 or l1 is null and l2 is not null.</li>
+ * </ul>
*/
public abstract int compare(final T l1, final T l2);
/**
- * used for computing average of <S> data values. Not providing the divide
- * method that takes two <S> values as it is not needed as of now.
+ * used for computing average of <S> data values. Not providing the divide
+ * method that takes two <S> values as it is not needed as of now.
* @param o
* @param l
* @return Average
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java
index e79a4d5..ecead8c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java
@@ -32,15 +32,16 @@ import org.apache.hadoop.hbase.protobuf.generated.FilterProtos;
import com.google.protobuf.InvalidProtocolBufferException;
/**
+ * <p>
* This filter is used to filter based on the column family. It takes an
* operator (equal, greater, not equal, etc) and a byte [] comparator for the
* column family portion of a key.
- * <p/>
+ * </p><p>
* This filter can be wrapped with {@link org.apache.hadoop.hbase.filter.WhileMatchFilter} and {@link org.apache.hadoop.hbase.filter.SkipFilter}
* to add more control.
- * <p/>
+ * </p><p>
* Multiple filters can be combined using {@link org.apache.hadoop.hbase.filter.FilterList}.
- * <p/>
+ * </p>
* If an already known column family is looked for, use {@link org.apache.hadoop.hbase.client.Get#addFamily(byte[])}
* directly rather than a filter.
*/
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/Filter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/Filter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/Filter.java
index a2f9015..f7598d6 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/Filter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/Filter.java
@@ -261,7 +261,6 @@ public abstract class Filter {
* @param pbBytes A pb serialized {@link Filter} instance
* @return An instance of {@link Filter} made from <code>bytes</code>
* @throws DeserializationException
- * @throws IOException in case an I/O or an filter specific failure needs to be signaled.
* @see #toByteArray
*/
public static Filter parseFrom(final byte [] pbBytes) throws DeserializationException {
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterBase.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterBase.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterBase.java
index 1bcd00a..08a6821 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterBase.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterBase.java
@@ -41,7 +41,7 @@ public abstract class FilterBase extends Filter {
* Filters that are purely stateless and do nothing in their reset() methods can inherit
* this null/empty implementation.
*
- * @inheritDoc
+ * {@inheritDoc}
*/
@Override
public void reset() throws IOException {
@@ -51,7 +51,7 @@ public abstract class FilterBase extends Filter {
* Filters that do not filter by row key can inherit this implementation that
* never filters anything. (ie: returns false).
*
- * @inheritDoc
+ * {@inheritDoc}
* @deprecated As of release 2.0.0, this will be removed in HBase 3.0.0.
* Instead use {@link #filterRowKey(Cell)}
*/
@@ -72,7 +72,7 @@ public abstract class FilterBase extends Filter {
* Filters that never filter all remaining can inherit this implementation that
* never stops the filter early.
*
- * @inheritDoc
+ * {@inheritDoc}
*/
@Override
public boolean filterAllRemaining() throws IOException {
@@ -82,7 +82,7 @@ public abstract class FilterBase extends Filter {
/**
* By default no transformation takes place
*
- * @inheritDoc
+ * {@inheritDoc}
*/
@Override
public Cell transformCell(Cell v) throws IOException {
@@ -93,7 +93,7 @@ public abstract class FilterBase extends Filter {
* Filters that never filter by modifying the returned List of Cells can
* inherit this implementation that does nothing.
*
- * @inheritDoc
+ * {@inheritDoc}
*/
@Override
public void filterRowCells(List<Cell> ignored) throws IOException {
@@ -103,7 +103,7 @@ public abstract class FilterBase extends Filter {
* Fitlers that never filter by modifying the returned List of Cells can
* inherit this implementation that does nothing.
*
- * @inheritDoc
+ * {@inheritDoc}
*/
@Override
public boolean hasFilterRow() {
@@ -115,7 +115,7 @@ public abstract class FilterBase extends Filter {
* {@link #filterKeyValue(Cell)} can inherit this implementation that
* never filters a row.
*
- * @inheritDoc
+ * {@inheritDoc}
*/
@Override
public boolean filterRow() throws IOException {
@@ -126,7 +126,7 @@ public abstract class FilterBase extends Filter {
* Filters that are not sure which key must be next seeked to, can inherit
* this implementation that, by default, returns a null Cell.
*
- * @inheritDoc
+ * {@inheritDoc}
*/
public Cell getNextCellHint(Cell currentCell) throws IOException {
return null;
@@ -136,7 +136,7 @@ public abstract class FilterBase extends Filter {
* By default, we require all scan's column families to be present. Our
* subclasses may be more precise.
*
- * @inheritDoc
+ * {@inheritDoc}
*/
public boolean isFamilyEssential(byte[] name) throws IOException {
return true;
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java
index 2f89251..8ba1ccb 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java
@@ -41,14 +41,14 @@ import com.google.protobuf.InvalidProtocolBufferException;
* Since you can use Filter Lists as children of Filter Lists, you can create a
* hierarchy of filters to be evaluated.
*
- * <br/>
+ * <br>
* {@link Operator#MUST_PASS_ALL} evaluates lazily: evaluation stops as soon as one filter does
* not include the KeyValue.
*
- * <br/>
+ * <br>
* {@link Operator#MUST_PASS_ONE} evaluates non-lazily: all filters are always evaluated.
*
- * <br/>
+ * <br>
* Defaults to {@link Operator#MUST_PASS_ALL}.
*/
@InterfaceAudience.Public
@@ -315,7 +315,7 @@ final public class FilterList extends Filter {
* Filters that never filter by modifying the returned List of Cells can
* inherit this implementation that does nothing.
*
- * @inheritDoc
+ * {@inheritDoc}
*/
@Override
public void filterRowCells(List<Cell> cells) throws IOException {
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java
index e2f159b..c041914 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java
@@ -44,7 +44,7 @@ import com.google.protobuf.InvalidProtocolBufferException;
* phoenix etc. However, both solutions are inefficient. Both of them can't utilize the range info
* to perform fast forwarding during scan which is quite time consuming. If the number of ranges
* are quite big (e.g. millions), join is a proper solution though it is slow. However, there are
- * cases that user wants to specify a small number of ranges to scan (e.g. <1000 ranges). Both
+ * cases that user wants to specify a small number of ranges to scan (e.g. <1000 ranges). Both
* solutions can't provide satisfactory performance in such case. MultiRowRangeFilter is to support
* such usec ase (scan multiple row key ranges), which can construct the row key ranges from user
* specified list and perform fast-forwarding during scan. Thus, the scan will be quite efficient.
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java
index 7c68dd2..adc9c54 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java
@@ -31,11 +31,11 @@ import com.google.common.base.Preconditions;
import com.google.protobuf.InvalidProtocolBufferException;
/**
* Implementation of Filter interface that limits results to a specific page
- * size. It terminates scanning once the number of filter-passed rows is >
+ * size. It terminates scanning once the number of filter-passed rows is >
* the given page size.
* <p>
* Note that this filter cannot guarantee that the number of results returned
- * to a client are <= page size. This is because the filter is applied
+ * to a client are <= page size. This is because the filter is applied
* separately on different region servers. It does however optimize the scan of
* individual HRegions by making sure that the page size is never exceeded
* locally.
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java
index 71ea3c3..3aced13 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java
@@ -39,8 +39,9 @@ import com.google.protobuf.InvalidProtocolBufferException;
* entire row if any of its weights are zero. In this case, we want to prevent
* rows from being emitted if a single key is filtered. Combine this filter
* with a {@link ValueFilter}:
+ * </p>
* <p>
- * <pre>
+ * <code>
* scan.setFilter(new SkipFilter(new ValueFilter(CompareOp.NOT_EQUAL,
* new BinaryComparator(Bytes.toBytes(0))));
* </code>
@@ -48,6 +49,7 @@ import com.google.protobuf.InvalidProtocolBufferException;
* (since ValueFilter will not pass that Cell).
* Without this filter, the other non-zero valued columns in the row would still
* be emitted.
+ * </p>
*/
@InterfaceAudience.Public
@InterfaceStability.Stable
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncRpcClient.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncRpcClient.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncRpcClient.java
index 2e4d0a6..005f03c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncRpcClient.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncRpcClient.java
@@ -422,7 +422,6 @@ public class AsyncRpcClient extends AbstractRpcClient {
* @param rpcTimeout default rpc operation timeout
*
* @return A rpc channel that goes via this rpc client instance.
- * @throws IOException when channel could not be created
*/
public RpcChannel createRpcChannel(final ServerName sn, final User user, int rpcTimeout) {
return new RpcChannelImplementation(this, sn, user, rpcTimeout);
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ConnectionId.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ConnectionId.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ConnectionId.java
index bbd2fc7..33fc880 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ConnectionId.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ConnectionId.java
@@ -24,7 +24,7 @@ import java.net.InetSocketAddress;
/**
* This class holds the address and the user ticket, etc. The client connections
- * to servers are uniquely identified by <remoteAddress, ticket, serviceName>
+ * to servers are uniquely identified by <remoteAddress, ticket, serviceName>
*/
@InterfaceAudience.Private
public class ConnectionId {
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ServerRpcController.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ServerRpcController.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ServerRpcController.java
index 5511cb1..aa407f7 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ServerRpcController.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ServerRpcController.java
@@ -35,7 +35,8 @@ import com.google.protobuf.RpcController;
* When implementing {@link com.google.protobuf.Service} defined methods,
* coprocessor endpoints can use the following pattern to pass exceptions back to the RPC client:
* <code>
- * public void myMethod(RpcController controller, MyRequest request, RpcCallback<MyResponse> done) {
+ * public void myMethod(RpcController controller, MyRequest request,
+ * RpcCallback<MyResponse> done) {
* MyResponse response = null;
* try {
* // do processing
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java
index e816f61..a87fd47 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java
@@ -252,7 +252,7 @@ public final class ProtobufUtil {
* to flag what follows as a protobuf in hbase. Prepend these bytes to all content written to
* znodes, etc.
* @param bytes Bytes to decorate
- * @return The passed <code>bytes</codes> with magic prepended (Creates a new
+ * @return The passed <code>bytes</code> with magic prepended (Creates a new
* byte array that is <code>bytes.length</code> plus {@link ProtobufMagic#PB_MAGIC}.length.
*/
public static byte [] prependPBMagic(final byte [] bytes) {
@@ -2120,7 +2120,7 @@ public final class ProtobufUtil {
}
/**
- * Convert a ListMultimap<String, TablePermission> where key is username
+ * Convert a ListMultimap<String, TablePermission> where key is username
* to a protobuf UserPermission
*
* @param perm the list of user and table permissions
@@ -2374,7 +2374,7 @@ public final class ProtobufUtil {
/**
* Convert a protobuf UserTablePermissions to a
- * ListMultimap<String, TablePermission> where key is username.
+ * ListMultimap<String, TablePermission> where key is username.
*
* @param proto the protobuf UserPermission
* @return the converted UserPermission
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/QuotaTableUtil.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/QuotaTableUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/QuotaTableUtil.java
index 9491795..a8fec87 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/QuotaTableUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/QuotaTableUtil.java
@@ -51,11 +51,11 @@ import org.apache.hadoop.hbase.util.Strings;
* Helper class to interact with the quota table.
* <pre>
* ROW-KEY FAM/QUAL DATA
- * n.<namespace> q:s <global-quotas>
- * t.<table> q:s <global-quotas>
- * u.<user> q:s <global-quotas>
- * u.<user> q:s.<table> <table-quotas>
- * u.<user> q:s.<ns>: <namespace-quotas>
+ * n.<namespace> q:s <global-quotas>
+ * t.<table> q:s <global-quotas>
+ * u.<user> q:s <global-quotas>
+ * u.<user> q:s.<table> <table-quotas>
+ * u.<user> q:s.<ns>: <namespace-quotas>
* </pre>
*/
@InterfaceAudience.Private
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/regionserver/BloomType.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/regionserver/BloomType.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/regionserver/BloomType.java
index 75c967d..50b8b15 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/regionserver/BloomType.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/regionserver/BloomType.java
@@ -34,7 +34,7 @@ public enum BloomType {
*/
ROW,
/**
- * Bloom enabled with Table row & column (family+qualifier) as Key
+ * Bloom enabled with Table row & column (family+qualifier) as Key
*/
ROWCOL
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/replication/ReplicationPeers.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/replication/ReplicationPeers.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/replication/ReplicationPeers.java
index 359dbff..8e80e06 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/replication/ReplicationPeers.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/replication/ReplicationPeers.java
@@ -125,7 +125,7 @@ public interface ReplicationPeers {
* have to be connected. The state is read directly from the backing store.
* @param peerId a short that identifies the cluster
* @return true if replication is enabled, false otherwise.
- * @throws IOException Throws if there's an error contacting the store
+ * @throws ReplicationException thrown if there's an error contacting the store
*/
boolean getStatusOfPeerFromBackingStore(String peerId) throws ReplicationException;
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/CellVisibility.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/CellVisibility.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/CellVisibility.java
index 4968548..765559f 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/CellVisibility.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/CellVisibility.java
@@ -24,7 +24,8 @@ import org.apache.hadoop.hbase.util.Bytes;
/**
* This contains a visibility expression which can be associated with a cell. When it is set with a
* Mutation, all the cells in that mutation will get associated with this expression. A visibility
- * expression can contain visibility labels combined with logical operators AND(&), OR(|) and NOT(!)
+ * expression can contain visibility labels combined with logical
+ * operators AND(&), OR(|) and NOT(!)
*/
@InterfaceAudience.Public
@InterfaceStability.Evolving
@@ -51,7 +52,7 @@ public class CellVisibility {
/**
* Helps in quoting authentication Strings. Use this if unicode characters to
* be used in expression or special characters like '(', ')',
- * '"','\','&','|','!'
+ * '"','\','&','|','!'
*/
public static String quote(String auth) {
return quote(Bytes.toBytes(auth));
@@ -60,7 +61,7 @@ public class CellVisibility {
/**
* Helps in quoting authentication Strings. Use this if unicode characters to
* be used in expression or special characters like '(', ')',
- * '"','\','&','|','!'
+ * '"','\','&','|','!'
*/
public static String quote(byte[] auth) {
int escapeChars = 0;
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java
index 5cbb066..e71886a 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java
@@ -1206,8 +1206,6 @@ public class ZKUtil {
* @param data data of node to create
* @param cb
* @param ctx
- * @throws KeeperException if unexpected zookeeper exception
- * @throws KeeperException.NodeExistsException if node already exists
*/
public static void asyncCreate(ZooKeeperWatcher zkw,
String znode, byte [] data, final AsyncCallback.StringCallback cb,
@@ -2027,7 +2025,6 @@ public class ZKUtil {
/**
* Recursively print the current state of ZK (non-transactional)
* @param root name of the root directory in zk to print
- * @throws KeeperException
*/
public static void logZKTree(ZooKeeperWatcher zkw, String root) {
if (!LOG.isDebugEnabled()) return;
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperWatcher.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperWatcher.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperWatcher.java
index d26874a..475e385 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperWatcher.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperWatcher.java
@@ -656,7 +656,6 @@ public class ZooKeeperWatcher implements Watcher, Abortable, Closeable {
/**
* Close the connection to ZooKeeper.
*
- * @throws InterruptedException
*/
@Override
public void close() {
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-common/src/main/java/org/apache/hadoop/hbase/Cell.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/Cell.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/Cell.java
index 8f299cc..46ebeeb 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/Cell.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/Cell.java
@@ -23,7 +23,8 @@ import org.apache.hadoop.hbase.classification.InterfaceStability;
/**
- * The unit of storage in HBase consisting of the following fields:<br/>
+ * The unit of storage in HBase consisting of the following fields:
+ * <br>
* <pre>
* 1) row
* 2) column family
@@ -33,30 +34,36 @@ import org.apache.hadoop.hbase.classification.InterfaceStability;
* 6) MVCC version
* 7) value
* </pre>
- * <p/>
+ * <p>
* Uniqueness is determined by the combination of row, column family, column qualifier,
* timestamp, and type.
- * <p/>
+ * </p>
+ * <p>
* The natural comparator will perform a bitwise comparison on row, column family, and column
* qualifier. Less intuitively, it will then treat the greater timestamp as the lesser value with
* the goal of sorting newer cells first.
- * <p/>
+ * </p>
+ * <p>
* This interface should not include methods that allocate new byte[]'s such as those used in client
* or debugging code. These users should use the methods found in the {@link CellUtil} class.
* Currently for to minimize the impact of existing applications moving between 0.94 and 0.96, we
* include the costly helper methods marked as deprecated.
- * <p/>
- * Cell implements Comparable<Cell> which is only meaningful when comparing to other keys in the
+ * </p>
+ * <p>
+ * Cell implements Comparable<Cell> which is only meaningful when
+ * comparing to other keys in the
* same table. It uses CellComparator which does not work on the -ROOT- and hbase:meta tables.
- * <p/>
+ * </p>
+ * <p>
* In the future, we may consider adding a boolean isOnHeap() method and a getValueBuffer() method
* that can be used to pass a value directly from an off-heap ByteBuffer to the network without
* copying into an on-heap byte[].
- * <p/>
+ * </p>
+ * <p>
* Historic note: the original Cell implementation (KeyValue) requires that all fields be encoded as
* consecutive bytes in the same byte[], whereas this interface allows fields to reside in separate
* byte[]'s.
- * <p/>
+ * </p>
*/
@InterfaceAudience.Public
@InterfaceStability.Evolving
@@ -77,7 +84,7 @@ public interface Cell {
int getRowOffset();
/**
- * @return Number of row bytes. Must be < rowArray.length - offset.
+ * @return Number of row bytes. Must be < rowArray.length - offset.
*/
short getRowLength();
@@ -97,7 +104,7 @@ public interface Cell {
int getFamilyOffset();
/**
- * @return Number of family bytes. Must be < familyArray.length - offset.
+ * @return Number of family bytes. Must be < familyArray.length - offset.
*/
byte getFamilyLength();
@@ -117,7 +124,7 @@ public interface Cell {
int getQualifierOffset();
/**
- * @return Number of qualifier bytes. Must be < qualifierArray.length - offset.
+ * @return Number of qualifier bytes. Must be < qualifierArray.length - offset.
*/
int getQualifierLength();
@@ -148,7 +155,7 @@ public interface Cell {
* cells in the memstore but is not retained forever. It may survive several flushes, but
* generally becomes irrelevant after the cell's row is no longer involved in any operations that
* require strict consistency.
- * @return mvccVersion (always >= 0 if exists), or 0 if it no longer exists
+ * @return mvccVersion (always >= 0 if exists), or 0 if it no longer exists
*/
@Deprecated
long getMvccVersion();
@@ -158,7 +165,7 @@ public interface Cell {
* exists for cells in the memstore but is not retained forever. It will be kept for
* {@link HConstants#KEEP_SEQID_PERIOD} days, but generally becomes irrelevant after the cell's
* row is no longer involved in any operations that require strict consistency.
- * @return seqId (always > 0 if exists), or 0 if it no longer exists
+ * @return seqId (always > 0 if exists), or 0 if it no longer exists
*/
long getSequenceId();
@@ -177,7 +184,7 @@ public interface Cell {
int getValueOffset();
/**
- * @return Number of value bytes. Must be < valueArray.length - offset.
+ * @return Number of value bytes. Must be < valueArray.length - offset.
*/
int getValueLength();
@@ -199,7 +206,7 @@ public interface Cell {
/**
* WARNING do not use, expensive. This gets an arraycopy of the cell's value.
*
- * Added to ease transition from 0.94 -> 0.96.
+ * Added to ease transition from 0.94 -> 0.96.
*
* @deprecated as of 0.96, use {@link CellUtil#cloneValue(Cell)}
*/
@@ -209,7 +216,7 @@ public interface Cell {
/**
* WARNING do not use, expensive. This gets an arraycopy of the cell's family.
*
- * Added to ease transition from 0.94 -> 0.96.
+ * Added to ease transition from 0.94 -> 0.96.
*
* @deprecated as of 0.96, use {@link CellUtil#cloneFamily(Cell)}
*/
@@ -219,7 +226,7 @@ public interface Cell {
/**
* WARNING do not use, expensive. This gets an arraycopy of the cell's qualifier.
*
- * Added to ease transition from 0.94 -> 0.96.
+ * Added to ease transition from 0.94 -> 0.96.
*
* @deprecated as of 0.96, use {@link CellUtil#cloneQualifier(Cell)}
*/
@@ -229,10 +236,10 @@ public interface Cell {
/**
* WARNING do not use, expensive. this gets an arraycopy of the cell's row.
*
- * Added to ease transition from 0.94 -> 0.96.
+ * Added to ease transition from 0.94 -> 0.96.
*
* @deprecated as of 0.96, use {@link CellUtil#getRowByte(Cell, int)}
*/
@Deprecated
byte[] getRow();
-}
\ No newline at end of file
+}
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-common/src/main/java/org/apache/hadoop/hbase/CellComparator.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellComparator.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellComparator.java
index e7ccbde..2d0c940 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellComparator.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellComparator.java
@@ -111,7 +111,7 @@ public class CellComparator implements Comparator<Cell>, Serializable {
* @param b
* @param ignoreSequenceid True if we are to compare the key portion only and ignore
* the sequenceid. Set to false to compare key and consider sequenceid.
- * @return 0 if equal, -1 if a < b, and +1 if a > b.
+ * @return 0 if equal, -1 if a < b, and +1 if a > b.
*/
private final int compare(final Cell a, final Cell b, boolean ignoreSequenceid) {
// row
@@ -682,4 +682,4 @@ public class CellComparator implements Comparator<Cell>, Serializable {
return result;
}
}
-}
\ No newline at end of file
+}
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScannable.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScannable.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScannable.java
index 79e677e..0adb7b5 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScannable.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScannable.java
@@ -22,7 +22,7 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience;
/**
* Implementer can return a CellScanner over its Cell content.
* Class name is ugly but mimicing java.util.Iterable only we are about the dumber
- * CellScanner rather than say Iterator<Cell>. See CellScanner class comment for why we go
+ * CellScanner rather than say Iterator<Cell>. See CellScanner class comment for why we go
* dumber than java.util.Iterator.
*/
@InterfaceAudience.Private
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScanner.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScanner.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScanner.java
index f337122..b0460b7 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScanner.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScanner.java
@@ -35,7 +35,7 @@ import org.apache.hadoop.hbase.classification.InterfaceStability;
* may or may not point to a reusable cell implementation, so users of the CellScanner should not,
* for example, accumulate a List of Cells. All of the references may point to the same object,
* which would be the latest state of the underlying Cell. In short, the Cell is mutable.
- * <p/>
+ * </p>
* Typical usage:
*
* <pre>
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-common/src/main/java/org/apache/hadoop/hbase/HConstants.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/HConstants.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/HConstants.java
index 674ef6e..89a3f34 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/HConstants.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/HConstants.java
@@ -1133,7 +1133,8 @@ public final class HConstants {
/**
* When using bucket cache, this is a float that EITHER represents a percentage of total heap
- * memory size to give to the cache (if < 1.0) OR, it is the capacity in megabytes of the cache.
+ * memory size to give to the cache (if < 1.0) OR, it is the capacity in
+ * megabytes of the cache.
*/
public static final String BUCKET_CACHE_SIZE_KEY = "hbase.bucketcache.size";
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
index 315e9a3..cd17bef 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
@@ -64,18 +64,21 @@ import com.google.common.annotations.VisibleForTesting;
* <p>
* KeyValue wraps a byte array and takes offsets and lengths into passed array at where to start
* interpreting the content as KeyValue. The KeyValue format inside a byte array is:
- * <code><keylength> <valuelength> <key> <value></code> Key is further decomposed as:
- * <code><rowlength> <row> <columnfamilylength> <columnfamily> <columnqualifier>
- * <timestamp> <keytype></code>
+ * <code><keylength> <valuelength> <key> <value></code>
+ * Key is further decomposed as:
+ * <code><rowlength> <row> <columnfamilylength>
+ * <columnfamily> <columnqualifier>
+ * <timestamp> <keytype></code>
* The <code>rowlength</code> maximum is <code>Short.MAX_SIZE</code>, column family length maximum
- * is <code>Byte.MAX_SIZE</code>, and column qualifier + key length must be <
+ * is <code>Byte.MAX_SIZE</code>, and column qualifier + key length must be <
* <code>Integer.MAX_SIZE</code>. The column does not contain the family/qualifier delimiter,
* {@link #COLUMN_FAMILY_DELIMITER}<br>
* KeyValue can optionally contain Tags. When it contains tags, it is added in the byte array after
- * the value part. The format for this part is: <code><tagslength><tagsbytes></code>.
+ * the value part. The format for this part is: <code><tagslength><tagsbytes></code>.
* <code>tagslength</code> maximum is <code>Short.MAX_SIZE</code>. The <code>tagsbytes</code>
* contain one or more tags where as each tag is of the form
- * <code><taglength><tagtype><tagbytes></code>. <code>tagtype</code> is one byte and
+ * <code><taglength><tagtype><tagbytes></code>.
+ * <code>tagtype</code> is one byte and
* <code>taglength</code> maximum is <code>Short.MAX_SIZE</code> and it includes 1 byte type length
* and actual tag bytes length.
*/
@@ -1163,7 +1166,7 @@ public class KeyValue implements Cell, HeapSize, Cloneable, SettableSequenceId,
* as JSON. Values are left out due to their tendency to be large. If needed,
* they can be added manually.
*
- * @return the Map<String,?> containing data from this key
+ * @return the Map<String,?> containing data from this key
*/
public Map<String, Object> toStringMap() {
Map<String, Object> stringMap = new HashMap<String, Object>();
@@ -1878,7 +1881,7 @@ public class KeyValue implements Cell, HeapSize, Cloneable, SettableSequenceId,
* Compares the only the user specified portion of a Key. This is overridden by MetaComparator.
* @param left
* @param right
- * @return 0 if equal, <0 if left smaller, >0 if right smaller
+ * @return 0 if equal, <0 if left smaller, >0 if right smaller
*/
protected int compareRowKey(final Cell left, final Cell right) {
return CellComparator.COMPARATOR.compareRows(left, right);
@@ -1893,7 +1896,7 @@ public class KeyValue implements Cell, HeapSize, Cloneable, SettableSequenceId,
* @param right
* @param roffset
* @param rlength
- * @return 0 if equal, <0 if left smaller, >0 if right smaller
+ * @return 0 if equal, <0 if left smaller, >0 if right smaller
*/
public int compareFlatKey(byte[] left, int loffset, int llength,
byte[] right, int roffset, int rlength) {
@@ -2005,7 +2008,7 @@ public class KeyValue implements Cell, HeapSize, Cloneable, SettableSequenceId,
* @param right
* @param roffset
* @param rlength
- * @return 0 if equal, <0 if left smaller, >0 if right smaller
+ * @return 0 if equal, <0 if left smaller, >0 if right smaller
*/
public int compareRows(byte [] left, int loffset, int llength,
byte [] right, int roffset, int rlength) {
@@ -2054,7 +2057,7 @@ public class KeyValue implements Cell, HeapSize, Cloneable, SettableSequenceId,
* @param right
* @param roffset
* @param rlength
- * @return 0 if equal, <0 if left smaller, >0 if right smaller
+ * @return 0 if equal, <0 if left smaller, >0 if right smaller
*/
@Override // SamePrefixComparator
public int compareIgnoringPrefix(int commonPrefix, byte[] left,
@@ -2292,7 +2295,7 @@ public class KeyValue implements Cell, HeapSize, Cloneable, SettableSequenceId,
* This is a HFile block index key optimization.
* @param leftKey
* @param rightKey
- * @return 0 if equal, <0 if left smaller, >0 if right smaller
+ * @return 0 if equal, <0 if left smaller, >0 if right smaller
* @deprecated Since 0.99.2;
*/
@Deprecated
http://git-wip-us.apache.org/repos/asf/hbase/blob/682b8ab8/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueUtil.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueUtil.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueUtil.java
index 5035666..407c017 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueUtil.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueUtil.java
@@ -526,7 +526,7 @@ public class KeyValueUtil {
/*************** misc **********************************/
/**
* @param cell
- * @return <code>cell<code> if it is an instance of {@link KeyValue} else we will return a
+ * @return <code>cell</code> if it is an instance of {@link KeyValue} else we will return a
* new {@link KeyValue} instance made from <code>cell</code>
* @deprecated without any replacement.
*/