You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by zh...@apache.org on 2022/07/19 02:10:35 UTC
[hbase] branch branch-2 updated: HBASE-27220 Apply the spotless format change in HBASE-27208 to our code base
This is an automated email from the ASF dual-hosted git repository.
zhangduo pushed a commit to branch branch-2
in repository https://gitbox.apache.org/repos/asf/hbase.git
The following commit(s) were added to refs/heads/branch-2 by this push:
new 99f2ab5aa8f HBASE-27220 Apply the spotless format change in HBASE-27208 to our code base
99f2ab5aa8f is described below
commit 99f2ab5aa8f01c056d8f0902d0c67fae0e8b104e
Author: Duo Zhang <zh...@apache.org>
AuthorDate: Tue Jul 19 10:00:23 2022 +0800
HBASE-27220 Apply the spotless format change in HBASE-27208 to our code base
Signed-off-by: Andrew Purtell <ap...@apache.org>
---
.../hadoop/hbase/io/asyncfs/AsyncFSOutput.java | 4 +-
.../hadoop/hbase/util/RecoverLeaseFSUtils.java | 4 +-
.../java/org/apache/hadoop/hbase/ClusterId.java | 8 +-
.../org/apache/hadoop/hbase/ClusterMetrics.java | 28 ++---
.../org/apache/hadoop/hbase/ClusterStatus.java | 8 +-
.../hadoop/hbase/CoprocessorEnvironment.java | 14 ++-
.../apache/hadoop/hbase/HBaseServerException.java | 4 +-
.../org/apache/hadoop/hbase/HColumnDescriptor.java | 12 +--
.../java/org/apache/hadoop/hbase/HRegionInfo.java | 48 +++------
.../org/apache/hadoop/hbase/HTableDescriptor.java | 12 +--
.../org/apache/hadoop/hbase/MetaTableAccessor.java | 16 +--
.../java/org/apache/hadoop/hbase/RegionLoad.java | 8 +-
.../org/apache/hadoop/hbase/RegionMetrics.java | 92 ++++------------
.../org/apache/hadoop/hbase/ServerMetrics.java | 40 ++-----
.../main/java/org/apache/hadoop/hbase/Size.java | 4 +-
.../java/org/apache/hadoop/hbase/UserMetrics.java | 24 ++---
.../java/org/apache/hadoop/hbase/client/Admin.java | 8 +-
.../org/apache/hadoop/hbase/client/Append.java | 4 +-
.../org/apache/hadoop/hbase/client/AsyncAdmin.java | 24 ++---
.../apache/hadoop/hbase/client/CheckAndMutate.java | 36 ++-----
.../hadoop/hbase/client/CheckAndMutateResult.java | 8 +-
.../hadoop/hbase/client/ClientIdGenerator.java | 4 +-
.../hadoop/hbase/client/ClusterConnection.java | 30 ++----
.../hbase/client/ColumnFamilyDescriptor.java | 96 +++++------------
.../org/apache/hadoop/hbase/client/Connection.java | 8 +-
.../hbase/client/ConnectionRegistryFactory.java | 4 +-
.../hadoop/hbase/client/CoprocessorDescriptor.java | 14 +--
.../org/apache/hadoop/hbase/client/HBaseAdmin.java | 26 ++---
.../org/apache/hadoop/hbase/client/HTable.java | 4 +-
.../hadoop/hbase/client/HTableMultiplexer.java | 4 +-
.../org/apache/hadoop/hbase/client/Increment.java | 4 +-
.../apache/hadoop/hbase/client/MultiResponse.java | 8 +-
.../hadoop/hbase/client/MutableRegionInfo.java | 20 ++--
.../org/apache/hadoop/hbase/client/Mutation.java | 24 ++---
.../apache/hadoop/hbase/client/NonceGenerator.java | 4 +-
.../java/org/apache/hadoop/hbase/client/Query.java | 12 +--
.../hbase/client/RegionAdminServiceCallable.java | 4 +-
.../org/apache/hadoop/hbase/client/RegionInfo.java | 66 ++++--------
.../hadoop/hbase/client/RegionReplicaUtil.java | 4 +-
.../hadoop/hbase/client/RegionServerCallable.java | 8 +-
.../hadoop/hbase/client/RequestController.java | 8 +-
.../org/apache/hadoop/hbase/client/Result.java | 4 +-
.../apache/hadoop/hbase/client/ResultScanner.java | 4 +-
.../java/org/apache/hadoop/hbase/client/Row.java | 4 +-
.../org/apache/hadoop/hbase/client/RowAccess.java | 8 +-
.../apache/hadoop/hbase/client/RowMutations.java | 4 +-
.../hadoop/hbase/client/RpcRetryingCallerImpl.java | 4 +-
.../java/org/apache/hadoop/hbase/client/Scan.java | 68 +++---------
.../hadoop/hbase/client/ScannerCallable.java | 4 +-
.../hadoop/hbase/client/TableDescriptor.java | 12 +--
.../hbase/client/TableDescriptorBuilder.java | 20 +---
.../org/apache/hadoop/hbase/client/TableState.java | 28 ++---
.../hbase/client/backoff/ClientBackoffPolicy.java | 4 +-
.../exceptions/PreemptiveFastFailException.java | 20 +---
.../hadoop/hbase/filter/BigDecimalComparator.java | 4 +-
.../hadoop/hbase/filter/BinaryComparator.java | 4 +-
.../hbase/filter/BinaryComponentComparator.java | 4 +-
.../hbase/filter/BinaryPrefixComparator.java | 4 +-
.../apache/hadoop/hbase/filter/BitComparator.java | 8 +-
.../hadoop/hbase/filter/ColumnCountGetFilter.java | 4 +-
.../hbase/filter/ColumnPaginationFilter.java | 4 +-
.../hadoop/hbase/filter/ColumnPrefixFilter.java | 4 +-
.../hadoop/hbase/filter/ColumnRangeFilter.java | 28 ++---
.../hadoop/hbase/filter/ColumnValueFilter.java | 16 +--
.../apache/hadoop/hbase/filter/CompareFilter.java | 8 +-
.../hadoop/hbase/filter/DependentColumnFilter.java | 16 +--
.../apache/hadoop/hbase/filter/FamilyFilter.java | 4 +-
.../org/apache/hadoop/hbase/filter/FilterList.java | 4 +-
.../hadoop/hbase/filter/FirstKeyOnlyFilter.java | 8 +-
.../FirstKeyValueMatchingQualifiersFilter.java | 4 +-
.../apache/hadoop/hbase/filter/FuzzyRowFilter.java | 4 +-
.../hadoop/hbase/filter/InclusiveStopFilter.java | 4 +-
.../apache/hadoop/hbase/filter/KeyOnlyFilter.java | 4 +-
.../apache/hadoop/hbase/filter/LongComparator.java | 4 +-
.../hadoop/hbase/filter/MultiRowRangeFilter.java | 12 +--
.../hbase/filter/MultipleColumnPrefixFilter.java | 4 +-
.../apache/hadoop/hbase/filter/NullComparator.java | 4 +-
.../org/apache/hadoop/hbase/filter/PageFilter.java | 4 +-
.../apache/hadoop/hbase/filter/PrefixFilter.java | 4 +-
.../hadoop/hbase/filter/QualifierFilter.java | 4 +-
.../hadoop/hbase/filter/RandomRowFilter.java | 8 +-
.../hadoop/hbase/filter/RegexStringComparator.java | 4 +-
.../org/apache/hadoop/hbase/filter/RowFilter.java | 4 +-
.../filter/SingleColumnValueExcludeFilter.java | 4 +-
.../hbase/filter/SingleColumnValueFilter.java | 16 +--
.../org/apache/hadoop/hbase/filter/SkipFilter.java | 4 +-
.../hadoop/hbase/filter/SubstringComparator.java | 4 +-
.../hadoop/hbase/filter/TimestampsFilter.java | 8 +-
.../apache/hadoop/hbase/filter/ValueFilter.java | 4 +-
.../hadoop/hbase/filter/WhileMatchFilter.java | 4 +-
.../hadoop/hbase/ipc/HBaseRpcController.java | 12 +--
.../hadoop/hbase/ipc/HBaseRpcControllerImpl.java | 4 +-
.../java/org/apache/hadoop/hbase/ipc/IPCUtil.java | 8 +-
.../hbase/ipc/RemoteWithExtrasException.java | 16 +--
.../hadoop/hbase/protobuf/ProtobufMagic.java | 4 +-
.../apache/hadoop/hbase/protobuf/ProtobufUtil.java | 4 +-
.../apache/hadoop/hbase/quotas/QuotaFilter.java | 12 +--
.../hadoop/hbase/quotas/SpaceQuotaSnapshot.java | 4 +-
.../hbase/quotas/SpaceQuotaSnapshotView.java | 4 +-
.../hadoop/hbase/security/access/Permission.java | 2 +-
.../hbase/security/visibility/CellVisibility.java | 4 +-
.../hadoop/hbase/shaded/protobuf/ProtobufUtil.java | 4 +-
.../hbase/shaded/protobuf/RequestConverter.java | 4 +-
.../hbase/snapshot/HBaseSnapshotException.java | 4 +-
.../apache/hadoop/hbase/zookeeper/ZNodePaths.java | 12 +--
.../org/apache/hadoop/hbase/ArrayBackedTag.java | 16 +--
.../hadoop/hbase/ByteBufferExtendedCell.java | 40 ++-----
.../main/java/org/apache/hadoop/hbase/Cell.java | 36 ++-----
.../org/apache/hadoop/hbase/CellScannable.java | 4 +-
.../java/org/apache/hadoop/hbase/CellScanner.java | 4 +-
.../java/org/apache/hadoop/hbase/CellUtil.java | 8 +-
.../java/org/apache/hadoop/hbase/ChoreService.java | 20 +---
.../java/org/apache/hadoop/hbase/ExtendedCell.java | 12 +--
.../apache/hadoop/hbase/HBaseConfiguration.java | 4 +-
.../java/org/apache/hadoop/hbase/KeyValue.java | 94 +++++------------
.../org/apache/hadoop/hbase/PrivateCellUtil.java | 16 +--
.../main/java/org/apache/hadoop/hbase/RawCell.java | 4 +-
.../apache/hadoop/hbase/RawCellBuilderFactory.java | 4 +-
.../org/apache/hadoop/hbase/ScheduledChore.java | 24 ++---
.../java/org/apache/hadoop/hbase/ServerName.java | 4 +-
.../java/org/apache/hadoop/hbase/Stoppable.java | 4 +-
.../java/org/apache/hadoop/hbase/TableName.java | 8 +-
.../src/main/java/org/apache/hadoop/hbase/Tag.java | 16 +--
.../main/java/org/apache/hadoop/hbase/TagUtil.java | 8 +-
.../org/apache/hadoop/hbase/codec/CellCodec.java | 4 +-
.../hadoop/hbase/codec/CellCodecWithTags.java | 4 +-
.../hadoop/hbase/conf/ConfigurationManager.java | 4 +-
.../hadoop/hbase/filter/ByteArrayComparable.java | 4 +-
.../hadoop/hbase/io/ByteArrayOutputStream.java | 8 +-
.../hadoop/hbase/io/ImmutableBytesWritable.java | 4 +-
.../java/org/apache/hadoop/hbase/io/TimeRange.java | 8 +-
.../io/encoding/BufferedDataBlockEncoder.java | 4 +-
.../hadoop/hbase/io/encoding/DataBlockEncoder.java | 4 +-
.../hbase/io/encoding/DataBlockEncoding.java | 8 +-
.../hadoop/hbase/io/encoding/EncodedDataBlock.java | 2 +-
.../io/encoding/HFileBlockDecodingContext.java | 4 +-
.../io/encoding/HFileBlockEncodingContext.java | 16 +--
.../apache/hadoop/hbase/io/hfile/BlockType.java | 4 +-
.../apache/hadoop/hbase/io/hfile/HFileContext.java | 4 +-
.../java/org/apache/hadoop/hbase/nio/ByteBuff.java | 20 +---
.../org/apache/hadoop/hbase/nio/MultiByteBuff.java | 12 +--
.../org/apache/hadoop/hbase/security/User.java | 4 +-
.../apache/hadoop/hbase/security/UserProvider.java | 4 +-
.../hadoop/hbase/util/AbstractByteRange.java | 4 +-
.../org/apache/hadoop/hbase/util/ByteRange.java | 4 +-
.../java/org/apache/hadoop/hbase/util/Bytes.java | 4 +-
.../apache/hadoop/hbase/util/ExceptionUtil.java | 8 +-
.../java/org/apache/hadoop/hbase/util/HashKey.java | 4 +-
.../java/org/apache/hadoop/hbase/util/Sleeper.java | 4 +-
.../hadoop/hbase/util/WindowMovingAverage.java | 8 +-
.../hadoop/hbase/HBaseCommonTestingUtility.java | 8 +-
.../apache/hadoop/hbase/net/BoundSocketMaker.java | 4 +-
.../master/MetricsAssignmentManagerSource.java | 32 ++----
.../hadoop/hbase/master/MetricsMasterSource.java | 4 +-
.../regionserver/MetricsRegionServerWrapper.java | 46 +++-----
.../hbase/regionserver/MetricsRegionWrapper.java | 32 ++----
.../regionserver/MetricsTableWrapperAggregate.java | 24 ++---
.../metrics2/lib/DynamicMetricsRegistry.java | 4 +-
.../org/apache/hadoop/hbase/util/JSONBean.java | 4 +-
.../hadoop/hbase/IntegrationTestMobCompaction.java | 4 +-
.../org/apache/hadoop/hbase/mapred/TableSplit.java | 10 +-
.../apache/hadoop/hbase/mapreduce/CellCreator.java | 4 +-
.../apache/hadoop/hbase/PerformanceEvaluation.java | 8 +-
.../mapreduce/TestRoundRobinTableInputFormat.java | 4 +-
.../hadoop/hbase/procedure2/LockAndQueue.java | 12 +--
.../apache/hadoop/hbase/procedure2/Procedure.java | 28 ++---
.../hadoop/hbase/procedure2/ProcedureEvent.java | 2 +-
.../hadoop/hbase/procedure2/ProcedureExecutor.java | 12 +--
.../hadoop/hbase/procedure2/ProcedureMetrics.java | 10 +-
.../hbase/procedure2/ProcedureScheduler.java | 4 +-
.../hbase/procedure2/store/ProcedureStore.java | 8 +-
.../hbase/procedure2/store/wal/BitSetNode.java | 20 +---
.../store/wal/ProcedureStoreTracker.java | 4 +-
.../hadoop/hbase/procedure2/util/DelayedUtil.java | 8 +-
.../hadoop/hbase/rest/ProtobufMessageHandler.java | 4 +-
.../org/apache/hadoop/hbase/rest/RESTServlet.java | 8 +-
.../apache/hadoop/hbase/rest/TableResource.java | 6 +-
.../apache/hadoop/hbase/rest/client/Client.java | 8 +-
.../apache/hadoop/hbase/rest/client/Cluster.java | 4 +-
.../apache/hadoop/hbase/rest/client/Response.java | 20 +---
.../apache/hadoop/hbase/rest/model/CellModel.java | 16 +--
.../hadoop/hbase/rest/model/CellSetModel.java | 4 +-
.../hadoop/hbase/rest/model/ColumnSchemaModel.java | 36 ++-----
.../hbase/rest/model/NamespacesInstanceModel.java | 4 +-
.../hadoop/hbase/rest/model/NamespacesModel.java | 4 +-
.../apache/hadoop/hbase/rest/model/RowModel.java | 8 +-
.../hadoop/hbase/rest/model/ScannerModel.java | 48 +++------
.../rest/model/StorageClusterStatusModel.java | 92 ++++------------
.../rest/model/StorageClusterVersionModel.java | 4 +-
.../hadoop/hbase/rest/model/TableInfoModel.java | 8 +-
.../hadoop/hbase/rest/model/TableListModel.java | 4 +-
.../apache/hadoop/hbase/rest/model/TableModel.java | 4 +-
.../hadoop/hbase/rest/model/TableRegionModel.java | 20 +---
.../hadoop/hbase/rest/model/TableSchemaModel.java | 28 ++---
.../hadoop/hbase/rest/model/VersionModel.java | 20 +---
.../hadoop/hbase/rsgroup/RSGroupAdminServer.java | 4 +-
.../hadoop/hbase/HDFSBlocksDistribution.java | 20 +---
.../org/apache/hadoop/hbase/LocalHBaseCluster.java | 16 +--
.../hadoop/hbase/MetaRegionLocationCache.java | 4 +-
.../main/java/org/apache/hadoop/hbase/Server.java | 16 +--
.../org/apache/hadoop/hbase/TableDescriptors.java | 8 +-
.../apache/hadoop/hbase/backup/HFileArchiver.java | 16 +--
.../backup/example/TableHFileArchiveTracker.java | 4 +-
.../hadoop/hbase/client/VersionInfoUtil.java | 4 +-
.../hadoop/hbase/client/locking/EntityLock.java | 4 +-
.../coordination/SplitLogManagerCoordination.java | 16 +--
.../coordination/SplitLogWorkerCoordination.java | 8 +-
.../coordination/ZkSplitLogWorkerCoordination.java | 4 +-
.../hadoop/hbase/coprocessor/BaseEnvironment.java | 4 +-
.../hbase/coprocessor/HasMasterServices.java | 4 +-
.../hbase/coprocessor/HasRegionServerServices.java | 4 +-
.../coprocessor/MasterCoprocessorEnvironment.java | 4 +-
.../coprocessor/RegionCoprocessorEnvironment.java | 14 +--
.../RegionServerCoprocessorEnvironment.java | 8 +-
.../coprocessor/WALCoprocessorEnvironment.java | 2 +-
.../apache/hadoop/hbase/executor/EventHandler.java | 4 +-
.../hadoop/hbase/executor/ExecutorService.java | 2 +-
.../apache/hadoop/hbase/executor/ExecutorType.java | 4 +-
.../hadoop/hbase/favored/FavoredNodesPlan.java | 4 +-
.../apache/hadoop/hbase/filter/FilterWrapper.java | 4 +-
.../hadoop/hbase/io/FSDataInputStreamWrapper.java | 4 +-
.../java/org/apache/hadoop/hbase/io/FileLink.java | 12 +--
.../java/org/apache/hadoop/hbase/io/HFileLink.java | 12 +--
.../apache/hadoop/hbase/io/hfile/BlockCache.java | 8 +-
.../hadoop/hbase/io/hfile/BlockCacheKey.java | 4 +-
.../hadoop/hbase/io/hfile/BlockCacheUtil.java | 24 ++---
.../apache/hadoop/hbase/io/hfile/CacheConfig.java | 22 ++--
.../apache/hadoop/hbase/io/hfile/Cacheable.java | 4 +-
.../org/apache/hadoop/hbase/io/hfile/HFile.java | 2 +-
.../apache/hadoop/hbase/io/hfile/HFileBlock.java | 28 ++---
.../hadoop/hbase/io/hfile/HFileBlockIndex.java | 28 ++---
.../hbase/io/hfile/HFileDataBlockEncoder.java | 2 +-
.../hadoop/hbase/io/hfile/HFileReaderImpl.java | 12 +--
.../apache/hadoop/hbase/io/hfile/HFileScanner.java | 12 +--
.../hadoop/hbase/io/hfile/HFileWriterImpl.java | 4 +-
.../hadoop/hbase/io/hfile/InlineBlockWriter.java | 4 +-
.../hadoop/hbase/io/hfile/LruCachedBlock.java | 4 +-
.../hadoop/hbase/io/hfile/LruCachedBlockQueue.java | 8 +-
.../hbase/io/hfile/bucket/CachedEntryQueue.java | 8 +-
.../hadoop/hbase/io/hfile/bucket/IOEngine.java | 4 +-
.../hadoop/hbase/io/util/MemorySizeUtil.java | 4 +-
.../ipc/FastPathBalancedQueueRpcExecutor.java | 4 +-
.../org/apache/hadoop/hbase/ipc/QueueBalancer.java | 4 +-
.../java/org/apache/hadoop/hbase/ipc/RpcCall.java | 44 ++------
.../apache/hadoop/hbase/ipc/RpcCallContext.java | 12 +--
.../org/apache/hadoop/hbase/ipc/RpcServer.java | 4 +-
.../hadoop/hbase/master/ActiveMasterManager.java | 8 +-
.../org/apache/hadoop/hbase/master/HMaster.java | 48 +++------
.../hadoop/hbase/master/MasterFileSystem.java | 20 +---
.../hadoop/hbase/master/MasterRpcServices.java | 4 +-
.../apache/hadoop/hbase/master/MasterServices.java | 104 +++++-------------
.../hadoop/hbase/master/MasterWalManager.java | 6 +-
.../hbase/master/MetricsAssignmentManager.java | 32 ++----
.../apache/hadoop/hbase/master/MetricsMaster.java | 4 +-
.../hbase/master/RegionPlacementMaintainer.java | 4 +-
.../apache/hadoop/hbase/master/ServerManager.java | 18 +---
.../hadoop/hbase/master/SnapshotSentinel.java | 8 +-
.../hbase/master/assignment/AssignmentManager.java | 4 +-
.../assignment/MergeTableRegionsProcedure.java | 4 +-
.../hbase/master/assignment/RegionStates.java | 20 ++--
.../hbase/master/assignment/UnassignProcedure.java | 2 +-
.../master/balancer/FavoredStochasticBalancer.java | 4 +-
.../apache/hadoop/hbase/master/hbck/HbckChore.java | 2 +-
.../hbase/master/janitor/CatalogJanitor.java | 4 +-
.../hbase/master/janitor/CatalogJanitorReport.java | 4 +-
.../hbase/master/janitor/ReportMakingVisitor.java | 8 +-
.../hbase/master/normalizer/NormalizationPlan.java | 4 +-
.../master/normalizer/SimpleRegionNormalizer.java | 4 +-
.../AbstractStateMachineRegionProcedure.java | 4 +-
.../master/procedure/EnableTableProcedure.java | 4 +-
.../master/procedure/ServerCrashProcedure.java | 4 +-
.../master/procedure/ServerProcedureInterface.java | 8 +-
.../master/procedure/TableProcedureInterface.java | 4 +-
.../hbase/master/snapshot/SnapshotFileCache.java | 4 +-
.../hbase/master/snapshot/SnapshotManager.java | 4 +-
.../hbase/master/zksyncer/ClientZKSyncer.java | 4 +-
.../apache/hadoop/hbase/procedure/Procedure.java | 2 +-
.../hbase/procedure/ProcedureCoordinator.java | 8 +-
.../hadoop/hbase/quotas/MasterQuotaManager.java | 4 +-
.../apache/hadoop/hbase/quotas/OperationQuota.java | 2 +-
.../apache/hadoop/hbase/quotas/QuotaLimiter.java | 6 +-
.../org/apache/hadoop/hbase/quotas/QuotaState.java | 4 +-
.../apache/hadoop/hbase/quotas/RateLimiter.java | 6 +-
.../quotas/RegionServerSpaceQuotaManager.java | 4 +-
.../apache/hadoop/hbase/quotas/UserQuotaState.java | 4 +-
.../hbase/regionserver/AbstractMemStore.java | 10 +-
.../hbase/regionserver/ChangedReadersObserver.java | 4 +-
.../apache/hadoop/hbase/regionserver/Chunk.java | 4 +-
.../hadoop/hbase/regionserver/CompactSplit.java | 12 +--
.../regionserver/CompositeImmutableSegment.java | 20 +---
.../ConstantSizeRegionSplitPolicy.java | 4 +-
.../FlushNonSloppyStoresFirstPolicy.java | 4 +-
.../hadoop/hbase/regionserver/FlushPolicy.java | 4 +-
.../apache/hadoop/hbase/regionserver/HRegion.java | 78 +++++---------
.../hbase/regionserver/HRegionFileSystem.java | 14 ++-
.../hadoop/hbase/regionserver/HRegionServer.java | 60 +++--------
.../apache/hadoop/hbase/regionserver/HStore.java | 20 +---
.../hadoop/hbase/regionserver/HStoreFile.java | 4 +-
.../hbase/regionserver/HeapMemoryManager.java | 4 +-
.../IncreasingToUpperBoundRegionSplitPolicy.java | 4 +-
.../hadoop/hbase/regionserver/KeyValueHeap.java | 4 +-
.../hadoop/hbase/regionserver/KeyValueScanner.java | 4 +-
.../hadoop/hbase/regionserver/LeaseManager.java | 6 +-
.../apache/hadoop/hbase/regionserver/MemStore.java | 4 +-
.../hadoop/hbase/regionserver/MemStoreFlusher.java | 16 +--
.../hadoop/hbase/regionserver/MemStoreSizing.java | 8 +-
.../hbase/regionserver/MemStoreSnapshot.java | 16 +--
.../hbase/regionserver/MetricsUserAggregate.java | 4 +-
.../regionserver/MiniBatchOperationInProgress.java | 4 +-
.../hadoop/hbase/regionserver/RSRpcServices.java | 16 +--
.../apache/hadoop/hbase/regionserver/Region.java | 40 ++++---
.../hbase/regionserver/RegionCoprocessorHost.java | 2 +-
.../hadoop/hbase/regionserver/RegionScanner.java | 8 +-
.../hbase/regionserver/RegionScannerImpl.java | 12 +--
.../hbase/regionserver/RegionServerAccounting.java | 16 +--
.../hbase/regionserver/RegionServerServices.java | 68 +++---------
.../hbase/regionserver/RegionSplitPolicy.java | 8 +-
.../hadoop/hbase/regionserver/RowProcessor.java | 8 +-
.../hadoop/hbase/regionserver/ScannerContext.java | 12 +--
.../apache/hadoop/hbase/regionserver/Segment.java | 8 +-
.../hadoop/hbase/regionserver/SegmentScanner.java | 4 +-
.../hadoop/hbase/regionserver/SplitLogWorker.java | 4 +-
.../apache/hadoop/hbase/regionserver/Store.java | 116 ++++++---------------
.../hbase/regionserver/StoreConfigInformation.java | 8 +-
.../hadoop/hbase/regionserver/StoreEngine.java | 16 +--
.../hadoop/hbase/regionserver/StoreFile.java | 36 ++-----
.../hadoop/hbase/regionserver/StoreFileInfo.java | 18 ++--
.../hbase/regionserver/StoreFileManager.java | 4 +-
.../hbase/regionserver/StoreFlushContext.java | 4 +-
.../hadoop/hbase/regionserver/StoreScanner.java | 8 +-
.../hbase/regionserver/StripeStoreFileManager.java | 4 +-
.../hbase/regionserver/TimeRangeTracker.java | 12 +--
.../compactions/CompactionConfiguration.java | 28 ++---
.../regionserver/compactions/CompactionPolicy.java | 4 +-
.../compactions/CompactionProgress.java | 12 +--
.../compactions/CompactionRequest.java | 28 ++---
.../compactions/DateTieredCompactionPolicy.java | 2 +-
.../regionserver/compactions/OffPeakHours.java | 8 +-
.../compactions/StripeCompactionPolicy.java | 16 +--
.../regionserver/handler/OpenRegionHandler.java | 8 +-
.../regionserver/querymatcher/ColumnCount.java | 12 +--
.../regionserver/querymatcher/ColumnTracker.java | 4 +-
.../regionserver/querymatcher/DeleteTracker.java | 4 +-
.../querymatcher/ScanQueryMatcher.java | 16 +--
.../hbase/regionserver/wal/AbstractFSWAL.java | 14 +--
.../hadoop/hbase/regionserver/wal/FSHLog.java | 16 +--
.../hadoop/hbase/regionserver/wal/FSWALEntry.java | 8 +-
.../hadoop/hbase/regionserver/wal/ReaderBase.java | 16 +--
.../hbase/regionserver/wal/RingBufferTruck.java | 4 +-
.../hbase/replication/ReplicationEndpoint.java | 4 +-
.../regionserver/ReplicationSource.java | 4 +-
.../regionserver/ReplicationSourceInterface.java | 32 ++----
.../regionserver/ReplicationSourceWALReader.java | 4 +-
.../replication/regionserver/WALEntryBatch.java | 36 ++-----
.../replication/regionserver/WALEntryStream.java | 12 +--
.../hbase/security/access/AccessControlFilter.java | 4 +-
.../hadoop/hbase/security/access/AuthResult.java | 4 +-
.../hbase/security/token/FsDelegationToken.java | 8 +-
.../security/visibility/VisibilityLabelsCache.java | 4 +-
.../hbase/snapshot/RestoreSnapshotHelper.java | 20 +---
.../apache/hadoop/hbase/snapshot/SnapshotInfo.java | 40 +++----
.../org/apache/hadoop/hbase/tool/CanaryTool.java | 4 +-
.../apache/hadoop/hbase/util/BloomFilterBase.java | 12 +--
.../hadoop/hbase/util/BloomFilterFactory.java | 20 ++--
.../apache/hadoop/hbase/util/ConnectionCache.java | 4 +-
.../hadoop/hbase/util/DirectMemoryUtils.java | 8 +-
.../java/org/apache/hadoop/hbase/util/FSUtils.java | 8 +-
.../org/apache/hadoop/hbase/util/HBaseFsck.java | 16 +--
.../apache/hadoop/hbase/util/HbckTableInfo.java | 4 +-
.../apache/hadoop/hbase/util/JVMClusterUtil.java | 4 +-
.../apache/hadoop/hbase/util/RegionSplitter.java | 4 +-
.../hadoop/hbase/util/RollingStatCalculator.java | 8 +-
.../hadoop/hbase/util/ServerRegionReplicaUtil.java | 16 +--
.../hadoop/hbase/util/YammerHistogramUtils.java | 6 +-
.../hbase/util/hbck/HFileCorruptionChecker.java | 32 ++----
.../hadoop/hbase/wal/AbstractFSWALProvider.java | 4 +-
.../wal/AbstractRecoveredEditsOutputSink.java | 4 +-
.../apache/hadoop/hbase/wal/AbstractWALRoller.java | 4 +-
.../org/apache/hadoop/hbase/wal/EntryBuffers.java | 4 +-
.../org/apache/hadoop/hbase/wal/OutputSink.java | 12 +--
.../main/java/org/apache/hadoop/hbase/wal/WAL.java | 4 +-
.../java/org/apache/hadoop/hbase/wal/WALEdit.java | 4 +-
.../java/org/apache/hadoop/hbase/wal/WALKey.java | 20 +---
.../org/apache/hadoop/hbase/wal/WALKeyImpl.java | 16 ++-
.../org/apache/hadoop/hbase/wal/WALProvider.java | 4 +-
.../org/apache/hadoop/hbase/MiniClusterRule.java | 4 +-
.../org/apache/hadoop/hbase/MiniHBaseCluster.java | 16 +--
.../hadoop/hbase/StartMiniClusterOption.java | 4 +-
.../hbase/io/hfile/TestHFileDataBlockEncoder.java | 4 +-
.../assignment/TestRaceBetweenSCPAndDTP.java | 4 +-
.../hadoop/hbase/master/procedure/TestHBCKSCP.java | 4 +-
.../hadoop/hbase/regionserver/TestHRegion.java | 4 +-
.../hbase/regionserver/TestRecoveredEdits.java | 4 +-
.../regionserver/wal/TestLogRollingNoCluster.java | 4 +-
.../TestMetaRegionReplicaReplicationEndpoint.java | 8 +-
.../regionserver/TestReplicationSourceManager.java | 4 +-
.../hadoop/hbase/util/HFileArchiveTestingUtil.java | 4 +-
.../hadoop/hbase/util/MultiThreadedAction.java | 2 +-
.../hbase/util/test/LoadTestDataGenerator.java | 4 +-
.../hadoop/hbase/client/AbstractTestShell.java | 8 +-
.../org/apache/hadoop/hbase/thrift/ImplType.java | 4 +-
.../hbase/thrift/ThriftHBaseServiceHandler.java | 2 +-
.../hadoop/hbase/thrift/TestThriftServer.java | 4 +-
.../hadoop/hbase/thrift/ThriftServerRunner.java | 4 +-
.../hbase/zookeeper/ClusterStatusTracker.java | 4 +-
.../hbase/zookeeper/MiniZooKeeperCluster.java | 4 +-
.../apache/hadoop/hbase/zookeeper/ZKListener.java | 4 +-
.../org/apache/hadoop/hbase/zookeeper/ZKUtil.java | 20 +---
.../apache/hadoop/hbase/zookeeper/ZKWatcher.java | 4 +-
.../apache/hadoop/hbase/HBaseZKTestingUtility.java | 4 +-
410 files changed, 1260 insertions(+), 3464 deletions(-)
diff --git a/hbase-asyncfs/src/main/java/org/apache/hadoop/hbase/io/asyncfs/AsyncFSOutput.java b/hbase-asyncfs/src/main/java/org/apache/hadoop/hbase/io/asyncfs/AsyncFSOutput.java
index b88b32bdb81..cd186f904a9 100644
--- a/hbase-asyncfs/src/main/java/org/apache/hadoop/hbase/io/asyncfs/AsyncFSOutput.java
+++ b/hbase-asyncfs/src/main/java/org/apache/hadoop/hbase/io/asyncfs/AsyncFSOutput.java
@@ -89,8 +89,6 @@ public interface AsyncFSOutput extends Closeable {
@Override
void close() throws IOException;
- /**
- * @return byteSize success synced to underlying filesystem.
- */
+ /** Returns byteSize success synced to underlying filesystem. */
long getSyncedLength();
}
diff --git a/hbase-asyncfs/src/main/java/org/apache/hadoop/hbase/util/RecoverLeaseFSUtils.java b/hbase-asyncfs/src/main/java/org/apache/hadoop/hbase/util/RecoverLeaseFSUtils.java
index e4a410aa9c3..0014185b85c 100644
--- a/hbase-asyncfs/src/main/java/org/apache/hadoop/hbase/util/RecoverLeaseFSUtils.java
+++ b/hbase-asyncfs/src/main/java/org/apache/hadoop/hbase/util/RecoverLeaseFSUtils.java
@@ -186,9 +186,7 @@ public final class RecoverLeaseFSUtils {
return recovered;
}
- /**
- * @return Detail to append to any log message around lease recovering.
- */
+ /** Returns Detail to append to any log message around lease recovering. */
private static String getLogMessageDetail(final int nbAttempt, final Path p,
final long startWaiting) {
return "attempt=" + nbAttempt + " on file=" + p + " after "
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterId.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterId.java
index e769e80847f..9b7a5de19bd 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterId.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterId.java
@@ -45,9 +45,7 @@ public class ClusterId {
this.id = uuid;
}
- /**
- * @return The clusterid serialized using pb w/ pb magic prefix
- */
+ /** Returns The clusterid serialized using pb w/ pb magic prefix */
public byte[] toByteArray() {
return ProtobufUtil.prependPBMagic(convert().toByteArray());
}
@@ -74,9 +72,7 @@ public class ClusterId {
}
}
- /**
- * @return A pb instance to represent this instance.
- */
+ /** Returns A pb instance to represent this instance. */
public ClusterIdProtos.ClusterId convert() {
ClusterIdProtos.ClusterId.Builder builder = ClusterIdProtos.ClusterId.newBuilder();
return builder.setClusterId(this.id).build();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterMetrics.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterMetrics.java
index 769d48496af..8be97fa1fda 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterMetrics.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterMetrics.java
@@ -69,33 +69,23 @@ import org.apache.yetus.audience.InterfaceAudience;
@InterfaceAudience.Public
public interface ClusterMetrics {
- /**
- * @return the HBase version string as reported by the HMaster
- */
+ /** Returns the HBase version string as reported by the HMaster */
@Nullable
String getHBaseVersion();
- /**
- * @return the names of region servers on the dead list
- */
+ /** Returns the names of region servers on the dead list */
List<ServerName> getDeadServerNames();
- /**
- * @return the names of region servers on the live list
- */
+ /** Returns the names of region servers on the live list */
Map<ServerName, ServerMetrics> getLiveServerMetrics();
- /**
- * @return the number of regions deployed on the cluster
- */
+ /** Returns the number of regions deployed on the cluster */
default int getRegionCount() {
return getLiveServerMetrics().entrySet().stream()
.mapToInt(v -> v.getValue().getRegionMetrics().size()).sum();
}
- /**
- * @return the number of requests since last report
- */
+ /** Returns the number of requests since last report */
default long getRequestCount() {
return getLiveServerMetrics().entrySet().stream()
.flatMap(v -> v.getValue().getRegionMetrics().values().stream())
@@ -109,9 +99,7 @@ public interface ClusterMetrics {
@Nullable
ServerName getMasterName();
- /**
- * @return the names of backup masters
- */
+ /** Returns the names of backup masters */
List<ServerName> getBackupMasterNames();
@InterfaceAudience.Private
@@ -142,9 +130,7 @@ public interface ClusterMetrics {
List<ServerName> getServersName();
- /**
- * @return the average cluster load
- */
+ /** Returns the average cluster load */
default double getAverageLoad() {
int serverSize = getLiveServerMetrics().size();
if (serverSize == 0) {
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterStatus.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterStatus.java
index b202f0ec43e..e218437694b 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterStatus.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterStatus.java
@@ -107,9 +107,7 @@ public class ClusterStatus implements ClusterMetrics {
this.metrics = metrics;
}
- /**
- * @return the names of region servers on the dead list
- */
+ /** Returns the names of region servers on the dead list */
@Override
public List<ServerName> getDeadServerNames() {
return metrics.getDeadServerNames();
@@ -187,9 +185,7 @@ public class ClusterStatus implements ClusterMetrics {
return metrics.getRegionStatesInTransition();
}
- /**
- * @return the HBase version string as reported by the HMaster
- */
+ /** Returns the HBase version string as reported by the HMaster */
public String getHBaseVersion() {
return metrics.getHBaseVersion();
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/CoprocessorEnvironment.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/CoprocessorEnvironment.java
index edbc5f479d6..94909f1c14e 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/CoprocessorEnvironment.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/CoprocessorEnvironment.java
@@ -28,19 +28,19 @@ import org.apache.yetus.audience.InterfaceStability;
@InterfaceStability.Evolving
public interface CoprocessorEnvironment<C extends Coprocessor> {
- /** @return the Coprocessor interface version */
+ /** Returns the Coprocessor interface version */
int getVersion();
- /** @return the HBase version as a string (e.g. "0.21.0") */
+ /** Returns the HBase version as a string (e.g. "0.21.0") */
String getHBaseVersion();
- /** @return the loaded coprocessor instance */
+ /** Returns the loaded coprocessor instance */
C getInstance();
- /** @return the priority assigned to the loaded coprocessor */
+ /** Returns the priority assigned to the loaded coprocessor */
int getPriority();
- /** @return the load sequence number */
+ /** Returns the load sequence number */
int getLoadSequence();
/**
@@ -49,8 +49,6 @@ public interface CoprocessorEnvironment<C extends Coprocessor> {
*/
Configuration getConfiguration();
- /**
- * @return the classloader for the loaded coprocessor instance
- */
+ /** Returns the classloader for the loaded coprocessor instance */
ClassLoader getClassLoader();
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HBaseServerException.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HBaseServerException.java
index 2e4ebbd0baa..3484995c1bf 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HBaseServerException.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HBaseServerException.java
@@ -63,9 +63,7 @@ public class HBaseServerException extends HBaseIOException {
this.serverOverloaded = serverOverloaded;
}
- /**
- * @return True if server was considered overloaded when exception was thrown
- */
+ /** Returns True if server was considered overloaded when exception was thrown */
public boolean isServerOverloaded() {
return serverOverloaded;
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java
index 4694afbb0b1..d55b417d482 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java
@@ -193,17 +193,13 @@ public class HColumnDescriptor implements ColumnFamilyDescriptor, Comparable<HCo
return ColumnFamilyDescriptorBuilder.isLegalColumnFamilyName(b);
}
- /**
- * @return Name of this column family
- */
+ /** Returns Name of this column family */
@Override
public byte[] getName() {
return delegatee.getName();
}
- /**
- * @return The name string of this column family
- */
+ /** Returns The name string of this column family */
@Override
public String getNameAsString() {
return delegatee.getNameAsString();
@@ -650,9 +646,7 @@ public class HColumnDescriptor implements ColumnFamilyDescriptor, Comparable<HCo
return delegatee.toString();
}
- /**
- * @return Column family descriptor with only the customized attributes.
- */
+ /** Returns Column family descriptor with only the customized attributes. */
@Override
public String toStringCustomizedValues() {
return delegatee.toStringCustomizedValues();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java
index 1769704a992..5e9abd31a7e 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java
@@ -104,7 +104,7 @@ public class HRegionInfo implements RegionInfo {
}
/**
- * @return Return a short, printable name for this region (usually encoded name) for us logging.
+ * Returns Return a short, printable name for this region (usually encoded name) for us logging.
*/
@Override
public String getShortNameToLog() {
@@ -425,7 +425,7 @@ public class HRegionInfo implements RegionInfo {
return RegionInfo.isEncodedRegionName(regionName);
}
- /** @return the regionId */
+ /** Returns the regionId */
@Override
public long getRegionId() {
return regionId;
@@ -440,9 +440,7 @@ public class HRegionInfo implements RegionInfo {
return regionName;
}
- /**
- * @return Region name as a String for use in logging, etc.
- */
+ /** Returns Region name as a String for use in logging, etc. */
@Override
public String getRegionNameAsString() {
if (RegionInfo.hasEncodedName(this.regionName)) {
@@ -456,9 +454,7 @@ public class HRegionInfo implements RegionInfo {
return Bytes.toStringBinary(this.regionName) + "." + this.getEncodedName();
}
- /**
- * @return the encoded region name
- */
+ /** Returns the encoded region name */
@Override
public synchronized String getEncodedName() {
if (this.encodedName == null) {
@@ -475,17 +471,13 @@ public class HRegionInfo implements RegionInfo {
return this.encodedNameAsBytes;
}
- /**
- * @return the startKey
- */
+ /** Returns the startKey */
@Override
public byte[] getStartKey() {
return startKey;
}
- /**
- * @return the endKey
- */
+ /** Returns the endKey */
@Override
public byte[] getEndKey() {
return endKey;
@@ -524,40 +516,30 @@ public class HRegionInfo implements RegionInfo {
return firstKeyInRange && lastKeyInRange;
}
- /**
- * @return true if the given row falls in this region.
- */
+ /** Returns true if the given row falls in this region. */
@Override
public boolean containsRow(byte[] row) {
return Bytes.compareTo(row, startKey) >= 0
&& (Bytes.compareTo(row, endKey) < 0 || Bytes.equals(endKey, HConstants.EMPTY_BYTE_ARRAY));
}
- /**
- * @return true if this region is from hbase:meta
- */
+ /** Returns true if this region is from hbase:meta */
public boolean isMetaTable() {
return isMetaRegion();
}
- /**
- * @return true if this region is a meta region
- */
+ /** Returns true if this region is a meta region */
@Override
public boolean isMetaRegion() {
return tableName.equals(HRegionInfo.FIRST_META_REGIONINFO.getTable());
}
- /**
- * @return true if this region is from a system table
- */
+ /** Returns true if this region is from a system table */
public boolean isSystemTable() {
return tableName.isSystemTable();
}
- /**
- * @return true if has been split and has daughters.
- */
+ /** Returns true if has been split and has daughters. */
@Override
public boolean isSplit() {
return this.split;
@@ -570,9 +552,7 @@ public class HRegionInfo implements RegionInfo {
this.split = split;
}
- /**
- * @return true if this region is offline.
- */
+ /** Returns true if this region is offline. */
@Override
public boolean isOffline() {
return this.offLine;
@@ -587,9 +567,7 @@ public class HRegionInfo implements RegionInfo {
this.offLine = offLine;
}
- /**
- * @return true if this is a split parent region.
- */
+ /** Returns true if this is a split parent region. */
@Override
public boolean isSplitParent() {
if (!isSplit()) return false;
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
index bdaad3beafe..6c8ea810d03 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
@@ -160,9 +160,7 @@ public class HTableDescriptor implements TableDescriptor, Comparable<HTableDescr
return delegatee.isMetaTable();
}
- /**
- * @return Getter for fetching an unmodifiable map.
- */
+ /** Returns Getter for fetching an unmodifiable map. */
@Override
public Map<Bytes, Bytes> getValues() {
return delegatee.getValues();
@@ -525,9 +523,7 @@ public class HTableDescriptor implements TableDescriptor, Comparable<HTableDescr
return delegatee.toStringCustomizedValues();
}
- /**
- * @return map of all table attributes formatted into string.
- */
+ /** Returns map of all table attributes formatted into string. */
public String toStringTableAttributes() {
return delegatee.toStringTableAttributes();
}
@@ -611,9 +607,7 @@ public class HTableDescriptor implements TableDescriptor, Comparable<HTableDescr
return hasRegionMemStoreReplication();
}
- /**
- * @return true if the read-replicas memstore replication is enabled.
- */
+ /** Returns true if the read-replicas memstore replication is enabled. */
@Override
public boolean hasRegionMemStoreReplication() {
return delegatee.hasRegionMemStoreReplication();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java
index 543cabde279..01b5f49a205 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java
@@ -304,9 +304,7 @@ public class MetaTableAccessor {
regionInfo.getReplicaId());
}
- /**
- * @return Return the {@link HConstants#CATALOG_FAMILY} row from hbase:meta.
- */
+ /** Returns Return the {@link HConstants#CATALOG_FAMILY} row from hbase:meta. */
public static Result getCatalogFamilyRow(Connection connection, RegionInfo ri)
throws IOException {
Get get = new Get(getMetaKeyForRegion(ri));
@@ -433,9 +431,7 @@ public class MetaTableAccessor {
return false;
}
- /**
- * @return True if the column in <code>cell</code> matches the regex 'info:merge.*'.
- */
+ /** Returns True if the column in <code>cell</code> matches the regex 'info:merge.*'. */
private static boolean isMergeQualifierPrefix(Cell cell) {
// Check to see if has family and that qualifier starts with the merge qualifier 'merge'
return CellUtil.matchingFamily(cell, HConstants.CATALOG_FAMILY)
@@ -825,9 +821,7 @@ public class MetaTableAccessor {
}
}
- /**
- * @return Get closest metatable region row to passed <code>row</code>
- */
+ /** Returns Get closest metatable region row to passed <code>row</code> */
@NonNull
private static RegionInfo getClosestRegionInfo(Connection connection,
@NonNull final TableName tableName, @NonNull final byte[] row) throws IOException {
@@ -1228,9 +1222,7 @@ public class MetaTableAccessor {
abstract void add(Result r);
- /**
- * @return Collected results; wait till visits complete to collect all possible results
- */
+ /** Returns Collected results; wait till visits complete to collect all possible results */
List<T> getResults() {
return this.results;
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/RegionLoad.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/RegionLoad.java
index 6184940313d..0a762bf78a6 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/RegionLoad.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/RegionLoad.java
@@ -332,9 +332,7 @@ public class RegionLoad implements RegionMetrics {
return (int) metrics.getUncompressedStoreFileSize().get(Size.Unit.KILOBYTE);
}
- /**
- * @return the data locality of region in the regionserver.
- */
+ /** Returns the data locality of region in the regionserver. */
@Override
public float getDataLocality() {
return metrics.getDataLocality();
@@ -355,9 +353,7 @@ public class RegionLoad implements RegionMetrics {
return metrics.getLastMajorCompactionTimestamp();
}
- /**
- * @return the reference count for the stores of this region
- */
+ /** Returns the reference count for the stores of this region */
public int getStoreRefCount() {
return metrics.getStoreRefCount();
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/RegionMetrics.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/RegionMetrics.java
index 73d036e7370..645a31a8552 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/RegionMetrics.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/RegionMetrics.java
@@ -28,58 +28,38 @@ import org.apache.yetus.audience.InterfaceAudience;
@InterfaceAudience.Public
public interface RegionMetrics {
- /**
- * @return the region name
- */
+ /** Returns the region name */
byte[] getRegionName();
- /**
- * @return the number of stores
- */
+ /** Returns the number of stores */
int getStoreCount();
- /**
- * @return the number of storefiles
- */
+ /** Returns the number of storefiles */
int getStoreFileCount();
- /**
- * @return the total size of the storefiles
- */
+ /** Returns the total size of the storefiles */
Size getStoreFileSize();
- /**
- * @return the memstore size
- */
+ /** Returns the memstore size */
Size getMemStoreSize();
- /**
- * @return the number of read requests made to region
- */
+ /** Returns the number of read requests made to region */
long getReadRequestCount();
- /**
- * @return the number of write requests made to region
- */
+ /** Returns the number of write requests made to region */
long getWriteRequestCount();
- /**
- * @return the number of write requests and read requests made to region
- */
+ /** Returns the number of write requests and read requests made to region */
default long getRequestCount() {
return getReadRequestCount() + getWriteRequestCount();
}
- /**
- * @return the region name as a string
- */
+ /** Returns the region name as a string */
default String getNameAsString() {
return Bytes.toStringBinary(getRegionName());
}
- /**
- * @return the number of filtered read requests made to region
- */
+ /** Returns the number of filtered read requests made to region */
long getFilteredReadRequestCount();
/**
@@ -90,29 +70,19 @@ public interface RegionMetrics {
*/
Size getStoreFileIndexSize();
- /**
- * @return The current total size of root-level indexes for the region
- */
+ /** Returns The current total size of root-level indexes for the region */
Size getStoreFileRootLevelIndexSize();
- /**
- * @return The total size of all index blocks, not just the root level
- */
+ /** Returns The total size of all index blocks, not just the root level */
Size getStoreFileUncompressedDataIndexSize();
- /**
- * @return The total size of all Bloom filter blocks, not just loaded into the block cache
- */
+ /** Returns The total size of all Bloom filter blocks, not just loaded into the block cache */
Size getBloomFilterSize();
- /**
- * @return the total number of cells in current compaction
- */
+ /** Returns the total number of cells in current compaction */
long getCompactingCellCount();
- /**
- * @return the number of already compacted kvs in current compaction
- */
+ /** Returns the number of already compacted kvs in current compaction */
long getCompactedCellCount();
/**
@@ -121,29 +91,19 @@ public interface RegionMetrics {
*/
long getCompletedSequenceId();
- /**
- * @return completed sequence id per store.
- */
+ /** Returns completed sequence id per store. */
Map<byte[], Long> getStoreSequenceId();
- /**
- * @return the uncompressed size of the storefiles
- */
+ /** Returns the uncompressed size of the storefiles */
Size getUncompressedStoreFileSize();
- /**
- * @return the data locality of region in the regionserver.
- */
+ /** Returns the data locality of region in the regionserver. */
float getDataLocality();
- /**
- * @return the timestamp of the oldest hfile for any store of this region.
- */
+ /** Returns the timestamp of the oldest hfile for any store of this region. */
long getLastMajorCompactionTimestamp();
- /**
- * @return the reference count for the stores of this region
- */
+ /** Returns the reference count for the stores of this region */
int getStoreRefCount();
/**
@@ -158,9 +118,7 @@ public interface RegionMetrics {
*/
float getDataLocalityForSsd();
- /**
- * @return the data at local weight of this region in the regionserver
- */
+ /** Returns the data at local weight of this region in the regionserver */
long getBlocksLocalWeight();
/**
@@ -169,13 +127,9 @@ public interface RegionMetrics {
*/
long getBlocksLocalWithSsdWeight();
- /**
- * @return the block total weight of this region
- */
+ /** Returns the block total weight of this region */
long getBlocksTotalWeight();
- /**
- * @return the compaction state of this region
- */
+ /** Returns the compaction state of this region */
CompactionState getCompactionState();
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerMetrics.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerMetrics.java
index 46678ac2460..5cf190c04f9 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerMetrics.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerMetrics.java
@@ -33,38 +33,26 @@ public interface ServerMetrics {
ServerName getServerName();
- /**
- * @return the version number of a regionserver.
- */
+ /** Returns the version number of a regionserver. */
default int getVersionNumber() {
return 0;
}
- /**
- * @return the string type version of a regionserver.
- */
+ /** Returns the string type version of a regionserver. */
default String getVersion() {
return "0.0.0";
}
- /**
- * @return the number of requests per second.
- */
+ /** Returns the number of requests per second. */
long getRequestCountPerSecond();
- /**
- * @return total Number of requests from the start of the region server.
- */
+ /** Returns total Number of requests from the start of the region server. */
long getRequestCount();
- /**
- * @return the amount of used heap
- */
+ /** Returns the amount of used heap */
Size getUsedHeapSize();
- /**
- * @return the maximum allowable size of the heap
- */
+ /** Returns the maximum allowable size of the heap */
Size getMaxHeapSize();
int getInfoServerPort();
@@ -87,14 +75,10 @@ public interface ServerMetrics {
@Nullable
ReplicationLoadSink getReplicationLoadSink();
- /**
- * @return region load metrics
- */
+ /** Returns region load metrics */
Map<byte[], RegionMetrics> getRegionMetrics();
- /**
- * @return metrics per user
- */
+ /** Returns metrics per user */
Map<byte[], UserMetrics> getUserMetrics();
/**
@@ -103,14 +87,10 @@ public interface ServerMetrics {
*/
Set<String> getCoprocessorNames();
- /**
- * @return the timestamp (server side) of generating this metrics
- */
+ /** Returns the timestamp (server side) of generating this metrics */
long getReportTimestamp();
- /**
- * @return the last timestamp (server side) of generating this metrics
- */
+ /** Returns the last timestamp (server side) of generating this metrics */
long getLastReportTimestamp();
/**
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/Size.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/Size.java
index c248849e363..99062635f77 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/Size.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/Size.java
@@ -69,9 +69,7 @@ public final class Size implements Comparable<Size> {
this.unit = Preconditions.checkNotNull(unit);
}
- /**
- * @return size unit
- */
+ /** Returns size unit */
public Unit getUnit() {
return unit;
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/UserMetrics.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/UserMetrics.java
index 2710aa9be27..05108c70e74 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/UserMetrics.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/UserMetrics.java
@@ -40,19 +40,13 @@ public interface UserMetrics {
long getFilteredReadRequestsCount();
}
- /**
- * @return the user name
- */
+ /** Returns the user name */
byte[] getUserName();
- /**
- * @return the number of read requests made by user
- */
+ /** Returns the number of read requests made by user */
long getReadRequestCount();
- /**
- * @return the number of write requests made by user
- */
+ /** Returns the number of write requests made by user */
long getWriteRequestCount();
/**
@@ -63,20 +57,14 @@ public interface UserMetrics {
return getReadRequestCount() + getWriteRequestCount();
}
- /**
- * @return the user name as a string
- */
+ /** Returns the user name as a string */
default String getNameAsString() {
return Bytes.toStringBinary(getUserName());
}
- /**
- * @return metrics per client(hostname)
- */
+ /** Returns metrics per client(hostname) */
Map<String, ClientMetrics> getClientMetrics();
- /**
- * @return count of filtered read requests for a user
- */
+ /** Returns count of filtered read requests for a user */
long getFilteredReadRequests();
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java
index a0505463f72..6a0913c23c7 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java
@@ -112,9 +112,7 @@ public interface Admin extends Abortable, Closeable {
@Override
boolean isAborted();
- /**
- * @return Connection used by this object.
- */
+ /** Returns Connection used by this object. */
Connection getConnection();
/**
@@ -1710,9 +1708,7 @@ public interface Admin extends Abortable, Closeable {
List<RegionMetrics> getRegionMetrics(ServerName serverName, TableName tableName)
throws IOException;
- /**
- * @return Configuration used by the instance.
- */
+ /** Returns Configuration used by the instance. */
Configuration getConfiguration();
/**
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Append.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Append.java
index c83bc76f56b..bf1a29196db 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Append.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Append.java
@@ -90,9 +90,7 @@ public class Append extends Mutation {
return this;
}
- /**
- * @return current setting for returnResults
- */
+ /** Returns current setting for returnResults */
// This method makes public the superclasses's protected method.
@Override
public boolean isReturnResults() {
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncAdmin.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncAdmin.java
index a47ca261e03..2d6d499042e 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncAdmin.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncAdmin.java
@@ -1051,34 +1051,24 @@ public interface AsyncAdmin {
CompletableFuture<Void> recommissionRegionServer(ServerName server,
List<byte[]> encodedRegionNames);
- /**
- * @return cluster status wrapped by {@link CompletableFuture}
- */
+ /** Returns cluster status wrapped by {@link CompletableFuture} */
CompletableFuture<ClusterMetrics> getClusterMetrics();
- /**
- * @return cluster status wrapped by {@link CompletableFuture}
- */
+ /** Returns cluster status wrapped by {@link CompletableFuture} */
CompletableFuture<ClusterMetrics> getClusterMetrics(EnumSet<Option> options);
- /**
- * @return current master server name wrapped by {@link CompletableFuture}
- */
+ /** Returns current master server name wrapped by {@link CompletableFuture} */
default CompletableFuture<ServerName> getMaster() {
return getClusterMetrics(EnumSet.of(Option.MASTER)).thenApply(ClusterMetrics::getMasterName);
}
- /**
- * @return current backup master list wrapped by {@link CompletableFuture}
- */
+ /** Returns current backup master list wrapped by {@link CompletableFuture} */
default CompletableFuture<Collection<ServerName>> getBackupMasters() {
return getClusterMetrics(EnumSet.of(Option.BACKUP_MASTERS))
.thenApply(ClusterMetrics::getBackupMasterNames);
}
- /**
- * @return current live region servers list wrapped by {@link CompletableFuture}
- */
+ /** Returns current live region servers list wrapped by {@link CompletableFuture} */
default CompletableFuture<Collection<ServerName>> getRegionServers() {
return getClusterMetrics(EnumSet.of(Option.SERVERS_NAME))
.thenApply(ClusterMetrics::getServersName);
@@ -1110,9 +1100,7 @@ public interface AsyncAdmin {
return future;
}
- /**
- * @return a list of master coprocessors wrapped by {@link CompletableFuture}
- */
+ /** Returns a list of master coprocessors wrapped by {@link CompletableFuture} */
default CompletableFuture<List<String>> getMasterCoprocessorNames() {
return getClusterMetrics(EnumSet.of(Option.MASTER_COPROCESSORS))
.thenApply(ClusterMetrics::getMasterCoprocessorNames);
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CheckAndMutate.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CheckAndMutate.java
index 9202e310e7f..e8949ee65ec 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CheckAndMutate.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CheckAndMutate.java
@@ -250,9 +250,7 @@ public final class CheckAndMutate implements Row {
this.action = action;
}
- /**
- * @return the row
- */
+ /** Returns the row */
@Override
public byte[] getRow() {
return row;
@@ -281,58 +279,42 @@ public final class CheckAndMutate implements Row {
return Bytes.hashCode(this.getRow());
}
- /**
- * @return the family to check
- */
+ /** Returns the family to check */
public byte[] getFamily() {
return family;
}
- /**
- * @return the qualifier to check
- */
+ /** Returns the qualifier to check */
public byte[] getQualifier() {
return qualifier;
}
- /**
- * @return the comparison operator
- */
+ /** Returns the comparison operator */
public CompareOperator getCompareOp() {
return op;
}
- /**
- * @return the expected value
- */
+ /** Returns the expected value */
public byte[] getValue() {
return value;
}
- /**
- * @return the filter to check
- */
+ /** Returns the filter to check */
public Filter getFilter() {
return filter;
}
- /**
- * @return whether this has a filter or not
- */
+ /** Returns whether this has a filter or not */
public boolean hasFilter() {
return filter != null;
}
- /**
- * @return the time range to check
- */
+ /** Returns the time range to check */
public TimeRange getTimeRange() {
return timeRange;
}
- /**
- * @return the action done if check succeeds
- */
+ /** Returns the action done if check succeeds */
public Row getAction() {
return action;
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CheckAndMutateResult.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CheckAndMutateResult.java
index 88b438a74a0..8ecb49e3d5f 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CheckAndMutateResult.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CheckAndMutateResult.java
@@ -32,16 +32,12 @@ public class CheckAndMutateResult {
this.result = result;
}
- /**
- * @return Whether the CheckAndMutate operation is successful or not
- */
+ /** Returns Whether the CheckAndMutate operation is successful or not */
public boolean isSuccess() {
return success;
}
- /**
- * @return It is used only for CheckAndMutate operations with Increment/Append. Otherwise null
- */
+ /** Returns It is used only for CheckAndMutate operations with Increment/Append. Otherwise null */
public Result getResult() {
return result;
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientIdGenerator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientIdGenerator.java
index 758cf508578..f5bedbe88d2 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientIdGenerator.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientIdGenerator.java
@@ -58,9 +58,7 @@ final class ClientIdGenerator {
return id;
}
- /**
- * @return PID of the current process, if it can be extracted from JVM name, or null.
- */
+ /** Returns PID of the current process, if it can be extracted from JVM name, or null. */
public static Long getPid() {
String name = ManagementFactory.getRuntimeMXBean().getName();
String[] nameParts = name.split("@");
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClusterConnection.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClusterConnection.java
index c7fde3d9214..849745b89eb 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClusterConnection.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClusterConnection.java
@@ -246,13 +246,11 @@ public interface ClusterConnection extends Connection {
void clearCaches(final ServerName sn);
/**
- * @return Nonce generator for this ClusterConnection; may be null if disabled in configuration.
+ * Returns Nonce generator for this ClusterConnection; may be null if disabled in configuration.
*/
NonceGenerator getNonceGenerator();
- /**
- * @return Default AsyncProcess associated with this connection.
- */
+ /** Returns Default AsyncProcess associated with this connection. */
AsyncProcess getAsyncProcess();
/**
@@ -263,34 +261,22 @@ public interface ClusterConnection extends Connection {
*/
RpcRetryingCallerFactory getNewRpcRetryingCallerFactory(Configuration conf);
- /**
- * @return Connection's RpcRetryingCallerFactory instance
- */
+ /** Returns Connection's RpcRetryingCallerFactory instance */
RpcRetryingCallerFactory getRpcRetryingCallerFactory();
- /**
- * @return Connection's RpcControllerFactory instance
- */
+ /** Returns Connection's RpcControllerFactory instance */
RpcControllerFactory getRpcControllerFactory();
- /**
- * @return a ConnectionConfiguration object holding parsed configuration values
- */
+ /** Returns a ConnectionConfiguration object holding parsed configuration values */
ConnectionConfiguration getConnectionConfiguration();
- /**
- * @return the current statistics tracker associated with this connection
- */
+ /** Returns the current statistics tracker associated with this connection */
ServerStatisticTracker getStatisticsTracker();
- /**
- * @return the configured client backoff policy
- */
+ /** Returns the configured client backoff policy */
ClientBackoffPolicy getBackoffPolicy();
- /**
- * @return the MetricsConnection instance associated with this connection.
- */
+ /** Returns the MetricsConnection instance associated with this connection. */
MetricsConnection getConnectionMetrics();
/**
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ColumnFamilyDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ColumnFamilyDescriptor.java
index 5f11f8d258a..b3d8e4fd600 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ColumnFamilyDescriptor.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ColumnFamilyDescriptor.java
@@ -77,39 +77,25 @@ public interface ColumnFamilyDescriptor {
return lcf.getConfiguration().hashCode() - rcf.getConfiguration().hashCode();
};
- /**
- * @return The storefile/hfile blocksize for this column family.
- */
+ /** Returns The storefile/hfile blocksize for this column family. */
int getBlocksize();
- /**
- * @return bloom filter type used for new StoreFiles in ColumnFamily
- */
+ /** Returns bloom filter type used for new StoreFiles in ColumnFamily */
BloomType getBloomFilterType();
- /**
- * @return Compression type setting.
- */
+ /** Returns Compression type setting. */
Compression.Algorithm getCompactionCompressionType();
- /**
- * @return Compression type setting for major compactions.
- */
+ /** Returns Compression type setting for major compactions. */
Compression.Algorithm getMajorCompactionCompressionType();
- /**
- * @return Compression type setting for minor compactions.
- */
+ /** Returns Compression type setting for minor compactions. */
Compression.Algorithm getMinorCompactionCompressionType();
- /**
- * @return Compression type setting.
- */
+ /** Returns Compression type setting. */
Compression.Algorithm getCompressionType();
- /**
- * @return an unmodifiable map.
- */
+ /** Returns an unmodifiable map. */
Map<String, String> getConfiguration();
/**
@@ -118,24 +104,16 @@ public interface ColumnFamilyDescriptor {
*/
String getConfigurationValue(String key);
- /**
- * @return replication factor set for this CF
- */
+ /** Returns replication factor set for this CF */
short getDFSReplication();
- /**
- * @return the data block encoding algorithm used in block cache and optionally on disk
- */
+ /** Returns the data block encoding algorithm used in block cache and optionally on disk */
DataBlockEncoding getDataBlockEncoding();
- /**
- * @return Return the raw crypto key attribute for the family, or null if not set
- */
+ /** Returns Return the raw crypto key attribute for the family, or null if not set */
byte[] getEncryptionKey();
- /**
- * @return Return the encryption algorithm in use by this family
- */
+ /** Returns Return the encryption algorithm in use by this family */
String getEncryptionType();
/**
@@ -144,19 +122,13 @@ public interface ColumnFamilyDescriptor {
*/
MemoryCompactionPolicy getInMemoryCompaction();
- /**
- * @return return the KeepDeletedCells
- */
+ /** Returns return the KeepDeletedCells */
KeepDeletedCells getKeepDeletedCells();
- /**
- * @return maximum number of versions
- */
+ /** Returns maximum number of versions */
int getMaxVersions();
- /**
- * @return The minimum number of versions to keep.
- */
+ /** Returns The minimum number of versions to keep. */
int getMinVersions();
/**
@@ -171,19 +143,13 @@ public interface ColumnFamilyDescriptor {
*/
long getMobThreshold();
- /**
- * @return a copy of Name of this column family
- */
+ /** Returns a copy of Name of this column family */
byte[] getName();
- /**
- * @return Name of this column family
- */
+ /** Returns Name of this column family */
String getNameAsString();
- /**
- * @return the scope tag
- */
+ /** Returns the scope tag */
int getScope();
/**
@@ -193,9 +159,7 @@ public interface ColumnFamilyDescriptor {
*/
String getStoragePolicy();
- /**
- * @return Time-to-live of cell contents, in seconds.
- */
+ /** Returns Time-to-live of cell contents, in seconds. */
int getTimeToLive();
/**
@@ -228,19 +192,13 @@ public interface ColumnFamilyDescriptor {
*/
boolean isBlockCacheEnabled();
- /**
- * @return true if we should cache bloomfilter blocks on write
- */
+ /** Returns true if we should cache bloomfilter blocks on write */
boolean isCacheBloomsOnWrite();
- /**
- * @return true if we should cache data blocks on write
- */
+ /** Returns true if we should cache data blocks on write */
boolean isCacheDataOnWrite();
- /**
- * @return true if we should cache index blocks on write
- */
+ /** Returns true if we should cache index blocks on write */
boolean isCacheIndexesOnWrite();
/**
@@ -249,9 +207,7 @@ public interface ColumnFamilyDescriptor {
*/
boolean isCompressTags();
- /**
- * @return true if we should evict cached blocks from the blockcache on close
- */
+ /** Returns true if we should evict cached blocks from the blockcache on close */
boolean isEvictBlocksOnClose();
/**
@@ -266,14 +222,10 @@ public interface ColumnFamilyDescriptor {
*/
boolean isMobEnabled();
- /**
- * @return true if we should prefetch blocks into the blockcache on open
- */
+ /** Returns true if we should prefetch blocks into the blockcache on open */
boolean isPrefetchBlocksOnOpen();
- /**
- * @return Column family descriptor with only the customized attributes.
- */
+ /** Returns Column family descriptor with only the customized attributes. */
String toStringCustomizedValues();
/**
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Connection.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Connection.java
index 4a8e47567ea..7db973d3f28 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Connection.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Connection.java
@@ -58,9 +58,7 @@ public interface Connection extends Abortable, Closeable {
* general methods.
*/
- /**
- * @return Configuration instance being used by this Connection instance.
- */
+ /** Returns Configuration instance being used by this Connection instance. */
Configuration getConfiguration();
/**
@@ -173,9 +171,7 @@ public interface Connection extends Abortable, Closeable {
*/
TableBuilder getTableBuilder(TableName tableName, ExecutorService pool);
- /**
- * @return the cluster ID unique to this HBase cluster.
- */
+ /** Returns the cluster ID unique to this HBase cluster. */
String getClusterId();
/**
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionRegistryFactory.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionRegistryFactory.java
index faaf61a4f52..bba75aabfc1 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionRegistryFactory.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionRegistryFactory.java
@@ -32,9 +32,7 @@ final class ConnectionRegistryFactory {
private ConnectionRegistryFactory() {
}
- /**
- * @return The connection registry implementation to use.
- */
+ /** Returns The connection registry implementation to use. */
static ConnectionRegistry getRegistry(Configuration conf) {
Class<? extends ConnectionRegistry> clazz =
conf.getClass(CLIENT_CONNECTION_REGISTRY_IMPL_CONF_KEY, ZKConnectionRegistry.class,
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CoprocessorDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CoprocessorDescriptor.java
index 3331c810700..40cb5abfa96 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CoprocessorDescriptor.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CoprocessorDescriptor.java
@@ -28,23 +28,17 @@ import org.apache.yetus.audience.InterfaceAudience;
*/
@InterfaceAudience.Public
public interface CoprocessorDescriptor {
- /**
- * @return the name of the class or interface represented by this object.
- */
+ /** Returns the name of the class or interface represented by this object. */
String getClassName();
/**
- * @return Path of the jar file. If it's null, the class will be loaded from default classloader.
+ * Returns Path of the jar file. If it's null, the class will be loaded from default classloader.
*/
Optional<String> getJarPath();
- /**
- * @return The order to execute this coprocessor
- */
+ /** Returns The order to execute this coprocessor */
int getPriority();
- /**
- * @return Arbitrary key-value parameter pairs passed into the coprocessor.
- */
+ /** Returns Arbitrary key-value parameter pairs passed into the coprocessor. */
Map<String, String> getProperties();
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java
index ca2cf11d3b2..90b499e416c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java
@@ -483,7 +483,7 @@ public class HBaseAdmin implements Admin {
}
}
- /** @return Connection used by this object. */
+ /** Returns Connection used by this object. */
@Override
public Connection getConnection() {
return connection;
@@ -3531,28 +3531,20 @@ public class HBaseAdmin implements Admin {
return getDescription();
}
- /**
- * @return the table name
- */
+ /** Returns the table name */
protected TableName getTableName() {
return tableName;
}
- /**
- * @return the table descriptor
- */
+ /** Returns the table descriptor */
protected TableDescriptor getTableDescriptor() throws IOException {
return getAdmin().getDescriptor(getTableName());
}
- /**
- * @return the operation type like CREATE, DELETE, DISABLE etc.
- */
+ /** Returns the operation type like CREATE, DELETE, DISABLE etc. */
public abstract String getOperationType();
- /**
- * @return a description of the operation
- */
+ /** Returns a description of the operation */
protected String getDescription() {
return "Operation: " + getOperationType() + ", " + "Table Name: "
+ tableName.getNameWithNamespaceInclAsString() + ", procId: " + procId;
@@ -3696,16 +3688,12 @@ public class HBaseAdmin implements Admin {
this.namespaceName = namespaceName;
}
- /**
- * @return the namespace name
- */
+ /** Returns the namespace name */
protected String getNamespaceName() {
return namespaceName;
}
- /**
- * @return the operation type like CREATE_NAMESPACE, DELETE_NAMESPACE, etc.
- */
+ /** Returns the operation type like CREATE_NAMESPACE, DELETE_NAMESPACE, etc. */
public abstract String getOperationType();
@Override
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java
index fcc928b5c63..acb5c9fda5d 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java
@@ -204,9 +204,7 @@ public class HTable implements Table {
this.locator = new HRegionLocator(tableName, connection);
}
- /**
- * @return maxKeyValueSize from configuration.
- */
+ /** Returns maxKeyValueSize from configuration. */
public static int getMaxKeyValueSize(Configuration conf) {
return conf.getInt(ConnectionConfiguration.MAX_KEYVALUE_SIZE_KEY, -1);
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableMultiplexer.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableMultiplexer.java
index fac13a831a6..7404d460b55 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableMultiplexer.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableMultiplexer.java
@@ -216,9 +216,7 @@ public class HTableMultiplexer {
return put(TableName.valueOf(tableName), put);
}
- /**
- * @return the current HTableMultiplexerStatus
- */
+ /** Returns the current HTableMultiplexerStatus */
public HTableMultiplexerStatus getHTableMultiplexerStatus() {
return new HTableMultiplexerStatus(serverToFlushWorkerMap);
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Increment.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Increment.java
index 2a9bec60329..0efd71e69d2 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Increment.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Increment.java
@@ -162,9 +162,7 @@ public class Increment extends Mutation {
return this;
}
- /**
- * @return current setting for returnResults
- */
+ /** Returns current setting for returnResults */
// This method makes public the superclasses's protected method.
@Override
public boolean isReturnResults() {
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/MultiResponse.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/MultiResponse.java
index 97a01e1bb6e..b768b76fdb6 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/MultiResponse.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/MultiResponse.java
@@ -44,9 +44,7 @@ public class MultiResponse extends AbstractResponse {
super();
}
- /**
- * @return Number of pairs in this container
- */
+ /** Returns Number of pairs in this container */
public int size() {
int size = 0;
for (RegionResult result : results.values()) {
@@ -68,9 +66,7 @@ public class MultiResponse extends AbstractResponse {
exceptions.put(regionName, ie);
}
- /**
- * @return the exception for the region, if any. Null otherwise.
- */
+ /** Returns the exception for the region, if any. Null otherwise. */
public Throwable getException(byte[] regionName) {
return exceptions.get(regionName);
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/MutableRegionInfo.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/MutableRegionInfo.java
index 22114f8f624..7627f8dc0e5 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/MutableRegionInfo.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/MutableRegionInfo.java
@@ -127,14 +127,14 @@ class MutableRegionInfo implements RegionInfo {
}
/**
- * @return Return a short, printable name for this region (usually encoded name) for us logging.
+ * Returns Return a short, printable name for this region (usually encoded name) for us logging.
*/
@Override
public String getShortNameToLog() {
return RegionInfo.prettyPrint(this.getEncodedName());
}
- /** @return the regionId */
+ /** Returns the regionId */
@Override
public long getRegionId() {
return regionId;
@@ -149,15 +149,13 @@ class MutableRegionInfo implements RegionInfo {
return regionName;
}
- /**
- * @return Region name as a String for use in logging, etc.
- */
+ /** Returns Region name as a String for use in logging, etc. */
@Override
public String getRegionNameAsString() {
return RegionInfo.getRegionNameAsString(this, this.regionName);
}
- /** @return the encoded region name */
+ /** Returns the encoded region name */
@Override
public String getEncodedName() {
return this.encodedName;
@@ -168,13 +166,13 @@ class MutableRegionInfo implements RegionInfo {
return this.encodedNameAsBytes;
}
- /** @return the startKey */
+ /** Returns the startKey */
@Override
public byte[] getStartKey() {
return startKey;
}
- /** @return the endKey */
+ /** Returns the endKey */
@Override
public byte[] getEndKey() {
return endKey;
@@ -219,15 +217,13 @@ class MutableRegionInfo implements RegionInfo {
|| Bytes.equals(endKey, HConstants.EMPTY_BYTE_ARRAY));
}
- /** @return true if this region is a meta region */
+ /** Returns true if this region is a meta region */
@Override
public boolean isMetaRegion() {
return tableName.equals(TableName.META_TABLE_NAME);
}
- /**
- * @return True if has been split and has daughters.
- */
+ /** Returns True if has been split and has daughters. */
@Override
public boolean isSplit() {
return this.split;
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Mutation.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Mutation.java
index 3c22b053a0d..408b02e44fa 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Mutation.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Mutation.java
@@ -354,9 +354,7 @@ public abstract class Mutation extends OperationWithAttributes
return this;
}
- /**
- * @return the set of clusterIds that have consumed the mutation
- */
+ /** Returns the set of clusterIds that have consumed the mutation */
public List<UUID> getClusterIds() {
List<UUID> clusterIds = new ArrayList<>();
byte[] bytes = getAttribute(CONSUMED_CLUSTER_IDS);
@@ -379,9 +377,7 @@ public abstract class Mutation extends OperationWithAttributes
return this;
}
- /**
- * @return CellVisibility associated with cells in this Mutation. n
- */
+ /** Returns CellVisibility associated with cells in this Mutation. n */
public CellVisibility getCellVisibility() throws DeserializationException {
byte[] cellVisibilityBytes = this.getAttribute(VisibilityConstants.VISIBILITY_LABELS_ATTR_KEY);
if (cellVisibilityBytes == null) return null;
@@ -437,16 +433,12 @@ public abstract class Mutation extends OperationWithAttributes
return size;
}
- /**
- * @return the number of different families
- */
+ /** Returns the number of different families */
public int numFamilies() {
return getFamilyCellMap().size();
}
- /**
- * @return Calculate what Mutation adds to class heap size.
- */
+ /** Returns Calculate what Mutation adds to class heap size. */
@Override
public long heapSize() {
long heapsize = MUTATION_OVERHEAD;
@@ -475,9 +467,7 @@ public abstract class Mutation extends OperationWithAttributes
return ClassSize.align(heapsize);
}
- /**
- * @return The serialized ACL for this operation, or null if none
- */
+ /** Returns The serialized ACL for this operation, or null if none */
public byte[] getACL() {
return getAttribute(AccessControlConstants.OP_ATTRIBUTE_ACL);
}
@@ -527,9 +517,7 @@ public abstract class Mutation extends OperationWithAttributes
return this;
}
- /**
- * @return current value for returnResults
- */
+ /** Returns current value for returnResults */
// Used by Increment and Append only.
@InterfaceAudience.Private
protected boolean isReturnResults() {
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/NonceGenerator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/NonceGenerator.java
index 3020be22105..83556e7e91a 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/NonceGenerator.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/NonceGenerator.java
@@ -29,9 +29,9 @@ public interface NonceGenerator {
static final String CLIENT_NONCES_ENABLED_KEY = "hbase.client.nonces.enabled";
- /** @return the nonce group (client ID) of this client manager. */
+ /** Returns the nonce group (client ID) of this client manager. */
long getNonceGroup();
- /** @return New nonce. */
+ /** Returns New nonce. */
long newNonce();
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Query.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Query.java
index 661065d0949..d752693dabc 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Query.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Query.java
@@ -75,18 +75,14 @@ public abstract class Query extends OperationWithAttributes {
return this;
}
- /**
- * @return The authorizations this Query is associated with. n
- */
+ /** Returns The authorizations this Query is associated with. n */
public Authorizations getAuthorizations() throws DeserializationException {
byte[] authorizationsBytes = this.getAttribute(VisibilityConstants.VISIBILITY_LABELS_ATTR_KEY);
if (authorizationsBytes == null) return null;
return ProtobufUtil.toAuthorizations(authorizationsBytes);
}
- /**
- * @return The serialized ACL for this operation, or null if none
- */
+ /** Returns The serialized ACL for this operation, or null if none */
public byte[] getACL() {
return getAttribute(AccessControlConstants.OP_ATTRIBUTE_ACL);
}
@@ -220,9 +216,7 @@ public abstract class Query extends OperationWithAttributes {
return this;
}
- /**
- * @return A map of column families to time ranges
- */
+ /** Returns A map of column families to time ranges */
public Map<byte[], TimeRange> getColumnFamilyTimeRange() {
return this.colFamTimeRangeMap;
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionAdminServiceCallable.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionAdminServiceCallable.java
index 301c4315d25..8b2ed345306 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionAdminServiceCallable.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionAdminServiceCallable.java
@@ -114,9 +114,7 @@ public abstract class RegionAdminServiceCallable<T> implements RetryingCallable<
}
}
- /**
- * @return {@link Connection} instance used by this Callable.
- */
+ /** Returns {@link Connection} instance used by this Callable. */
Connection getConnection() {
return this.connection;
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfo.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfo.java
index b9a5969b4d2..de3f53707b8 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfo.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfo.java
@@ -152,13 +152,11 @@ public interface RegionInfo extends Comparable<RegionInfo> {
};
/**
- * @return Return a short, printable name for this region (usually encoded name) for us logging.
+ * Returns Return a short, printable name for this region (usually encoded name) for us logging.
*/
String getShortNameToLog();
- /**
- * @return the regionId.
- */
+ /** Returns the regionId. */
long getRegionId();
/**
@@ -167,44 +165,28 @@ public interface RegionInfo extends Comparable<RegionInfo> {
*/
byte[] getRegionName();
- /**
- * @return Region name as a String for use in logging, etc.
- */
+ /** Returns Region name as a String for use in logging, etc. */
String getRegionNameAsString();
- /**
- * @return the encoded region name.
- */
+ /** Returns the encoded region name. */
String getEncodedName();
- /**
- * @return the encoded region name as an array of bytes.
- */
+ /** Returns the encoded region name as an array of bytes. */
byte[] getEncodedNameAsBytes();
- /**
- * @return the startKey.
- */
+ /** Returns the startKey. */
byte[] getStartKey();
- /**
- * @return the endKey.
- */
+ /** Returns the endKey. */
byte[] getEndKey();
- /**
- * @return current table name of the region
- */
+ /** Returns current table name of the region */
TableName getTable();
- /**
- * @return returns region replica id
- */
+ /** Returns returns region replica id */
int getReplicaId();
- /**
- * @return True if has been split and has daughters.
- */
+ /** Returns True if has been split and has daughters. */
boolean isSplit();
/**
@@ -223,9 +205,7 @@ public interface RegionInfo extends Comparable<RegionInfo> {
@Deprecated
boolean isSplitParent();
- /**
- * @return true if this region is a meta region.
- */
+ /** Returns true if this region is a meta region. */
boolean isMetaRegion();
/**
@@ -236,9 +216,7 @@ public interface RegionInfo extends Comparable<RegionInfo> {
*/
boolean containsRange(byte[] rangeStartKey, byte[] rangeEndKey);
- /**
- * @return true if the given row falls in this region.
- */
+ /** Returns true if the given row falls in this region. */
boolean containsRow(byte[] row);
/**
@@ -253,9 +231,7 @@ public interface RegionInfo extends Comparable<RegionInfo> {
&& (regionName[regionName.length - 1] == RegionInfo.ENC_SEPARATOR);
}
- /**
- * @return the encodedName
- */
+ /** Returns the encodedName */
@InterfaceAudience.Private
static String encodeRegionName(final byte[] regionName) {
String encodedName;
@@ -371,7 +347,7 @@ public interface RegionInfo extends Comparable<RegionInfo> {
}
/**
- * @return A deserialized {@link RegionInfo} or null if we failed deserialize or passed bytes null
+ * Returns A deserialized {@link RegionInfo} or null if we failed deserialize or passed bytes null
*/
@InterfaceAudience.Private
static RegionInfo parseFromOrNull(final byte[] bytes) {
@@ -380,7 +356,7 @@ public interface RegionInfo extends Comparable<RegionInfo> {
}
/**
- * @return A deserialized {@link RegionInfo} or null if we failed deserialize or passed bytes null
+ * Returns A deserialized {@link RegionInfo} or null if we failed deserialize or passed bytes null
*/
@InterfaceAudience.Private
static RegionInfo parseFromOrNull(final byte[] bytes, int offset, int len) {
@@ -778,16 +754,12 @@ public interface RegionInfo extends Comparable<RegionInfo> {
return ris;
}
- /**
- * @return True if this is first Region in Table
- */
+ /** Returns True if this is first Region in Table */
default boolean isFirst() {
return Bytes.equals(getStartKey(), HConstants.EMPTY_START_ROW);
}
- /**
- * @return True if this is last Region in Table
- */
+ /** Returns True if this is last Region in Table */
default boolean isLast() {
return Bytes.equals(getEndKey(), HConstants.EMPTY_END_ROW);
}
@@ -809,9 +781,7 @@ public interface RegionInfo extends Comparable<RegionInfo> {
return getTable().equals(other.getTable()) && areAdjacent(this, other);
}
- /**
- * @return True if RegionInfo is degenerate... if startKey > endKey.
- */
+ /** Returns True if RegionInfo is degenerate... if startKey > endKey. */
default boolean isDegenerate() {
return !isLast() && Bytes.compareTo(getStartKey(), getEndKey()) > 0;
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionReplicaUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionReplicaUtil.java
index ea022820950..4475a01c0f9 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionReplicaUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionReplicaUtil.java
@@ -73,12 +73,12 @@ public class RegionReplicaUtil {
return getRegionInfoForReplica(regionInfo, DEFAULT_REPLICA_ID);
}
- /** @return true if this replicaId corresponds to default replica for the region */
+ /** Returns true if this replicaId corresponds to default replica for the region */
public static boolean isDefaultReplica(int replicaId) {
return DEFAULT_REPLICA_ID == replicaId;
}
- /** @return true if this region is a default replica for the region */
+ /** Returns true if this region is a default replica for the region */
public static boolean isDefaultReplica(RegionInfo hri) {
return hri.getReplicaId() == DEFAULT_REPLICA_ID;
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionServerCallable.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionServerCallable.java
index 8bdc1543d74..abe4058b99d 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionServerCallable.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionServerCallable.java
@@ -155,9 +155,7 @@ public abstract class RegionServerCallable<T, S> implements RetryingCallable<T>
}
}
- /**
- * @return {@link ClusterConnection} instance used by this Callable.
- */
+ /** Returns {@link ClusterConnection} instance used by this Callable. */
protected ClusterConnection getConnection() {
return (ClusterConnection) this.connection;
}
@@ -200,9 +198,7 @@ public abstract class RegionServerCallable<T, S> implements RetryingCallable<T>
return ConnectionUtils.getPauseTime(pause, tries);
}
- /**
- * @return the HRegionInfo for the current region
- */
+ /** Returns the HRegionInfo for the current region */
public HRegionInfo getHRegionInfo() {
if (this.location == null) {
return null;
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RequestController.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RequestController.java
index 66d864be7d4..325286fd94c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RequestController.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RequestController.java
@@ -67,9 +67,7 @@ public interface RequestController {
void reset() throws InterruptedIOException;
}
- /**
- * @return A new checker for evaluating a batch rows.
- */
+ /** Returns A new checker for evaluating a batch rows. */
Checker newChecker();
/**
@@ -86,9 +84,7 @@ public interface RequestController {
*/
void decTaskCounters(Collection<byte[]> regions, ServerName sn);
- /**
- * @return The number of running task.
- */
+ /** Returns The number of running task. */
long getNumberOfTasksInProgress();
/**
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Result.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Result.java
index 75f6cb95868..2ae13c6ce1c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Result.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Result.java
@@ -671,9 +671,7 @@ public class Result implements CellScannable, CellScanner {
return this.cells == null || this.cells.length == 0;
}
- /**
- * @return the size of the underlying Cell []
- */
+ /** Returns the size of the underlying Cell [] */
public int size() {
return this.cells == null ? 0 : this.cells.length;
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ResultScanner.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ResultScanner.java
index 09f4cd92c7a..89b0a6d2efa 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ResultScanner.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ResultScanner.java
@@ -114,8 +114,6 @@ public interface ResultScanner extends Closeable, Iterable<Result> {
*/
boolean renewLease();
- /**
- * @return the scan metrics, or {@code null} if we do not enable metrics.
- */
+ /** Returns the scan metrics, or {@code null} if we do not enable metrics. */
ScanMetrics getScanMetrics();
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Row.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Row.java
index ad1089f2aad..683304ff758 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Row.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Row.java
@@ -28,9 +28,7 @@ import org.apache.yetus.audience.InterfaceAudience;
public interface Row extends Comparable<Row> {
Comparator<Row> COMPARATOR = (v1, v2) -> Bytes.compareTo(v1.getRow(), v2.getRow());
- /**
- * @return The row.
- */
+ /** Returns The row. */
byte[] getRow();
/**
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RowAccess.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RowAccess.java
index ba613bb1773..cd1f4756079 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RowAccess.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RowAccess.java
@@ -26,13 +26,9 @@ import org.apache.yetus.audience.InterfaceAudience;
*/
@InterfaceAudience.Private
public interface RowAccess<T> extends Iterable<T> {
- /**
- * @return true if there are no elements.
- */
+ /** Returns true if there are no elements. */
boolean isEmpty();
- /**
- * @return the number of elements in this list.
- */
+ /** Returns the number of elements in this list. */
int size();
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RowMutations.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RowMutations.java
index 6b2535989bc..1d9b94a54cc 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RowMutations.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RowMutations.java
@@ -155,9 +155,7 @@ public class RowMutations implements Row {
return row;
}
- /**
- * @return An unmodifiable list of the current mutations.
- */
+ /** Returns An unmodifiable list of the current mutations. */
public List<Mutation> getMutations() {
return Collections.unmodifiableList(mutations);
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerImpl.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerImpl.java
index c36a02a885d..2c82cdd54bc 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerImpl.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerImpl.java
@@ -176,9 +176,7 @@ public class RpcRetryingCallerImpl<T> implements RpcRetryingCaller<T> {
}
}
- /**
- * @return Calculate how long a single call took
- */
+ /** Returns Calculate how long a single call took */
private long singleCallDuration(final long expectedSleep) {
return (EnvironmentEdgeManager.currentTime() - tracker.getStartTime()) + expectedSleep;
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Scan.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Scan.java
index 677ddc0bb3e..d417d8e8453 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Scan.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Scan.java
@@ -687,9 +687,7 @@ public class Scan extends Query {
return this;
}
- /**
- * @return the maximum result size in bytes. See {@link #setMaxResultSize(long)}
- */
+ /** Returns the maximum result size in bytes. See {@link #setMaxResultSize(long)} */
public long getMaxResultSize() {
return maxResultSize;
}
@@ -727,9 +725,7 @@ public class Scan extends Query {
return this.familyMap;
}
- /**
- * @return the number of families in familyMap
- */
+ /** Returns the number of families in familyMap */
public int numFamilies() {
if (hasFamilies()) {
return this.familyMap.size();
@@ -737,16 +733,12 @@ public class Scan extends Query {
return 0;
}
- /**
- * @return true if familyMap is non empty, false otherwise
- */
+ /** Returns true if familyMap is non empty, false otherwise */
public boolean hasFamilies() {
return !this.familyMap.isEmpty();
}
- /**
- * @return the keys of the familyMap
- */
+ /** Returns the keys of the familyMap */
public byte[][] getFamilies() {
if (hasFamilies()) {
return this.familyMap.keySet().toArray(new byte[0][0]);
@@ -754,51 +746,37 @@ public class Scan extends Query {
return null;
}
- /**
- * @return the startrow
- */
+ /** Returns the startrow */
public byte[] getStartRow() {
return this.startRow;
}
- /**
- * @return if we should include start row when scan
- */
+ /** Returns if we should include start row when scan */
public boolean includeStartRow() {
return includeStartRow;
}
- /**
- * @return the stoprow
- */
+ /** Returns the stoprow */
public byte[] getStopRow() {
return this.stopRow;
}
- /**
- * @return if we should include stop row when scan
- */
+ /** Returns if we should include stop row when scan */
public boolean includeStopRow() {
return includeStopRow;
}
- /**
- * @return the max number of versions to fetch
- */
+ /** Returns the max number of versions to fetch */
public int getMaxVersions() {
return this.maxVersions;
}
- /**
- * @return maximum number of values to return for a single call to next()
- */
+ /** Returns maximum number of values to return for a single call to next() */
public int getBatch() {
return this.batch;
}
- /**
- * @return maximum number of values to return per row per CF
- */
+ /** Returns maximum number of values to return per row per CF */
public int getMaxResultsPerColumnFamily() {
return this.storeLimit;
}
@@ -811,9 +789,7 @@ public class Scan extends Query {
return this.storeOffset;
}
- /**
- * @return caching the number of rows fetched when calling next on a scanner
- */
+ /** Returns caching the number of rows fetched when calling next on a scanner */
public int getCaching() {
return this.caching;
}
@@ -833,9 +809,7 @@ public class Scan extends Query {
return filter;
}
- /**
- * @return true is a filter has been specified, false if not
- */
+ /** Returns true is a filter has been specified, false if not */
public boolean hasFilter() {
return filter != null;
}
@@ -996,9 +970,7 @@ public class Scan extends Query {
return this;
}
- /**
- * @return True if this Scan is in "raw" mode.
- */
+ /** Returns True if this Scan is in "raw" mode. */
public boolean isRaw() {
byte[] attr = getAttribute(RAW_ATTR);
return attr == null ? false : Bytes.toBoolean(attr);
@@ -1098,9 +1070,7 @@ public class Scan extends Query {
return this;
}
- /**
- * @return True if collection of scan metrics is enabled. For advanced users.
- */
+ /** Returns True if collection of scan metrics is enabled. For advanced users. */
public boolean isScanMetricsEnabled() {
byte[] attr = getAttribute(Scan.SCAN_ATTRIBUTES_METRICS_ENABLE);
return attr == null ? false : Bytes.toBoolean(attr);
@@ -1129,9 +1099,7 @@ public class Scan extends Query {
return this;
}
- /**
- * @return the limit of rows for this scan
- */
+ /** Returns the limit of rows for this scan */
public int getLimit() {
return limit;
}
@@ -1163,9 +1131,7 @@ public class Scan extends Query {
PREAD
}
- /**
- * @return the read type for this scan
- */
+ /** Returns the read type for this scan */
public ReadType getReadType() {
return readType;
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ScannerCallable.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ScannerCallable.java
index 06130f5a889..9c4f6649b19 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ScannerCallable.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ScannerCallable.java
@@ -396,9 +396,7 @@ public class ScannerCallable extends ClientServiceCallable<Result[]> {
this.renew = val;
}
- /**
- * @return the HRegionInfo for the current region
- */
+ /** Returns the HRegionInfo for the current region */
@Override
public HRegionInfo getHRegionInfo() {
if (!instantiated) {
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableDescriptor.java
index 85f412e2df2..ff97f419bac 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableDescriptor.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableDescriptor.java
@@ -153,9 +153,7 @@ public interface TableDescriptor {
// those tables with the highest priority (From Yi Liang over on HBASE-18109).
int getPriority();
- /**
- * @return Returns the configured replicas per region
- */
+ /** Returns Returns the configured replicas per region */
int getRegionReplication();
/**
@@ -200,9 +198,7 @@ public interface TableDescriptor {
*/
String getValue(String key);
- /**
- * @return Getter for fetching an unmodifiable map.
- */
+ /** Returns Getter for fetching an unmodifiable map. */
Map<Bytes, Bytes> getValues();
/**
@@ -219,9 +215,7 @@ public interface TableDescriptor {
*/
boolean hasColumnFamily(final byte[] name);
- /**
- * @return true if the read-replicas memstore replication is enabled.
- */
+ /** Returns true if the read-replicas memstore replication is enabled. */
boolean hasRegionMemStoreReplication();
/**
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableDescriptorBuilder.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableDescriptorBuilder.java
index a005a752527..1d2b69c7146 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableDescriptorBuilder.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableDescriptorBuilder.java
@@ -1110,9 +1110,7 @@ public class TableDescriptorBuilder {
return families.containsKey(familyName);
}
- /**
- * @return Name of this table and then a map of all of the column family descriptors.
- */
+ /** Returns Name of this table and then a map of all of the column family descriptors. */
@Override
public String toString() {
StringBuilder s = new StringBuilder();
@@ -1135,9 +1133,7 @@ public class TableDescriptorBuilder {
return s.toString();
}
- /**
- * @return map of all table attributes formatted into string.
- */
+ /** Returns map of all table attributes formatted into string. */
public String toStringTableAttributes() {
return getValues(true).toString();
}
@@ -1247,9 +1243,7 @@ public class TableDescriptorBuilder {
return false;
}
- /**
- * @return hash code
- */
+ /** Returns hash code */
@Override
public int hashCode() {
int result = this.name.hashCode();
@@ -1297,9 +1291,7 @@ public class TableDescriptorBuilder {
return setValue(REGION_REPLICATION_KEY, Integer.toString(regionReplication));
}
- /**
- * @return true if the read-replicas memstore replication is enabled.
- */
+ /** Returns true if the read-replicas memstore replication is enabled. */
@Override
public boolean hasRegionMemStoreReplication() {
return getOrDefault(REGION_MEMSTORE_REPLICATION_KEY, Boolean::valueOf,
@@ -1550,9 +1542,7 @@ public class TableDescriptorBuilder {
return getOrDefault(OWNER_KEY, Function.identity(), null);
}
- /**
- * @return the bytes in pb format
- */
+ /** Returns the bytes in pb format */
private byte[] toByteArray() {
return ProtobufUtil.prependPBMagic(ProtobufUtil.toTableSchema(this).toByteArray());
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableState.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableState.java
index ffd8cf8409d..6ecd97a75c9 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableState.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableState.java
@@ -94,44 +94,32 @@ public class TableState {
private final TableName tableName;
private final State state;
- /**
- * @return True if table is {@link State#ENABLED}.
- */
+ /** Returns True if table is {@link State#ENABLED}. */
public boolean isEnabled() {
return isInStates(State.ENABLED);
}
- /**
- * @return True if table is {@link State#ENABLING}.
- */
+ /** Returns True if table is {@link State#ENABLING}. */
public boolean isEnabling() {
return isInStates(State.ENABLING);
}
- /**
- * @return True if {@link State#ENABLED} or {@link State#ENABLING}
- */
+ /** Returns True if {@link State#ENABLED} or {@link State#ENABLING} */
public boolean isEnabledOrEnabling() {
return isInStates(State.ENABLED, State.ENABLING);
}
- /**
- * @return True if table is disabled.
- */
+ /** Returns True if table is disabled. */
public boolean isDisabled() {
return isInStates(State.DISABLED);
}
- /**
- * @return True if table is disabling.
- */
+ /** Returns True if table is disabling. */
public boolean isDisabling() {
return isInStates(State.DISABLING);
}
- /**
- * @return True if {@link State#DISABLED} or {@link State#DISABLED}
- */
+ /** Returns True if {@link State#DISABLED} or {@link State#DISABLED} */
public boolean isDisabledOrDisabling() {
return isInStates(State.DISABLED, State.DISABLING);
}
@@ -146,9 +134,7 @@ public class TableState {
this.state = state;
}
- /**
- * @return table state
- */
+ /** Returns table state */
public State getState() {
return state;
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/backoff/ClientBackoffPolicy.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/backoff/ClientBackoffPolicy.java
index ee66bcada1b..1799c27ce2c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/backoff/ClientBackoffPolicy.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/backoff/ClientBackoffPolicy.java
@@ -32,8 +32,6 @@ public interface ClientBackoffPolicy {
public static final String BACKOFF_POLICY_CLASS = "hbase.client.statistics.backoff-policy";
- /**
- * @return the number of ms to wait on the client based on the
- */
+ /** Returns the number of ms to wait on the client based on the */
public long getBackoffTime(ServerName serverName, byte[] region, ServerStatistics stats);
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/exceptions/PreemptiveFastFailException.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/exceptions/PreemptiveFastFailException.java
index 6d28f3288fd..356e90b2992 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/exceptions/PreemptiveFastFailException.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/exceptions/PreemptiveFastFailException.java
@@ -67,37 +67,27 @@ public class PreemptiveFastFailException extends ConnectException {
this.guaranteedClientSideOnly = guaranteedClientSideOnly;
}
- /**
- * @return time of the fist failure
- */
+ /** Returns time of the fist failure */
public long getFirstFailureAt() {
return timeOfFirstFailureMilliSec;
}
- /**
- * @return time of the latest attempt
- */
+ /** Returns time of the latest attempt */
public long getLastAttemptAt() {
return timeOfLatestAttemptMilliSec;
}
- /**
- * @return failure count
- */
+ /** Returns failure count */
public long getFailureCount() {
return failureCount;
}
- /**
- * @return true if operation was attempted by server, false otherwise.
- */
+ /** Returns true if operation was attempted by server, false otherwise. */
public boolean wasOperationAttemptedByServer() {
return false;
}
- /**
- * @return true if we know no mutation made it to the server, false otherwise.
- */
+ /** Returns true if we know no mutation made it to the server, false otherwise. */
public boolean isGuaranteedClientSideOnly() {
return guaranteedClientSideOnly;
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BigDecimalComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BigDecimalComparator.java
index bfd285975ff..669a591a961 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BigDecimalComparator.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BigDecimalComparator.java
@@ -72,9 +72,7 @@ public class BigDecimalComparator extends ByteArrayComparable {
return this.bigDecimal.compareTo(that);
}
- /**
- * @return The comparator serialized using pb
- */
+ /** Returns The comparator serialized using pb */
@Override
public byte[] toByteArray() {
ComparatorProtos.BigDecimalComparator.Builder builder =
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryComparator.java
index 0c8274a8611..7aff979f4c2 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryComparator.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryComparator.java
@@ -54,9 +54,7 @@ public class BinaryComparator extends org.apache.hadoop.hbase.filter.ByteArrayCo
return ByteBufferUtils.compareTo(this.value, 0, this.value.length, value, offset, length);
}
- /**
- * @return The comparator serialized using pb
- */
+ /** Returns The comparator serialized using pb */
@Override
public byte[] toByteArray() {
ComparatorProtos.BinaryComparator.Builder builder =
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryComponentComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryComponentComparator.java
index c479d757d1c..88622ed2eba 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryComponentComparator.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryComponentComparator.java
@@ -73,9 +73,7 @@ public class BinaryComponentComparator extends ByteArrayComparable {
return result;
}
- /**
- * @return The comparator serialized using pb
- */
+ /** Returns The comparator serialized using pb */
@Override
public byte[] toByteArray() {
ComparatorProtos.BinaryComponentComparator.Builder builder =
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryPrefixComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryPrefixComparator.java
index f97fd070be6..fcf447dcb0c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryPrefixComparator.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryPrefixComparator.java
@@ -58,9 +58,7 @@ public class BinaryPrefixComparator extends ByteArrayComparable {
return ByteBufferUtils.compareTo(this.value, 0, this.value.length, value, offset, length);
}
- /**
- * @return The comparator serialized using pb
- */
+ /** Returns The comparator serialized using pb */
@Override
public byte[] toByteArray() {
ComparatorProtos.BinaryPrefixComparator.Builder builder =
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java
index 15ca8890aba..8abdcf4e8a6 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java
@@ -57,16 +57,12 @@ public class BitComparator extends ByteArrayComparable {
this.bitOperator = bitOperator;
}
- /**
- * @return the bitwise operator
- */
+ /** Returns the bitwise operator */
public BitwiseOp getOperator() {
return bitOperator;
}
- /**
- * @return The comparator serialized using pb
- */
+ /** Returns The comparator serialized using pb */
@Override
public byte[] toByteArray() {
ComparatorProtos.BitComparator.Builder builder = ComparatorProtos.BitComparator.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java
index f1e5644fd8a..a7cd0489ca5 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java
@@ -84,9 +84,7 @@ public class ColumnCountGetFilter extends FilterBase {
return new ColumnCountGetFilter(limit);
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() {
FilterProtos.ColumnCountGetFilter.Builder builder =
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java
index f11005e43d2..e9cf7dce76d 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java
@@ -157,9 +157,7 @@ public class ColumnPaginationFilter extends FilterBase {
return new ColumnPaginationFilter(limit, offset);
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() {
FilterProtos.ColumnPaginationFilter.Builder builder =
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java
index 86401564ed5..de76b8c5059 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java
@@ -108,9 +108,7 @@ public class ColumnPrefixFilter extends FilterBase {
return new ColumnPrefixFilter(columnPrefix);
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() {
FilterProtos.ColumnPrefixFilter.Builder builder = FilterProtos.ColumnPrefixFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java
index dca4609763c..1abd49fff3c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java
@@ -65,44 +65,32 @@ public class ColumnRangeFilter extends FilterBase {
this.maxColumnInclusive = maxColumnInclusive;
}
- /**
- * @return if min column range is inclusive.
- */
+ /** Returns if min column range is inclusive. */
public boolean isMinColumnInclusive() {
return minColumnInclusive;
}
- /**
- * @return if max column range is inclusive.
- */
+ /** Returns if max column range is inclusive. */
public boolean isMaxColumnInclusive() {
return maxColumnInclusive;
}
- /**
- * @return the min column range for the filter
- */
+ /** Returns the min column range for the filter */
public byte[] getMinColumn() {
return this.minColumn;
}
- /**
- * @return true if min column is inclusive, false otherwise
- */
+ /** Returns true if min column is inclusive, false otherwise */
public boolean getMinColumnInclusive() {
return this.minColumnInclusive;
}
- /**
- * @return the max column range for the filter
- */
+ /** Returns the max column range for the filter */
public byte[] getMaxColumn() {
return this.maxColumn;
}
- /**
- * @return true if max column is inclusive, false otherwise
- */
+ /** Returns true if max column is inclusive, false otherwise */
public boolean getMaxColumnInclusive() {
return this.maxColumnInclusive;
}
@@ -161,9 +149,7 @@ public class ColumnRangeFilter extends FilterBase {
return new ColumnRangeFilter(minColumn, minColumnInclusive, maxColumn, maxColumnInclusive);
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() {
FilterProtos.ColumnRangeFilter.Builder builder = FilterProtos.ColumnRangeFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnValueFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnValueFilter.java
index 0074fe40a3a..0b3b85dbb47 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnValueFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnValueFilter.java
@@ -75,23 +75,17 @@ public class ColumnValueFilter extends FilterBase {
return op;
}
- /**
- * @return the comparator
- */
+ /** Returns the comparator */
public ByteArrayComparable getComparator() {
return comparator;
}
- /**
- * @return the column family
- */
+ /** Returns the column family */
public byte[] getFamily() {
return family;
}
- /**
- * @return the qualifier
- */
+ /** Returns the qualifier */
public byte[] getQualifier() {
return qualifier;
}
@@ -161,9 +155,7 @@ public class ColumnValueFilter extends FilterBase {
return new ColumnValueFilter(family, qualifier, operator, comparator);
}
- /**
- * @return A pb instance to represent this instance.
- */
+ /** Returns A pb instance to represent this instance. */
FilterProtos.ColumnValueFilter convert() {
FilterProtos.ColumnValueFilter.Builder builder = FilterProtos.ColumnValueFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/CompareFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/CompareFilter.java
index 7fc8be99396..5b25e3b2ae1 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/CompareFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/CompareFilter.java
@@ -113,9 +113,7 @@ public abstract class CompareFilter extends FilterBase {
return op;
}
- /**
- * @return the comparator
- */
+ /** Returns the comparator */
public ByteArrayComparable getComparator() {
return comparator;
}
@@ -279,9 +277,7 @@ public abstract class CompareFilter extends FilterBase {
return arguments;
}
- /**
- * @return A pb instance to represent this instance.
- */
+ /** Returns A pb instance to represent this instance. */
FilterProtos.CompareFilter convert() {
FilterProtos.CompareFilter.Builder builder = FilterProtos.CompareFilter.newBuilder();
HBaseProtos.CompareType compareOp = CompareType.valueOf(this.op.name());
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java
index c9a688028ca..0a39f4c94f2 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java
@@ -112,23 +112,17 @@ public class DependentColumnFilter extends CompareFilter {
this(family, qualifier, dropDependentColumn, CompareOp.NO_OP, null);
}
- /**
- * @return the column family
- */
+ /** Returns the column family */
public byte[] getFamily() {
return this.columnFamily;
}
- /**
- * @return the column qualifier
- */
+ /** Returns the column qualifier */
public byte[] getQualifier() {
return this.columnQualifier;
}
- /**
- * @return true if we should drop the dependent column, false otherwise
- */
+ /** Returns true if we should drop the dependent column, false otherwise */
public boolean dropDependentColumn() {
return this.dropDependentColumn;
}
@@ -219,9 +213,7 @@ public class DependentColumnFilter extends CompareFilter {
}
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() {
FilterProtos.DependentColumnFilter.Builder builder =
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java
index 489b1363fd7..917ac522de1 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java
@@ -92,9 +92,7 @@ public class FamilyFilter extends CompareFilter {
return new FamilyFilter(compareOp, comparator);
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() {
FilterProtos.FamilyFilter.Builder builder = FilterProtos.FamilyFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java
index 6daf97e931e..35313837843 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java
@@ -192,9 +192,7 @@ final public class FilterList extends FilterBase {
return filterListBase.filterRow();
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() throws IOException {
FilterProtos.FilterList.Builder builder = FilterProtos.FilterList.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.java
index fd6eddf005e..d779d36eef7 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.java
@@ -71,9 +71,7 @@ public class FirstKeyOnlyFilter extends FilterBase {
return new FirstKeyOnlyFilter();
}
- /**
- * @return true if first KV has been found.
- */
+ /** Returns true if first KV has been found. */
protected boolean hasFoundKV() {
return this.foundKV;
}
@@ -85,9 +83,7 @@ public class FirstKeyOnlyFilter extends FilterBase {
this.foundKV = value;
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() {
FilterProtos.FirstKeyOnlyFilter.Builder builder = FilterProtos.FirstKeyOnlyFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyValueMatchingQualifiersFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyValueMatchingQualifiersFilter.java
index d10a6ea8e90..b49e51f348c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyValueMatchingQualifiersFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyValueMatchingQualifiersFilter.java
@@ -83,9 +83,7 @@ public class FirstKeyValueMatchingQualifiersFilter extends FirstKeyOnlyFilter {
return false;
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() {
FilterProtos.FirstKeyValueMatchingQualifiersFilter.Builder builder =
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java
index 5f3e2b3da73..6036aebdd57 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java
@@ -284,9 +284,7 @@ public class FuzzyRowFilter extends FilterBase {
return done;
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() {
FilterProtos.FuzzyRowFilter.Builder builder = FilterProtos.FuzzyRowFilter.newBuilder()
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java
index b8d43cc5c39..67a214e5650 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java
@@ -81,9 +81,7 @@ public class InclusiveStopFilter extends FilterBase {
return new InclusiveStopFilter(stopRowKey);
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() {
FilterProtos.InclusiveStopFilter.Builder builder =
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/KeyOnlyFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/KeyOnlyFilter.java
index 6332d404ad1..60b3a5c0480 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/KeyOnlyFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/KeyOnlyFilter.java
@@ -99,9 +99,7 @@ public class KeyOnlyFilter extends FilterBase {
return filter;
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() {
FilterProtos.KeyOnlyFilter.Builder builder = FilterProtos.KeyOnlyFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/LongComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/LongComparator.java
index ead0ee10447..8f9a7f1dfc3 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/LongComparator.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/LongComparator.java
@@ -53,9 +53,7 @@ public class LongComparator extends ByteArrayComparable {
return Long.compare(longValue, that);
}
- /**
- * @return The comparator serialized using pb
- */
+ /** Returns The comparator serialized using pb */
@Override
public byte[] toByteArray() {
ComparatorProtos.LongComparator.Builder builder = ComparatorProtos.LongComparator.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java
index 55026e8f1bf..af07cc42b11 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java
@@ -182,9 +182,7 @@ public class MultiRowRangeFilter extends FilterBase {
return PrivateCellUtil.createFirstOnRow(comparisonData, 0, (short) comparisonData.length);
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() {
FilterProtos.MultiRowRangeFilter.Builder builder =
@@ -475,16 +473,12 @@ public class MultiRowRangeFilter extends FilterBase {
return stopRow;
}
- /**
- * @return if start row is inclusive.
- */
+ /** Returns if start row is inclusive. */
public boolean isStartRowInclusive() {
return startRowInclusive;
}
- /**
- * @return if stop row is inclusive.
- */
+ /** Returns if stop row is inclusive. */
public boolean isStopRowInclusive() {
return stopRowInclusive;
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java
index 1744c20968e..e7c868e3bbb 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java
@@ -122,9 +122,7 @@ public class MultipleColumnPrefixFilter extends FilterBase {
return new MultipleColumnPrefixFilter(prefixes);
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() {
FilterProtos.MultipleColumnPrefixFilter.Builder builder =
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/NullComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/NullComparator.java
index fc0562ecb3e..6153e1e8080 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/NullComparator.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/NullComparator.java
@@ -63,9 +63,7 @@ public class NullComparator extends ByteArrayComparable {
return value != null ? 1 : 0;
}
- /**
- * @return The comparator serialized using pb
- */
+ /** Returns The comparator serialized using pb */
@Override
public byte[] toByteArray() {
ComparatorProtos.NullComparator.Builder builder = ComparatorProtos.NullComparator.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java
index 603778cac57..45c8a1f1a3f 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java
@@ -97,9 +97,7 @@ public class PageFilter extends FilterBase {
return new PageFilter(pageSize);
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() {
FilterProtos.PageFilter.Builder builder = FilterProtos.PageFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java
index c6fdcc88875..51752d2b72f 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java
@@ -107,9 +107,7 @@ public class PrefixFilter extends FilterBase {
return new PrefixFilter(prefix);
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() {
FilterProtos.PrefixFilter.Builder builder = FilterProtos.PrefixFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/QualifierFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/QualifierFilter.java
index 6ee70a79e9f..7f0b8642003 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/QualifierFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/QualifierFilter.java
@@ -86,9 +86,7 @@ public class QualifierFilter extends CompareFilter {
return new QualifierFilter(compareOp, comparator);
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() {
FilterProtos.QualifierFilter.Builder builder = FilterProtos.QualifierFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RandomRowFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RandomRowFilter.java
index ff73dadb9d8..656ceaf3e1a 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RandomRowFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RandomRowFilter.java
@@ -44,9 +44,7 @@ public class RandomRowFilter extends FilterBase {
this.chance = chance;
}
- /**
- * @return The chance that a row gets included.
- */
+ /** Returns The chance that a row gets included. */
public float getChance() {
return chance;
}
@@ -107,9 +105,7 @@ public class RandomRowFilter extends FilterBase {
filterOutRow = false;
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() {
FilterProtos.RandomRowFilter.Builder builder = FilterProtos.RandomRowFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RegexStringComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RegexStringComparator.java
index 75272c5f241..6841c80194a 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RegexStringComparator.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RegexStringComparator.java
@@ -145,9 +145,7 @@ public class RegexStringComparator extends ByteArrayComparable {
return engine.compareTo(value, offset, length);
}
- /**
- * @return The comparator serialized using pb
- */
+ /** Returns The comparator serialized using pb */
@Override
public byte[] toByteArray() {
return engine.toByteArray();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java
index 6c51e788091..fdefe382439 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java
@@ -106,9 +106,7 @@ public class RowFilter extends CompareFilter {
return new RowFilter(compareOp, comparator);
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() {
FilterProtos.RowFilter.Builder builder = FilterProtos.RowFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueExcludeFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueExcludeFilter.java
index a410884a8ab..bbd8ec67994 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueExcludeFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueExcludeFilter.java
@@ -163,9 +163,7 @@ public class SingleColumnValueExcludeFilter extends SingleColumnValueFilter {
return filter;
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() {
FilterProtos.SingleColumnValueExcludeFilter.Builder builder =
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java
index b049867d2ef..a68fdcd0577 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java
@@ -192,23 +192,17 @@ public class SingleColumnValueFilter extends FilterBase {
return op;
}
- /**
- * @return the comparator
- */
+ /** Returns the comparator */
public org.apache.hadoop.hbase.filter.ByteArrayComparable getComparator() {
return comparator;
}
- /**
- * @return the family
- */
+ /** Returns the family */
public byte[] getFamily() {
return columnFamily;
}
- /**
- * @return the qualifier
- */
+ /** Returns the qualifier */
public byte[] getQualifier() {
return columnQualifier;
}
@@ -356,9 +350,7 @@ public class SingleColumnValueFilter extends FilterBase {
return builder.build();
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() {
return convert().toByteArray();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java
index e3b334b8820..4b23a87537d 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java
@@ -102,9 +102,7 @@ public class SkipFilter extends FilterBase {
return true;
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() throws IOException {
FilterProtos.SkipFilter.Builder builder = FilterProtos.SkipFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SubstringComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SubstringComparator.java
index b8e33c438fe..b5fe22dca24 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SubstringComparator.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SubstringComparator.java
@@ -66,9 +66,7 @@ public class SubstringComparator extends ByteArrayComparable {
return Bytes.toString(value, offset, length).toLowerCase(Locale.ROOT).contains(substr) ? 0 : 1;
}
- /**
- * @return The comparator serialized using pb
- */
+ /** Returns The comparator serialized using pb */
@Override
public byte[] toByteArray() {
ComparatorProtos.SubstringComparator.Builder builder =
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/TimestampsFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/TimestampsFilter.java
index 2ed44ba77e4..eaeaaac3df1 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/TimestampsFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/TimestampsFilter.java
@@ -77,9 +77,7 @@ public class TimestampsFilter extends FilterBase {
init();
}
- /**
- * @return the list of timestamps
- */
+ /** Returns the list of timestamps */
public List<Long> getTimestamps() {
List<Long> list = new ArrayList<>(timestamps.size());
list.addAll(timestamps);
@@ -165,9 +163,7 @@ public class TimestampsFilter extends FilterBase {
return new TimestampsFilter(timestamps);
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() {
FilterProtos.TimestampsFilter.Builder builder = FilterProtos.TimestampsFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ValueFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ValueFilter.java
index 57a1c5c7c2a..e726d88f720 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ValueFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ValueFilter.java
@@ -87,9 +87,7 @@ public class ValueFilter extends CompareFilter {
return new ValueFilter(compareOp, comparator);
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() {
FilterProtos.ValueFilter.Builder builder = FilterProtos.ValueFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/WhileMatchFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/WhileMatchFilter.java
index c58e080b565..e31cbb97378 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/WhileMatchFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/WhileMatchFilter.java
@@ -107,9 +107,7 @@ public class WhileMatchFilter extends FilterBase {
return true;
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() throws IOException {
FilterProtos.WhileMatchFilter.Builder builder = FilterProtos.WhileMatchFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRpcController.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRpcController.java
index b33771e5b58..01412631ef7 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRpcController.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRpcController.java
@@ -60,9 +60,7 @@ public interface HBaseRpcController extends RpcController, CellScannable {
*/
void setPriority(final TableName tn);
- /**
- * @return The priority of this request
- */
+ /** Returns The priority of this request */
int getPriority();
int getCallTimeout();
@@ -110,16 +108,12 @@ public interface HBaseRpcController extends RpcController, CellScannable {
*/
void notifyOnCancel(RpcCallback<Object> callback, CancellationCallback action) throws IOException;
- /**
- * @return True if this Controller is carrying the RPC target Region's RegionInfo.
- */
+ /** Returns True if this Controller is carrying the RPC target Region's RegionInfo. */
default boolean hasRegionInfo() {
return false;
}
- /**
- * @return Target Region's RegionInfo or null if not available or pertinent.
- */
+ /** Returns Target Region's RegionInfo or null if not available or pertinent. */
default RegionInfo getRegionInfo() {
return null;
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRpcControllerImpl.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRpcControllerImpl.java
index a4c51ae62ff..f3e19aeadae 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRpcControllerImpl.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRpcControllerImpl.java
@@ -101,9 +101,7 @@ public class HBaseRpcControllerImpl implements HBaseRpcController {
return this.regionInfo;
}
- /**
- * @return One-shot cell scanner (you cannot back it up and restart)
- */
+ /** Returns One-shot cell scanner (you cannot back it up and restart) */
@Override
public CellScanner cellScanner() {
return cellScanner;
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/IPCUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/IPCUtil.java
index 57f8da98eff..b509dcbd27b 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/IPCUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/IPCUtil.java
@@ -96,9 +96,7 @@ class IPCUtil {
return totalSize;
}
- /**
- * @return Size on the wire when the two messages are written with writeDelimitedTo
- */
+ /** Returns Size on the wire when the two messages are written with writeDelimitedTo */
public static int getTotalSizeWhenWrittenDelimited(Message... messages) {
int totalSize = 0;
for (Message m : messages) {
@@ -149,9 +147,7 @@ class IPCUtil {
serverOverloaded);
}
- /**
- * @return True if the exception is a fatal connection exception.
- */
+ /** Returns True if the exception is a fatal connection exception. */
static boolean isFatalConnectionException(final ExceptionResponse e) {
return e.getExceptionClassName().equals(FatalConnectionException.class.getName());
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RemoteWithExtrasException.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RemoteWithExtrasException.java
index 62d0bb1d455..a25708fe58e 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RemoteWithExtrasException.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RemoteWithExtrasException.java
@@ -123,30 +123,22 @@ public class RemoteWithExtrasException extends RemoteException {
return ex;
}
- /**
- * @return null if not set
- */
+ /** Returns null if not set */
public String getHostname() {
return this.hostname;
}
- /**
- * @return -1 if not set
- */
+ /** Returns -1 if not set */
public int getPort() {
return this.port;
}
- /**
- * @return True if origin exception was a do not retry type.
- */
+ /** Returns True if origin exception was a do not retry type. */
public boolean isDoNotRetry() {
return this.doNotRetry;
}
- /**
- * @return True if the server was considered overloaded when the exception was thrown.
- */
+ /** Returns True if the server was considered overloaded when the exception was thrown. */
public boolean isServerOverloaded() {
return serverOverloaded;
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufMagic.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufMagic.java
index 7c6f780e069..15a196cbe5b 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufMagic.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufMagic.java
@@ -77,9 +77,7 @@ public class ProtobufMagic {
return compareTo(PB_MAGIC, 0, PB_MAGIC.length, bytes, offset, PB_MAGIC.length) == 0;
}
- /**
- * @return Length of {@link #PB_MAGIC}
- */
+ /** Returns Length of {@link #PB_MAGIC} */
public static int lengthOfPBMagic() {
return PB_MAGIC.length;
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java
index a7bb4d0e01b..76ce56318ed 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java
@@ -223,9 +223,7 @@ public final class ProtobufUtil {
}
}
- /**
- * @return Length of {@link ProtobufMagic#lengthOfPBMagic()}
- */
+ /** Returns Length of {@link ProtobufMagic#lengthOfPBMagic()} */
public static int lengthOfPBMagic() {
return ProtobufMagic.lengthOfPBMagic();
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/QuotaFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/QuotaFilter.java
index 9ddd408f845..9dd44c56d43 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/QuotaFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/QuotaFilter.java
@@ -92,32 +92,32 @@ public class QuotaFilter {
return this;
}
- /** @return true if the filter is empty */
+ /** Returns true if the filter is empty */
public boolean isNull() {
return !hasFilters;
}
- /** @return the QuotaType types that we want to filter one */
+ /** Returns the QuotaType types that we want to filter one */
public Set<QuotaType> getTypeFilters() {
return types;
}
- /** @return the Namespace filter regex */
+ /** Returns the Namespace filter regex */
public String getNamespaceFilter() {
return namespaceRegex;
}
- /** @return the Table filter regex */
+ /** Returns the Table filter regex */
public String getTableFilter() {
return tableRegex;
}
- /** @return the User filter regex */
+ /** Returns the User filter regex */
public String getUserFilter() {
return userRegex;
}
- /** @return the RegionServer filter regex */
+ /** Returns the RegionServer filter regex */
public String getRegionServerFilter() {
return regionServerRegex;
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/SpaceQuotaSnapshot.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/SpaceQuotaSnapshot.java
index 4a64a170152..0dd4fd3b7d2 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/SpaceQuotaSnapshot.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/SpaceQuotaSnapshot.java
@@ -74,9 +74,7 @@ public class SpaceQuotaSnapshot implements SpaceQuotaSnapshotView {
return policy;
}
- /**
- * @return {@code true} if the quota is being violated, {@code false} otherwise.
- */
+ /** Returns {@code true} if the quota is being violated, {@code false} otherwise. */
@Override
public boolean isInViolation() {
return inViolation;
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/SpaceQuotaSnapshotView.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/SpaceQuotaSnapshotView.java
index 4bc47b9cddd..f7acff7a283 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/SpaceQuotaSnapshotView.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/SpaceQuotaSnapshotView.java
@@ -39,9 +39,7 @@ public interface SpaceQuotaSnapshotView {
*/
Optional<SpaceViolationPolicy> getPolicy();
- /**
- * @return {@code true} if the quota is being violated, {@code false} otherwise.
- */
+ /** Returns {@code true} if the quota is being violated, {@code false} otherwise. */
boolean isInViolation();
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/Permission.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/Permission.java
index b3ac386689e..ace6f302a19 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/Permission.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/Permission.java
@@ -203,7 +203,7 @@ public class Permission extends VersionedWritable {
return raw.toString();
}
- /** @return the object version number */
+ /** Returns the object version number */
@Override
public byte getVersion() {
return VERSION;
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/CellVisibility.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/CellVisibility.java
index 8abaee00509..05b1f6c596d 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/CellVisibility.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/CellVisibility.java
@@ -35,9 +35,7 @@ public class CellVisibility {
this.expression = expression;
}
- /**
- * @return The visibility expression
- */
+ /** Returns The visibility expression */
public String getExpression() {
return this.expression;
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java
index cae868a01a9..cb063938848 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java
@@ -334,9 +334,7 @@ public final class ProtobufUtil {
}
}
- /**
- * @return Length of {@link ProtobufMagic#lengthOfPBMagic()}
- */
+ /** Returns Length of {@link ProtobufMagic#lengthOfPBMagic()} */
public static int lengthOfPBMagic() {
return ProtobufMagic.lengthOfPBMagic();
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java
index 09f7895b977..0ca16009d6c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java
@@ -841,9 +841,7 @@ public final class RequestConverter {
ProtobufUtil.toMutationNoData(MutationType.APPEND, append, mutationBuilder, nonce)));
}
- /**
- * @return whether or not the rowMutations has a Increment or Append
- */
+ /** Returns whether or not the rowMutations has a Increment or Append */
private static boolean buildNoDataRegionAction(final RowMutations rowMutations,
final List<CellScannable> cells, long nonce, final RegionAction.Builder regionActionBuilder,
final ClientProtos.Action.Builder actionBuilder, final MutationProto.Builder mutationBuilder)
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/HBaseSnapshotException.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/HBaseSnapshotException.java
index 5fd62a86008..65481f7c9f6 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/HBaseSnapshotException.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/HBaseSnapshotException.java
@@ -69,9 +69,7 @@ public class HBaseSnapshotException extends DoNotRetryIOException {
super(message, cause);
}
- /**
- * @return the description of the snapshot that is being failed
- */
+ /** Returns the description of the snapshot that is being failed */
public SnapshotDescription getSnapshotDescription() {
return this.description;
}
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java
index 003878e7b21..4208ea113e4 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java
@@ -137,9 +137,7 @@ public class ZNodePaths {
.append("]").toString();
}
- /**
- * @return the znode string corresponding to a replicaId
- */
+ /** Returns the znode string corresponding to a replicaId */
public String getZNodeForReplica(int replicaId) {
if (RegionReplicaUtil.isDefaultReplica(replicaId)) {
return joinZNode(baseZNode, metaZNodePrefix);
@@ -169,16 +167,12 @@ public class ZNodePaths {
: Integer.parseInt(znode.substring(metaZNodePrefix.length() + 1));
}
- /**
- * @return True if meta znode.
- */
+ /** Returns True if meta znode. */
public boolean isMetaZNodePrefix(String znode) {
return znode != null && znode.startsWith(this.metaZNodePrefix);
}
- /**
- * @return True is the fully qualified path is for meta location
- */
+ /** Returns True is the fully qualified path is for meta location */
public boolean isMetaZNodePath(String path) {
int prefixLen = baseZNode.length() + 1;
return path.length() > prefixLen && isMetaZNodePrefix(path.substring(prefixLen));
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/ArrayBackedTag.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/ArrayBackedTag.java
index 434f731883a..efa64f09e72 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/ArrayBackedTag.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/ArrayBackedTag.java
@@ -91,33 +91,25 @@ public class ArrayBackedTag implements Tag {
this.type = bytes[offset + TAG_LENGTH_SIZE];
}
- /**
- * @return The byte array backing this Tag.
- */
+ /** Returns The byte array backing this Tag. */
@Override
public byte[] getValueArray() {
return this.bytes;
}
- /**
- * @return the tag type
- */
+ /** Returns the tag type */
@Override
public byte getType() {
return this.type;
}
- /**
- * @return Length of actual tag bytes within the backed buffer
- */
+ /** Returns Length of actual tag bytes within the backed buffer */
@Override
public int getValueLength() {
return this.length - INFRASTRUCTURE_SIZE;
}
- /**
- * @return Offset of actual tag bytes within the backed buffer
- */
+ /** Returns Offset of actual tag bytes within the backed buffer */
@Override
public int getValueOffset() {
return this.offset + INFRASTRUCTURE_SIZE;
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/ByteBufferExtendedCell.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/ByteBufferExtendedCell.java
index 824625074b5..cb3d379849e 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/ByteBufferExtendedCell.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/ByteBufferExtendedCell.java
@@ -60,53 +60,33 @@ import org.apache.yetus.audience.InterfaceAudience;
*/
@InterfaceAudience.Private
public abstract class ByteBufferExtendedCell implements ExtendedCell {
- /**
- * @return The {@link ByteBuffer} containing the row bytes.
- */
+ /** Returns The {@link ByteBuffer} containing the row bytes. */
public abstract ByteBuffer getRowByteBuffer();
- /**
- * @return Position in the {@link ByteBuffer} where row bytes start
- */
+ /** Returns Position in the {@link ByteBuffer} where row bytes start */
public abstract int getRowPosition();
- /**
- * @return The {@link ByteBuffer} containing the column family bytes.
- */
+ /** Returns The {@link ByteBuffer} containing the column family bytes. */
public abstract ByteBuffer getFamilyByteBuffer();
- /**
- * @return Position in the {@link ByteBuffer} where column family bytes start
- */
+ /** Returns Position in the {@link ByteBuffer} where column family bytes start */
public abstract int getFamilyPosition();
- /**
- * @return The {@link ByteBuffer} containing the column qualifier bytes.
- */
+ /** Returns The {@link ByteBuffer} containing the column qualifier bytes. */
public abstract ByteBuffer getQualifierByteBuffer();
- /**
- * @return Position in the {@link ByteBuffer} where column qualifier bytes start
- */
+ /** Returns Position in the {@link ByteBuffer} where column qualifier bytes start */
public abstract int getQualifierPosition();
- /**
- * @return The {@link ByteBuffer} containing the value bytes.
- */
+ /** Returns The {@link ByteBuffer} containing the value bytes. */
public abstract ByteBuffer getValueByteBuffer();
- /**
- * @return Position in the {@link ByteBuffer} where value bytes start
- */
+ /** Returns Position in the {@link ByteBuffer} where value bytes start */
public abstract int getValuePosition();
- /**
- * @return The {@link ByteBuffer} containing the tag bytes.
- */
+ /** Returns The {@link ByteBuffer} containing the tag bytes. */
public abstract ByteBuffer getTagsByteBuffer();
- /**
- * @return Position in the {@link ByteBuffer} where tag bytes start
- */
+ /** Returns Position in the {@link ByteBuffer} where tag bytes start */
public abstract int getTagsPosition();
}
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/Cell.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/Cell.java
index 4b7ab303175..4fb254b35b4 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/Cell.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/Cell.java
@@ -68,14 +68,10 @@ public interface Cell extends HeapSize {
*/
byte[] getRowArray();
- /**
- * @return Array index of first row byte
- */
+ /** Returns Array index of first row byte */
int getRowOffset();
- /**
- * @return Number of row bytes. Must be < rowArray.length - offset.
- */
+ /** Returns Number of row bytes. Must be < rowArray.length - offset. */
short getRowLength();
// 2) Family
@@ -87,14 +83,10 @@ public interface Cell extends HeapSize {
*/
byte[] getFamilyArray();
- /**
- * @return Array index of first family byte
- */
+ /** Returns Array index of first family byte */
int getFamilyOffset();
- /**
- * @return Number of family bytes. Must be < familyArray.length - offset.
- */
+ /** Returns Number of family bytes. Must be < familyArray.length - offset. */
byte getFamilyLength();
// 3) Qualifier
@@ -105,14 +97,10 @@ public interface Cell extends HeapSize {
*/
byte[] getQualifierArray();
- /**
- * @return Array index of first qualifier byte
- */
+ /** Returns Array index of first qualifier byte */
int getQualifierOffset();
- /**
- * @return Number of qualifier bytes. Must be < qualifierArray.length - offset.
- */
+ /** Returns Number of qualifier bytes. Must be < qualifierArray.length - offset. */
int getQualifierLength();
// 4) Timestamp
@@ -154,19 +142,13 @@ public interface Cell extends HeapSize {
*/
byte[] getValueArray();
- /**
- * @return Array index of first value byte
- */
+ /** Returns Array index of first value byte */
int getValueOffset();
- /**
- * @return Number of value bytes. Must be < valueArray.length - offset.
- */
+ /** Returns Number of value bytes. Must be < valueArray.length - offset. */
int getValueLength();
- /**
- * @return Serialized size (defaults to include tag length if has some tags).
- */
+ /** Returns Serialized size (defaults to include tag length if has some tags). */
int getSerializedSize();
/**
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScannable.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScannable.java
index 454f5be3bee..b194f08d4c2 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScannable.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScannable.java
@@ -26,8 +26,6 @@ import org.apache.yetus.audience.InterfaceAudience;
*/
@InterfaceAudience.Public
public interface CellScannable {
- /**
- * @return A CellScanner over the contained {@link Cell}s
- */
+ /** Returns A CellScanner over the contained {@link Cell}s */
CellScanner cellScanner();
}
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScanner.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScanner.java
index df2f24b5b1b..8d2573765be 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScanner.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScanner.java
@@ -47,9 +47,7 @@ import org.apache.yetus.audience.InterfaceAudience;
*/
@InterfaceAudience.Public
public interface CellScanner {
- /**
- * @return the current Cell which may be mutable
- */
+ /** Returns the current Cell which may be mutable */
Cell current();
/**
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellUtil.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellUtil.java
index 687114e0921..604aa1dff0d 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellUtil.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellUtil.java
@@ -856,9 +856,7 @@ public final class CellUtil {
return matchingFamily(left, fam) && matchingQualifier(left, qual);
}
- /**
- * @return True if matching column family and the qualifier starts with <code>qual</code>
- */
+ /** Returns True if matching column family and the qualifier starts with <code>qual</code> */
public static boolean matchingColumnFamilyAndQualifierPrefix(final Cell left, final byte[] fam,
final byte[] qual) {
return matchingFamily(left, fam) && PrivateCellUtil.qualifierStartsWith(left, qual);
@@ -976,9 +974,7 @@ public final class CellUtil {
return t == Type.DeleteColumn.getCode() || t == Type.DeleteFamily.getCode();
}
- /**
- * @return True if this cell is a Put.
- */
+ /** Returns True if this cell is a Put. */
@SuppressWarnings("deprecation")
public static boolean isPut(Cell cell) {
return cell.getTypeByte() == Type.Put.getCode();
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/ChoreService.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/ChoreService.java
index 1bba8d49120..b93610f26d0 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/ChoreService.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/ChoreService.java
@@ -228,9 +228,7 @@ public class ChoreService {
}
}
- /**
- * @return true when the chore is scheduled with the implementer of this interface
- */
+ /** Returns true when the chore is scheduled with the implementer of this interface */
@InterfaceAudience.Private
public synchronized boolean isChoreScheduled(ScheduledChore chore) {
return chore != null && scheduledChores.containsKey(chore)
@@ -248,9 +246,7 @@ public class ChoreService {
rescheduleChore(chore);
}
- /**
- * @return number of chores that this service currently has scheduled
- */
+ /** Returns number of chores that this service currently has scheduled */
int getNumberOfScheduledChores() {
return scheduledChores.size();
}
@@ -263,9 +259,7 @@ public class ChoreService {
return choresMissingStartTime.size();
}
- /**
- * @return number of threads in the core pool of the underlying ScheduledThreadPoolExecutor
- */
+ /** Returns number of threads in the core pool of the underlying ScheduledThreadPoolExecutor */
int getCorePoolSize() {
return scheduler.getCorePoolSize();
}
@@ -375,16 +369,12 @@ public class ChoreService {
choresMissingStartTime.clear();
}
- /**
- * @return true when the service is shutdown and thus cannot be used anymore
- */
+ /** Returns true when the service is shutdown and thus cannot be used anymore */
public boolean isShutdown() {
return scheduler.isShutdown();
}
- /**
- * @return true when the service is shutdown and all threads have terminated
- */
+ /** Returns true when the service is shutdown and all threads have terminated */
public boolean isTerminated() {
return scheduler.isTerminated();
}
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/ExtendedCell.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/ExtendedCell.java
index b36e1c90aa5..b3b7a1c5e57 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/ExtendedCell.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/ExtendedCell.java
@@ -86,9 +86,7 @@ public interface ExtendedCell extends RawCell, HeapSize {
getValueLength(), getTagsLength(), withTags);
}
- /**
- * @return Serialized size (defaults to include tag length).
- */
+ /** Returns Serialized size (defaults to include tag length). */
@Override
default int getSerializedSize() {
return getSerializedSize(true);
@@ -155,9 +153,7 @@ public interface ExtendedCell extends RawCell, HeapSize {
*/
byte[] getTagsArray();
- /**
- * @return the first offset where the tags start in the Cell
- */
+ /** Returns the first offset where the tags start in the Cell */
int getTagsOffset();
/**
@@ -170,8 +166,6 @@ public interface ExtendedCell extends RawCell, HeapSize {
*/
int getTagsLength();
- /**
- * @return The byte representation of the KeyValue.TYPE of this cell: one of Put, Delete, etc
- */
+ /** Returns The byte representation of the KeyValue.TYPE of this cell: one of Put, Delete, etc */
byte getTypeByte();
}
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/HBaseConfiguration.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/HBaseConfiguration.java
index 91f232f2dc8..0573fe5e45c 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/HBaseConfiguration.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/HBaseConfiguration.java
@@ -185,9 +185,7 @@ public class HBaseConfiguration extends Configuration {
}
}
- /**
- * @return whether to show HBase Configuration in servlet
- */
+ /** Returns whether to show HBase Configuration in servlet */
public static boolean isShowConfInServlet() {
boolean isShowConf = false;
try {
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
index c8956bb02f4..85b8d9d6d31 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
@@ -1164,16 +1164,12 @@ public class KeyValue implements ExtendedCell, Cloneable {
return this.bytes;
}
- /**
- * @return Offset into {@link #getBuffer()} at which this KeyValue starts.
- */
+ /** Returns Offset into {@link #getBuffer()} at which this KeyValue starts. */
public int getOffset() {
return this.offset;
}
- /**
- * @return Length of bytes this KeyValue occupies in {@link #getBuffer()}.
- */
+ /** Returns Length of bytes this KeyValue occupies in {@link #getBuffer()}. */
public int getLength() {
return length;
}
@@ -1197,9 +1193,7 @@ public class KeyValue implements ExtendedCell, Cloneable {
return klength + vlength;
}
- /**
- * @return Key offset in backing buffer..
- */
+ /** Returns Key offset in backing buffer.. */
public int getKeyOffset() {
return this.offset + ROW_OFFSET;
}
@@ -1208,33 +1202,27 @@ public class KeyValue implements ExtendedCell, Cloneable {
return Bytes.toStringBinary(getBuffer(), getKeyOffset(), getKeyLength());
}
- /**
- * @return Length of key portion.
- */
+ /** Returns Length of key portion. */
public int getKeyLength() {
return Bytes.toInt(this.bytes, this.offset);
}
/**
- * @return the backing array of the entire KeyValue (all KeyValue fields are in a single array)
+ * Returns the backing array of the entire KeyValue (all KeyValue fields are in a single array)
*/
@Override
public byte[] getValueArray() {
return bytes;
}
- /**
- * @return the value offset
- */
+ /** Returns the value offset */
@Override
public int getValueOffset() {
int voffset = getKeyOffset() + getKeyLength();
return voffset;
}
- /**
- * @return Value length
- */
+ /** Returns Value length */
@Override
public int getValueLength() {
int vlength = Bytes.toInt(this.bytes, this.offset + Bytes.SIZEOF_INT);
@@ -1242,63 +1230,51 @@ public class KeyValue implements ExtendedCell, Cloneable {
}
/**
- * @return the backing array of the entire KeyValue (all KeyValue fields are in a single array)
+ * Returns the backing array of the entire KeyValue (all KeyValue fields are in a single array)
*/
@Override
public byte[] getRowArray() {
return bytes;
}
- /**
- * @return Row offset
- */
+ /** Returns Row offset */
@Override
public int getRowOffset() {
return this.offset + ROW_KEY_OFFSET;
}
- /**
- * @return Row length
- */
+ /** Returns Row length */
@Override
public short getRowLength() {
return Bytes.toShort(this.bytes, getKeyOffset());
}
/**
- * @return the backing array of the entire KeyValue (all KeyValue fields are in a single array)
+ * Returns the backing array of the entire KeyValue (all KeyValue fields are in a single array)
*/
@Override
public byte[] getFamilyArray() {
return bytes;
}
- /**
- * @return Family offset
- */
+ /** Returns Family offset */
@Override
public int getFamilyOffset() {
return getFamilyOffset(getFamilyLengthPosition(getRowLength()));
}
- /**
- * @return Family offset
- */
+ /** Returns Family offset */
int getFamilyOffset(int familyLenPosition) {
return familyLenPosition + Bytes.SIZEOF_BYTE;
}
- /**
- * @return Family length
- */
+ /** Returns Family length */
@Override
public byte getFamilyLength() {
return getFamilyLength(getFamilyLengthPosition(getRowLength()));
}
- /**
- * @return Family length
- */
+ /** Returns Family length */
public byte getFamilyLength(int famLenPos) {
return this.bytes[famLenPos];
}
@@ -1308,60 +1284,46 @@ public class KeyValue implements ExtendedCell, Cloneable {
}
/**
- * @return the backing array of the entire KeyValue (all KeyValue fields are in a single array)
+ * Returns the backing array of the entire KeyValue (all KeyValue fields are in a single array)
*/
@Override
public byte[] getQualifierArray() {
return bytes;
}
- /**
- * @return Qualifier offset
- */
+ /** Returns Qualifier offset */
@Override
public int getQualifierOffset() {
return getQualifierOffset(getFamilyOffset());
}
- /**
- * @return Qualifier offset
- */
+ /** Returns Qualifier offset */
private int getQualifierOffset(int foffset) {
return getQualifierOffset(foffset, getFamilyLength());
}
- /**
- * @return Qualifier offset
- */
+ /** Returns Qualifier offset */
int getQualifierOffset(int foffset, int flength) {
return foffset + flength;
}
- /**
- * @return Qualifier length
- */
+ /** Returns Qualifier length */
@Override
public int getQualifierLength() {
return getQualifierLength(getRowLength(), getFamilyLength());
}
- /**
- * @return Qualifier length
- */
+ /** Returns Qualifier length */
private int getQualifierLength(int rlength, int flength) {
return getQualifierLength(getKeyLength(), rlength, flength);
}
- /**
- * @return Qualifier length
- */
+ /** Returns Qualifier length */
int getQualifierLength(int keyLength, int rlength, int flength) {
return keyLength - (int) getKeyDataStructureSize(rlength, flength, 0);
}
- /**
- * @return Timestamp offset
- */
+ /** Returns Timestamp offset */
public int getTimestampOffset() {
return getTimestampOffset(getKeyLength());
}
@@ -1374,9 +1336,7 @@ public class KeyValue implements ExtendedCell, Cloneable {
return getKeyOffset() + keylength - TIMESTAMP_TYPE_SIZE;
}
- /**
- * @return True if this KeyValue has a LATEST_TIMESTAMP timestamp.
- */
+ /** Returns True if this KeyValue has a LATEST_TIMESTAMP timestamp. */
public boolean isLatestTimestamp() {
return Bytes.equals(getBuffer(), getTimestampOffset(), Bytes.SIZEOF_LONG,
HConstants.LATEST_TIMESTAMP_BYTES, 0, Bytes.SIZEOF_LONG);
@@ -1442,9 +1402,7 @@ public class KeyValue implements ExtendedCell, Cloneable {
return Bytes.toLong(this.bytes, tsOffset);
}
- /**
- * @return KeyValue.TYPE byte representation
- */
+ /** Returns KeyValue.TYPE byte representation */
@Override
public byte getTypeByte() {
return getTypeByte(getKeyLength());
@@ -1481,7 +1439,7 @@ public class KeyValue implements ExtendedCell, Cloneable {
}
/**
- * @return the backing array of the entire KeyValue (all KeyValue fields are in a single array)
+ * Returns the backing array of the entire KeyValue (all KeyValue fields are in a single array)
*/
@Override
public byte[] getTagsArray() {
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/PrivateCellUtil.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/PrivateCellUtil.java
index 525433209a3..b3e70132dfc 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/PrivateCellUtil.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/PrivateCellUtil.java
@@ -100,16 +100,12 @@ public final class PrivateCellUtil {
return buffer;
}
- /**
- * @return A new cell which is having the extra tags also added to it.
- */
+ /** Returns A new cell which is having the extra tags also added to it. */
public static Cell createCell(Cell cell, List<Tag> tags) {
return createCell(cell, TagUtil.fromList(tags));
}
- /**
- * @return A new cell which is having the extra tags also added to it.
- */
+ /** Returns A new cell which is having the extra tags also added to it. */
public static Cell createCell(Cell cell, byte[] tags) {
if (cell instanceof ByteBufferExtendedCell) {
return new TagRewriteByteBufferExtendedCell((ByteBufferExtendedCell) cell, tags);
@@ -839,9 +835,7 @@ public final class PrivateCellUtil {
return KeyValue.Type.Delete.getCode() <= type && type <= KeyValue.Type.DeleteFamily.getCode();
}
- /**
- * @return True if this cell is a {@link KeyValue.Type#Delete} type.
- */
+ /** Returns True if this cell is a {@link KeyValue.Type#Delete} type. */
public static boolean isDeleteType(Cell cell) {
return cell.getTypeByte() == KeyValue.Type.Delete.getCode();
}
@@ -862,9 +856,7 @@ public final class PrivateCellUtil {
return cell.getTypeByte() == KeyValue.Type.Delete.getCode();
}
- /**
- * @return True if this cell is a delete family or column type.
- */
+ /** Returns True if this cell is a delete family or column type. */
public static boolean isDeleteColumnOrFamily(Cell cell) {
int t = cell.getTypeByte();
return t == KeyValue.Type.DeleteColumn.getCode() || t == KeyValue.Type.DeleteFamily.getCode();
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/RawCell.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/RawCell.java
index 3961e565721..b4d26dbfee9 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/RawCell.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/RawCell.java
@@ -69,9 +69,7 @@ public interface RawCell extends Cell {
}
}
- /**
- * @return A new cell which is having the extra tags also added to it.
- */
+ /** Returns A new cell which is having the extra tags also added to it. */
public static Cell createCell(Cell cell, List<Tag> tags) {
return PrivateCellUtil.createCell(cell, tags);
}
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/RawCellBuilderFactory.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/RawCellBuilderFactory.java
index 50a320b1033..dec7b083e74 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/RawCellBuilderFactory.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/RawCellBuilderFactory.java
@@ -28,9 +28,7 @@ import org.apache.yetus.audience.InterfaceAudience;
@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.UNITTEST)
public final class RawCellBuilderFactory {
- /**
- * @return the cell that is created
- */
+ /** Returns the cell that is created */
public static RawCellBuilder create() {
return new KeyValueBuilder();
}
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/ScheduledChore.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/ScheduledChore.java
index 68c8d69410a..f6b6537f5d4 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/ScheduledChore.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/ScheduledChore.java
@@ -200,17 +200,13 @@ public abstract class ScheduledChore implements Runnable {
return timeOfThisRun - timeOfLastRun;
}
- /**
- * @return true when the time between runs exceeds the acceptable threshold
- */
+ /** Returns true when the time between runs exceeds the acceptable threshold */
private synchronized boolean missedStartTime() {
return isValidTime(timeOfLastRun) && isValidTime(timeOfThisRun)
&& getTimeBetweenRuns() > getMaximumAllowedTimeBetweenRuns();
}
- /**
- * @return max allowed time in millis between runs.
- */
+ /** Returns max allowed time in millis between runs. */
private double getMaximumAllowedTimeBetweenRuns() {
// Threshold used to determine if the Chore's current run started too late
return 1.5 * timeUnit.toMillis(period);
@@ -224,9 +220,7 @@ public abstract class ScheduledChore implements Runnable {
return time > 0 && time <= EnvironmentEdgeManager.currentTime();
}
- /**
- * @return false when the Chore is not currently scheduled with a ChoreService
- */
+ /** Returns false when the Chore is not currently scheduled with a ChoreService */
public synchronized boolean triggerNow() {
if (choreService == null) {
return false;
@@ -261,16 +255,12 @@ public abstract class ScheduledChore implements Runnable {
return stopper;
}
- /**
- * @return period to execute chore in getTimeUnit() units
- */
+ /** Returns period to execute chore in getTimeUnit() units */
public int getPeriod() {
return period;
}
- /**
- * @return initial delay before executing chore in getTimeUnit() units
- */
+ /** Returns initial delay before executing chore in getTimeUnit() units */
public long getInitialDelay() {
return initialDelay;
}
@@ -295,9 +285,7 @@ public abstract class ScheduledChore implements Runnable {
return timeOfThisRun;
}
- /**
- * @return true when this Chore is scheduled with a ChoreService
- */
+ /** Returns true when this Chore is scheduled with a ChoreService */
public synchronized boolean isScheduled() {
return choreService != null && choreService.isChoreScheduled(this);
}
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/ServerName.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/ServerName.java
index c65a821ad37..c6008561969 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/ServerName.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/ServerName.java
@@ -412,9 +412,7 @@ public class ServerName implements Comparable<ServerName>, Serializable {
return SERVERNAME_PATTERN.matcher(str).matches() ? valueOf(str) : valueOf(str, NON_STARTCODE);
}
- /**
- * @return true if the String follows the pattern of {@link #toString()}, false otherwise.
- */
+ /** Returns true if the String follows the pattern of {@link #toString()}, false otherwise. */
public static boolean isFullServerName(final String str) {
if (str == null || str.isEmpty()) {
return false;
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/Stoppable.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/Stoppable.java
index f2d9b6b4ead..5d6b4957298 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/Stoppable.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/Stoppable.java
@@ -30,8 +30,6 @@ public interface Stoppable {
*/
void stop(String why);
- /**
- * @return True if {@link #stop(String)} has been closed.
- */
+ /** Returns True if {@link #stop(String)} has been closed. */
boolean isStopped();
}
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/TableName.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/TableName.java
index 43732ad6c8a..18eba1ebb53 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/TableName.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/TableName.java
@@ -76,9 +76,7 @@ public final class TableName implements Comparable<TableName> {
/** One globally disallowed name */
public static final String DISALLOWED_TABLE_NAME = "zookeeper";
- /**
- * @return True if <code>tn</code> is the hbase:meta table name.
- */
+ /** Returns True if <code>tn</code> is the hbase:meta table name. */
public static boolean isMetaTableName(final TableName tn) {
return tn.equals(TableName.META_TABLE_NAME);
}
@@ -254,9 +252,7 @@ public final class TableName implements Comparable<TableName> {
return qualifierAsString;
}
- /**
- * @return A pointer to TableName as String bytes.
- */
+ /** Returns A pointer to TableName as String bytes. */
public byte[] toBytes() {
return name;
}
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/Tag.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/Tag.java
index cf8798a0e6b..fd0aa43190b 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/Tag.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/Tag.java
@@ -46,19 +46,13 @@ public interface Tag {
*/
public static final byte CUSTOM_TAG_TYPE_RANGE = (byte) 64;
- /**
- * @return the tag type
- */
+ /** Returns the tag type */
byte getType();
- /**
- * @return Offset of tag value within the backed buffer
- */
+ /** Returns Offset of tag value within the backed buffer */
int getValueOffset();
- /**
- * @return Length of tag value within the backed buffer
- */
+ /** Returns Length of tag value within the backed buffer */
int getValueLength();
/**
@@ -73,9 +67,7 @@ public interface Tag {
*/
byte[] getValueArray();
- /**
- * @return The {@link java.nio.ByteBuffer} containing the value bytes.
- */
+ /** Returns The {@link java.nio.ByteBuffer} containing the value bytes. */
ByteBuffer getValueByteBuffer();
/**
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/TagUtil.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/TagUtil.java
index 53158800cfd..c3c49410210 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/TagUtil.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/TagUtil.java
@@ -65,9 +65,7 @@ public final class TagUtil {
return StreamUtils.readRawVarint32(tag.getValueByteBuffer(), offset);
}
- /**
- * @return A List<Tag> of any Tags found in <code>cell</code> else null.
- */
+ /** Returns A List<Tag> of any Tags found in <code>cell</code> else null. */
public static List<Tag> carryForwardTags(final Cell cell) {
return carryForwardTags(null, cell);
}
@@ -108,9 +106,7 @@ public final class TagUtil {
return b;
}
- /**
- * @return Carry forward the TTL tag.
- */
+ /** Returns Carry forward the TTL tag. */
public static List<Tag> carryForwardTTLTag(final List<Tag> tagsOrNull, final long ttl) {
if (ttl == Long.MAX_VALUE) {
return tagsOrNull;
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/codec/CellCodec.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/codec/CellCodec.java
index 10a363a5f40..e7facdbfbf2 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/codec/CellCodec.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/codec/CellCodec.java
@@ -97,9 +97,7 @@ public class CellCodec implements Codec {
.setTimestamp(timestamp).setType(type).setValue(value).setSequenceId(memstoreTS).build();
}
- /**
- * @return Byte array read from the stream. n
- */
+ /** Returns Byte array read from the stream. n */
private byte[] readByteArray(final InputStream in) throws IOException {
byte[] intArray = new byte[Bytes.SIZEOF_INT];
IOUtils.readFully(in, intArray);
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/codec/CellCodecWithTags.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/codec/CellCodecWithTags.java
index fe9417b23c9..75e3d48d9fa 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/codec/CellCodecWithTags.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/codec/CellCodecWithTags.java
@@ -101,9 +101,7 @@ public class CellCodecWithTags implements Codec {
.setTags(tags).build();
}
- /**
- * @return Byte array read from the stream. n
- */
+ /** Returns Byte array read from the stream. n */
private byte[] readByteArray(final InputStream in) throws IOException {
byte[] intArray = new byte[Bytes.SIZEOF_INT];
IOUtils.readFully(in, intArray);
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/conf/ConfigurationManager.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/conf/ConfigurationManager.java
index 91d2a7be266..0605b3e5190 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/conf/ConfigurationManager.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/conf/ConfigurationManager.java
@@ -117,9 +117,7 @@ public class ConfigurationManager {
}
}
- /**
- * @return the number of observers.
- */
+ /** Returns the number of observers. */
public int getNumObservers() {
synchronized (configurationObservers) {
return configurationObservers.size();
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/filter/ByteArrayComparable.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/filter/ByteArrayComparable.java
index 332d44a84e5..480680b00f3 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/filter/ByteArrayComparable.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/filter/ByteArrayComparable.java
@@ -46,9 +46,7 @@ public abstract class ByteArrayComparable implements Comparable<byte[]> {
return value;
}
- /**
- * @return The comparator serialized using pb
- */
+ /** Returns The comparator serialized using pb */
public abstract byte[] toByteArray();
/**
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ByteArrayOutputStream.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ByteArrayOutputStream.java
index 64bef1b83dc..d6b4c34aebb 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ByteArrayOutputStream.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ByteArrayOutputStream.java
@@ -116,16 +116,12 @@ public class ByteArrayOutputStream extends OutputStream implements ByteBufferWri
buff.put(buf, 0, pos);
}
- /**
- * @return the underlying array where the data gets accumulated
- */
+ /** Returns the underlying array where the data gets accumulated */
public byte[] getBuffer() {
return this.buf;
}
- /**
- * @return The current size of the buffer.
- */
+ /** Returns The current size of the buffer. */
public int size() {
return this.pos;
}
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ImmutableBytesWritable.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ImmutableBytesWritable.java
index 4074eed46e5..da897b04ae1 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ImmutableBytesWritable.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ImmutableBytesWritable.java
@@ -122,9 +122,7 @@ public class ImmutableBytesWritable implements WritableComparable<ImmutableBytes
return this.length;
}
- /**
- * @return the number of valid bytes in the buffer
- */
+ /** Returns the number of valid bytes in the buffer */
public int getLength() {
if (this.bytes == null) {
throw new IllegalStateException(
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/TimeRange.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/TimeRange.java
index b848887bdb6..0c96039c4f8 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/TimeRange.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/TimeRange.java
@@ -162,16 +162,12 @@ public class TimeRange {
}
}
- /**
- * @return the smallest timestamp that should be considered
- */
+ /** Returns the smallest timestamp that should be considered */
public long getMin() {
return minStamp;
}
- /**
- * @return the biggest timestamp that should be considered
- */
+ /** Returns the biggest timestamp that should be considered */
public long getMax() {
return maxStamp;
}
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.java
index b155d9b854e..c2bd9d39da7 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.java
@@ -1001,9 +1001,7 @@ abstract class BufferedDataBlockEncoder extends AbstractDataBlockEncoder {
abstract protected void decodeNext();
}
- /**
- * @return unencoded size added
- */
+ /** Returns unencoded size added */
protected final int afterEncodingKeyValue(Cell cell, DataOutputStream out,
HFileBlockDefaultEncodingContext encodingCtx) throws IOException {
int size = 0;
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoder.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoder.java
index 266dc26d594..7f13b2c6f66 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoder.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoder.java
@@ -129,9 +129,7 @@ public interface DataBlockEncoder {
*/
ByteBuffer getValueShallowCopy();
- /**
- * @return the Cell at the current position. Includes memstore timestamp.
- */
+ /** Returns the Cell at the current position. Includes memstore timestamp. */
Cell getCell();
/** Set position to beginning of given block */
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoding.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoding.java
index 54d045ba95d..71808adf75d 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoding.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoding.java
@@ -77,16 +77,12 @@ public enum DataBlockEncoding {
this.encoderCls = encoderClsName;
}
- /**
- * @return name converted to bytes.
- */
+ /** Returns name converted to bytes. */
public byte[] getNameInBytes() {
return Bytes.toBytes(toString());
}
- /**
- * @return The id of a data block encoder.
- */
+ /** Returns The id of a data block encoder. */
public short getId() {
return id;
}
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/EncodedDataBlock.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/EncodedDataBlock.java
index 4092a1ed2d6..d09ddc9126f 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/EncodedDataBlock.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/EncodedDataBlock.java
@@ -223,7 +223,7 @@ public class EncodedDataBlock {
return getCompressedSize(comprAlgo, compressor, compressedBytes, 0, compressedBytes.length);
}
- /** @return encoded data with header */
+ /** Returns encoded data with header */
private byte[] getEncodedData() {
if (cachedEncodedData != null) {
return cachedEncodedData;
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/HFileBlockDecodingContext.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/HFileBlockDecodingContext.java
index 16de5fb2669..6835a8bac3c 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/HFileBlockDecodingContext.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/HFileBlockDecodingContext.java
@@ -39,8 +39,6 @@ public interface HFileBlockDecodingContext {
void prepareDecoding(int onDiskSizeWithoutHeader, int uncompressedSizeWithoutHeader,
ByteBuff blockBufferWithoutHeader, ByteBuff onDiskBlock) throws IOException;
- /**
- * @return HFile meta information
- */
+ /** Returns HFile meta information */
HFileContext getHFileContext();
}
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/HFileBlockEncodingContext.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/HFileBlockEncodingContext.java
index 15fe5eee5ea..fcac1549b9f 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/HFileBlockEncodingContext.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/HFileBlockEncodingContext.java
@@ -31,14 +31,10 @@ import org.apache.yetus.audience.InterfaceAudience;
@InterfaceAudience.Private
public interface HFileBlockEncodingContext {
- /**
- * @return the block type after encoding
- */
+ /** Returns the block type after encoding */
BlockType getBlockType();
- /**
- * @return the {@link DataBlockEncoding} encoding used
- */
+ /** Returns the {@link DataBlockEncoding} encoding used */
DataBlockEncoding getDataBlockEncoding();
/**
@@ -52,9 +48,7 @@ public interface HFileBlockEncodingContext {
*/
void close();
- /**
- * @return HFile context information
- */
+ /** Returns HFile context information */
HFileContext getHFileContext();
/**
@@ -62,9 +56,7 @@ public interface HFileBlockEncodingContext {
*/
void setEncodingState(EncodingState state);
- /**
- * @return the encoding state
- */
+ /** Returns the encoding state */
EncodingState getEncodingState();
/**
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockType.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockType.java
index 5c8c366a228..3bd98e6388c 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockType.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockType.java
@@ -210,9 +210,7 @@ public enum BlockType {
}
}
- /**
- * @return whether this block type is encoded or unencoded data block
- */
+ /** Returns whether this block type is encoded or unencoded data block */
public final boolean isData() {
return this == DATA || this == ENCODED_DATA;
}
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileContext.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileContext.java
index 843c65a5f41..5064c4a4b8f 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileContext.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileContext.java
@@ -119,9 +119,7 @@ public class HFileContext implements HeapSize, Cloneable {
: CellComparator.getInstance();
}
- /**
- * @return true when on-disk blocks are compressed, and/or encrypted; false otherwise.
- */
+ /** Returns true when on-disk blocks are compressed, and/or encrypted; false otherwise. */
public boolean isCompressedOrEncrypted() {
Compression.Algorithm compressAlgo = getCompression();
boolean compressed = compressAlgo != null && compressAlgo != Compression.Algorithm.NONE;
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/ByteBuff.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/ByteBuff.java
index 2925fab161b..2285917364f 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/ByteBuff.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/ByteBuff.java
@@ -80,9 +80,7 @@ public abstract class ByteBuff implements HBaseReferenceCounted {
/******************************* Methods for ByteBuff **************************************/
- /**
- * @return this ByteBuff's current position
- */
+ /** Returns this ByteBuff's current position */
public abstract int position();
/**
@@ -102,9 +100,7 @@ public abstract class ByteBuff implements HBaseReferenceCounted {
*/
public abstract ByteBuff moveBack(int len);
- /**
- * @return the total capacity of this ByteBuff.
- */
+ /** Returns the total capacity of this ByteBuff. */
public abstract int capacity();
/**
@@ -263,19 +259,13 @@ public abstract class ByteBuff implements HBaseReferenceCounted {
*/
public abstract ByteBuff put(byte[] src);
- /**
- * @return true or false if the underlying BB support hasArray
- */
+ /** Returns true or false if the underlying BB support hasArray */
public abstract boolean hasArray();
- /**
- * @return the byte[] if the underlying BB has single BB and hasArray true
- */
+ /** Returns the byte[] if the underlying BB has single BB and hasArray true */
public abstract byte[] array();
- /**
- * @return the arrayOffset of the byte[] incase of a single BB backed ByteBuff
- */
+ /** Returns the arrayOffset of the byte[] incase of a single BB backed ByteBuff */
public abstract int arrayOffset();
/**
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/MultiByteBuff.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/MultiByteBuff.java
index a3c0b5c528c..042b21f6c00 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/MultiByteBuff.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/MultiByteBuff.java
@@ -131,17 +131,13 @@ public class MultiByteBuff extends ByteBuff {
throw new UnsupportedOperationException();
}
- /**
- * @return false. MBB does not support array based operations
- */
+ /** Returns false. MBB does not support array based operations */
@Override
public boolean hasArray() {
return false;
}
- /**
- * @return the total capacity of this MultiByteBuffer.
- */
+ /** Returns the total capacity of this MultiByteBuffer. */
@Override
public int capacity() {
checkRefCount();
@@ -384,9 +380,7 @@ public class MultiByteBuff extends ByteBuff {
return getLong(index, itemIndex);
}
- /**
- * @return this MBB's current position
- */
+ /** Returns this MBB's current position */
@Override
public int position() {
checkRefCount();
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/security/User.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/security/User.java
index a353f11c488..e7d37069ac5 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/security/User.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/security/User.java
@@ -129,9 +129,7 @@ public abstract class User {
ugi.addToken(token);
}
- /**
- * @return true if user credentials are obtained from keytab.
- */
+ /** Returns true if user credentials are obtained from keytab. */
public boolean isLoginFromKeytab() {
return ugi.isFromKeytab();
}
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/security/UserProvider.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/security/UserProvider.java
index 5f1f65f40ee..f6766bd1f3c 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/security/UserProvider.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/security/UserProvider.java
@@ -149,9 +149,7 @@ public class UserProvider extends BaseConfigurable {
return user == null ? null : user.getName();
}
- /**
- * @return <tt>true</tt> if security is enabled, <tt>false</tt> otherwise
- */
+ /** Returns <tt>true</tt> if security is enabled, <tt>false</tt> otherwise */
public boolean isHBaseSecurityEnabled() {
return User.isHBaseSecurityEnabled(this.getConf());
}
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/AbstractByteRange.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/AbstractByteRange.java
index f2bf5cfa36d..c5fbf52fd46 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/AbstractByteRange.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/AbstractByteRange.java
@@ -152,9 +152,7 @@ public abstract class AbstractByteRange implements ByteRange {
return isEmpty(this);
}
- /**
- * @return true when {@code range} is of zero length, false otherwise.
- */
+ /** Returns true when {@code range} is of zero length, false otherwise. */
public static boolean isEmpty(ByteRange range) {
return range == null || range.getLength() == 0;
}
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteRange.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteRange.java
index 129f17d54df..64bd5cb3b6c 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteRange.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteRange.java
@@ -120,9 +120,7 @@ public interface ByteRange extends Comparable<ByteRange> {
*/
public ByteRange setLength(int length);
- /**
- * @return true when this range is of zero length, false otherwise.
- */
+ /** Returns true when this range is of zero length, false otherwise. */
public boolean isEmpty();
/**
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Bytes.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Bytes.java
index 5505f5f3bdf..1955aecf433 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Bytes.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Bytes.java
@@ -230,9 +230,7 @@ public class Bytes implements Comparable<Bytes> {
return this.length;
}
- /**
- * @return the number of valid bytes in the buffer
- */
+ /** Returns the number of valid bytes in the buffer */
public int getLength() {
if (this.bytes == null) {
throw new IllegalStateException(
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ExceptionUtil.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ExceptionUtil.java
index 26625a2862e..d3b0e82b1dc 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ExceptionUtil.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ExceptionUtil.java
@@ -33,9 +33,7 @@ public final class ExceptionUtil {
private ExceptionUtil() {
}
- /**
- * @return true if the throwable comes an interruption, false otherwise.
- */
+ /** Returns true if the throwable comes an interruption, false otherwise. */
public static boolean isInterrupt(Throwable t) {
if (t instanceof InterruptedException) {
return true;
@@ -59,9 +57,7 @@ public final class ExceptionUtil {
}
}
- /**
- * @return an InterruptedIOException if t was an interruption, null otherwise
- */
+ /** Returns an InterruptedIOException if t was an interruption, null otherwise */
public static InterruptedIOException asInterrupt(Throwable t) {
if (t instanceof SocketTimeoutException) {
return null;
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/HashKey.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/HashKey.java
index cf573aa394e..e4e89f7711f 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/HashKey.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/HashKey.java
@@ -36,8 +36,6 @@ public abstract class HashKey<T> {
*/
public abstract byte get(int pos);
- /**
- * @return The number of bytes in this HashKey
- */
+ /** Returns The number of bytes in this HashKey */
public abstract int length();
}
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Sleeper.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Sleeper.java
index 23ca92251c1..8fdcc189ea8 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Sleeper.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Sleeper.java
@@ -105,9 +105,7 @@ public class Sleeper {
}
}
- /**
- * @return the sleep period in milliseconds
- */
+ /** Returns the sleep period in milliseconds */
public final int getPeriod() {
return period;
}
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/WindowMovingAverage.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/WindowMovingAverage.java
index b79109cc991..b79924a352c 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/WindowMovingAverage.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/WindowMovingAverage.java
@@ -66,9 +66,7 @@ public class WindowMovingAverage extends MovingAverage {
return oneRound;
}
- /**
- * @return number of statistics
- */
+ /** Returns number of statistics */
protected int getNumberOfStatistics() {
return lastN.length;
}
@@ -85,9 +83,7 @@ public class WindowMovingAverage extends MovingAverage {
return lastN[index];
}
- /**
- * @return index of most recent
- */
+ /** Returns index of most recent */
protected int getMostRecentPosistion() {
return mostRecent;
}
diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/HBaseCommonTestingUtility.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/HBaseCommonTestingUtility.java
index cbc7717ca97..76086a3e732 100644
--- a/hbase-common/src/test/java/org/apache/hadoop/hbase/HBaseCommonTestingUtility.java
+++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/HBaseCommonTestingUtility.java
@@ -161,18 +161,14 @@ public class HBaseCommonTestingUtility {
conf.set(propertyName, newDir.getAbsolutePath());
}
- /**
- * @return True if we should delete testing dirs on exit.
- */
+ /** Returns True if we should delete testing dirs on exit. */
boolean deleteOnExit() {
String v = System.getProperty("hbase.testing.preserve.testdir");
// Let default be true, to delete on exit.
return v == null ? true : !Boolean.parseBoolean(v);
}
- /**
- * @return True if we removed the test dirs
- */
+ /** Returns True if we removed the test dirs */
public boolean cleanupTestDir() {
if (deleteDir(this.dataTestDir)) {
this.dataTestDir = null;
diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/net/BoundSocketMaker.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/net/BoundSocketMaker.java
index 69f072a0c6d..c8eb875ec0e 100644
--- a/hbase-common/src/test/java/org/apache/hadoop/hbase/net/BoundSocketMaker.java
+++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/net/BoundSocketMaker.java
@@ -53,9 +53,7 @@ public final class BoundSocketMaker implements Closeable {
return this.socket.getLocalPort();
}
- /**
- * @return Returns a bound socket; be sure to close when done.
- */
+ /** Returns Returns a bound socket; be sure to close when done. */
private ServerSocket get(String hostname, Supplier<Integer> randomPortMaker) {
ServerSocket ss = null;
int port = -1;
diff --git a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/master/MetricsAssignmentManagerSource.java b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/master/MetricsAssignmentManagerSource.java
index 76a70edbddf..6870c60838d 100644
--- a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/master/MetricsAssignmentManagerSource.java
+++ b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/master/MetricsAssignmentManagerSource.java
@@ -166,43 +166,27 @@ public interface MetricsAssignmentManagerSource extends BaseSource {
*/
void incrementOperationCounter();
- /**
- * @return {@link OperationMetrics} containing common metrics for assign region operation
- */
+ /** Returns {@link OperationMetrics} containing common metrics for assign region operation */
OperationMetrics getAssignMetrics();
- /**
- * @return {@link OperationMetrics} containing common metrics for unassign region operation
- */
+ /** Returns {@link OperationMetrics} containing common metrics for unassign region operation */
OperationMetrics getUnassignMetrics();
- /**
- * @return {@link OperationMetrics} containing common metrics for move region operation
- */
+ /** Returns {@link OperationMetrics} containing common metrics for move region operation */
OperationMetrics getMoveMetrics();
- /**
- * @return {@link OperationMetrics} containing common metrics for reopen region operation
- */
+ /** Returns {@link OperationMetrics} containing common metrics for reopen region operation */
OperationMetrics getReopenMetrics();
- /**
- * @return {@link OperationMetrics} containing common metrics for open region request
- */
+ /** Returns {@link OperationMetrics} containing common metrics for open region request */
OperationMetrics getOpenMetrics();
- /**
- * @return {@link OperationMetrics} containing common metrics for close region request
- */
+ /** Returns {@link OperationMetrics} containing common metrics for close region request */
OperationMetrics getCloseMetrics();
- /**
- * @return {@link OperationMetrics} containing common metrics for split operation
- */
+ /** Returns {@link OperationMetrics} containing common metrics for split operation */
OperationMetrics getSplitMetrics();
- /**
- * @return {@link OperationMetrics} containing common metrics for merge operation
- */
+ /** Returns {@link OperationMetrics} containing common metrics for merge operation */
OperationMetrics getMergeMetrics();
}
diff --git a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/master/MetricsMasterSource.java b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/master/MetricsMasterSource.java
index 06b2167a243..5952bdc4d8e 100644
--- a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/master/MetricsMasterSource.java
+++ b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/master/MetricsMasterSource.java
@@ -92,8 +92,6 @@ public interface MetricsMasterSource extends BaseSource {
*/
void incRequests(final long inc);
- /**
- * @return {@link OperationMetrics} containing common metrics for server crash operation
- */
+ /** Returns {@link OperationMetrics} containing common metrics for server crash operation */
OperationMetrics getServerCrashMetrics();
}
diff --git a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionServerWrapper.java b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionServerWrapper.java
index 1d5dcb23d83..46de6742db6 100644
--- a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionServerWrapper.java
+++ b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionServerWrapper.java
@@ -116,24 +116,16 @@ public interface MetricsRegionServerWrapper {
*/
long getStoreFileSize();
- /**
- * @return Max age of store files hosted on this region server
- */
+ /** Returns Max age of store files hosted on this region server */
long getMaxStoreFileAge();
- /**
- * @return Min age of store files hosted on this region server
- */
+ /** Returns Min age of store files hosted on this region server */
long getMinStoreFileAge();
- /**
- * @return Average age of store files hosted on this region server
- */
+ /** Returns Average age of store files hosted on this region server */
long getAvgStoreFileAge();
- /**
- * @return Number of reference files on this region server
- */
+ /** Returns Number of reference files on this region server */
long getNumReferenceFiles();
/**
@@ -507,43 +499,29 @@ public interface MetricsRegionServerWrapper {
*/
double getMobFileCacheHitPercent();
- /**
- * @return Count of hedged read operations
- */
+ /** Returns Count of hedged read operations */
long getHedgedReadOps();
- /**
- * @return Count of times a hedged read beat out the primary read.
- */
+ /** Returns Count of times a hedged read beat out the primary read. */
long getHedgedReadWins();
- /**
- * @return Count of times a hedged read executes in current thread
- */
+ /** Returns Count of times a hedged read executes in current thread */
long getHedgedReadOpsInCurThread();
- /**
- * @return Number of total bytes read from HDFS.
- */
+ /** Returns Number of total bytes read from HDFS. */
long getTotalBytesRead();
- /**
- * @return Number of bytes read from the local HDFS DataNode.
- */
+ /** Returns Number of bytes read from the local HDFS DataNode. */
long getLocalBytesRead();
- /**
- * @return Number of bytes read locally through HDFS short circuit.
- */
+ /** Returns Number of bytes read locally through HDFS short circuit. */
long getShortCircuitBytesRead();
- /**
- * @return Number of bytes read locally through HDFS zero copy.
- */
+ /** Returns Number of bytes read locally through HDFS zero copy. */
long getZeroCopyBytesRead();
/**
- * @return Count of requests blocked because the memstore size is larger than blockingMemStoreSize
+ * Returns Count of requests blocked because the memstore size is larger than blockingMemStoreSize
*/
long getBlockedRequestsCount();
diff --git a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionWrapper.java b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionWrapper.java
index b20d1c1ec80..2782af18b83 100644
--- a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionWrapper.java
+++ b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionWrapper.java
@@ -75,24 +75,16 @@ public interface MetricsRegionWrapper {
*/
long getFilteredReadRequestCount();
- /**
- * @return Max age of store files under this region
- */
+ /** Returns Max age of store files under this region */
long getMaxStoreFileAge();
- /**
- * @return Min age of store files under this region
- */
+ /** Returns Min age of store files under this region */
long getMinStoreFileAge();
- /**
- * @return Average age of store files under this region
- */
+ /** Returns Average age of store files under this region */
long getAvgStoreFileAge();
- /**
- * @return Number of reference files under this region
- */
+ /** Returns Number of reference files under this region */
long getNumReferenceFiles();
/**
@@ -108,9 +100,7 @@ public interface MetricsRegionWrapper {
long getNumCompactionsCompleted();
- /**
- * @return Age of the last major compaction
- */
+ /** Returns Age of the last major compaction */
long getLastMajorCompactionAge();
/**
@@ -151,9 +141,7 @@ public interface MetricsRegionWrapper {
*/
int getReplicaId();
- /**
- * @return the number of references active on the store
- */
+ /** Returns the number of references active on the store */
long getStoreRefCount();
/**
@@ -162,14 +150,10 @@ public interface MetricsRegionWrapper {
*/
long getMaxCompactedStoreFileRefCount();
- /**
- * @return the number of row reads completely on memstore per store
- */
+ /** Returns the number of row reads completely on memstore per store */
Map<String, Long> getMemstoreOnlyRowReadsCount();
- /**
- * @return the number of row reads on memstore and file per store
- */
+ /** Returns the number of row reads on memstore and file per store */
Map<String, Long> getMixedRowReadsCount();
}
diff --git a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsTableWrapperAggregate.java b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsTableWrapperAggregate.java
index 95489cb10b0..e18ba51b3c2 100644
--- a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsTableWrapperAggregate.java
+++ b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsTableWrapperAggregate.java
@@ -88,33 +88,21 @@ public interface MetricsTableWrapperAggregate {
*/
long getMaxStoreFiles(String table);
- /**
- * @return Max age of store files for this table
- */
+ /** Returns Max age of store files for this table */
long getMaxStoreFileAge(String table);
- /**
- * @return Min age of store files for this table
- */
+ /** Returns Min age of store files for this table */
long getMinStoreFileAge(String table);
- /**
- * @return Average age of store files for this table
- */
+ /** Returns Average age of store files for this table */
long getAvgStoreFileAge(String table);
- /**
- * @return Number of reference files for this table
- */
+ /** Returns Number of reference files for this table */
long getNumReferenceFiles(String table);
- /**
- * @return number of row reads completely from memstore per store for this table
- */
+ /** Returns number of row reads completely from memstore per store for this table */
Map<String, Long> getMemstoreOnlyRowReadsCount(String table);
- /**
- * @return number of row reads from file and memstore per store for this table
- */
+ /** Returns number of row reads from file and memstore per store for this table */
Map<String, Long> getMixedRowReadsCount(String table);
}
diff --git a/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/metrics2/lib/DynamicMetricsRegistry.java b/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/metrics2/lib/DynamicMetricsRegistry.java
index 7177f322210..f1b1196cea4 100644
--- a/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/metrics2/lib/DynamicMetricsRegistry.java
+++ b/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/metrics2/lib/DynamicMetricsRegistry.java
@@ -68,9 +68,7 @@ public class DynamicMetricsRegistry {
metricsInfo = info;
}
- /**
- * @return the info object of the metrics registry
- */
+ /** Returns the info object of the metrics registry */
public MetricsInfo info() {
return metricsInfo;
}
diff --git a/hbase-http/src/main/java/org/apache/hadoop/hbase/util/JSONBean.java b/hbase-http/src/main/java/org/apache/hadoop/hbase/util/JSONBean.java
index 2360dcb32a2..687b1e34041 100644
--- a/hbase-http/src/main/java/org/apache/hadoop/hbase/util/JSONBean.java
+++ b/hbase-http/src/main/java/org/apache/hadoop/hbase/util/JSONBean.java
@@ -129,9 +129,7 @@ public class JSONBean {
};
}
- /**
- * @return Return non-zero if failed to find bean. 0
- */
+ /** Returns Return non-zero if failed to find bean. 0 */
private static int write(JsonWriter writer, MBeanServer mBeanServer, ObjectName qry,
String attribute, boolean description, ObjectName excluded) throws IOException {
LOG.debug("Listing beans for {}", qry);
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestMobCompaction.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestMobCompaction.java
index 51f9bca5580..a5172997eba 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestMobCompaction.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestMobCompaction.java
@@ -162,9 +162,7 @@ public class IntegrationTestMobCompaction extends IntegrationTestBase {
LOG.debug("Done initializing/checking cluster");
}
- /**
- * @return status of CLI execution
- */
+ /** Returns status of CLI execution */
@Override
public int runTestFromCommandLine() throws Exception {
testMobCompaction();
diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapred/TableSplit.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapred/TableSplit.java
index 8cc03b22254..2cb63ba7a6a 100644
--- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapred/TableSplit.java
+++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapred/TableSplit.java
@@ -56,27 +56,27 @@ public class TableSplit implements InputSplit, Comparable<TableSplit> {
this(TableName.valueOf(tableName), startRow, endRow, location);
}
- /** @return table name */
+ /** Returns table name */
public TableName getTable() {
return this.m_tableName;
}
- /** @return table name */
+ /** Returns table name */
public byte[] getTableName() {
return this.m_tableName.getName();
}
- /** @return starting row key */
+ /** Returns starting row key */
public byte[] getStartRow() {
return this.m_startRow;
}
- /** @return end row key */
+ /** Returns end row key */
public byte[] getEndRow() {
return this.m_endRow;
}
- /** @return the region's hostname */
+ /** Returns the region's hostname */
public String getRegionLocation() {
return this.m_regionLocation;
}
diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/CellCreator.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/CellCreator.java
index bc2f8040db1..a48ba49058a 100644
--- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/CellCreator.java
+++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/CellCreator.java
@@ -120,9 +120,7 @@ public class CellCreator {
qlength, timestamp, KeyValue.Type.Put, value, voffset, vlength, tags);
}
- /**
- * @return Visibility expression resolver
- */
+ /** Returns Visibility expression resolver */
public VisibilityExpressionResolver getVisibilityExpressionResolver() {
return this.visExpResolver;
}
diff --git a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java
index beef9308ca6..ee36048cb3e 100644
--- a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java
+++ b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java
@@ -1406,16 +1406,12 @@ public class PerformanceEvaluation extends Configured implements Tool {
}
}
- /**
- * @return Subset of the histograms' calculation.
- */
+ /** Returns Subset of the histograms' calculation. */
public String getShortLatencyReport() {
return YammerHistogramUtils.getShortHistogramReport(this.latencyHistogram);
}
- /**
- * @return Subset of the histograms' calculation.
- */
+ /** Returns Subset of the histograms' calculation. */
public String getShortValueSizeReport() {
return YammerHistogramUtils.getShortHistogramReport(this.valueSizeHistogram);
}
diff --git a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestRoundRobinTableInputFormat.java b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestRoundRobinTableInputFormat.java
index df9d1597857..34bd76937d3 100644
--- a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestRoundRobinTableInputFormat.java
+++ b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestRoundRobinTableInputFormat.java
@@ -77,9 +77,7 @@ public class TestRoundRobinTableInputFormat {
}
}
- /**
- * @return Splits made out of {@link #KEYS}. Splits are for five Servers. Length is ZERO!
- */
+ /** Returns Splits made out of {@link #KEYS}. Splits are for five Servers. Length is ZERO! */
private List<InputSplit> createSplits() {
List<InputSplit> splits = new ArrayList<>(KEYS.length - 1);
for (int i = 0; i < KEYS.length - 1; i++) {
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/LockAndQueue.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/LockAndQueue.java
index e6b7d7d0397..fa6c72c490f 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/LockAndQueue.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/LockAndQueue.java
@@ -105,9 +105,7 @@ public class LockAndQueue implements LockStatus {
// try/release Shared/Exclusive lock
// ======================================================================
- /**
- * @return whether we have succesfully acquired the shared lock.
- */
+ /** Returns whether we have succesfully acquired the shared lock. */
public boolean trySharedLock(Procedure<?> proc) {
if (hasExclusiveLock() && !hasLockAccess(proc)) {
return false;
@@ -119,9 +117,7 @@ public class LockAndQueue implements LockStatus {
return true;
}
- /**
- * @return whether we should wake the procedures waiting on the lock here.
- */
+ /** Returns whether we should wake the procedures waiting on the lock here. */
public boolean releaseSharedLock() {
// hasExclusiveLock could be true, it usually means we acquire shared lock while we or our
// parent have held the xlock. And since there is still an exclusive lock, we do not need to
@@ -137,9 +133,7 @@ public class LockAndQueue implements LockStatus {
return true;
}
- /**
- * @return whether we should wake the procedures waiting on the lock here.
- */
+ /** Returns whether we should wake the procedures waiting on the lock here. */
public boolean releaseExclusiveLock(Procedure<?> proc) {
if (
exclusiveLockOwnerProcedure == null
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/Procedure.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/Procedure.java
index ec46e523fc5..a71026e6007 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/Procedure.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/Procedure.java
@@ -628,9 +628,7 @@ public abstract class Procedure<TEnvironment> implements Comparable<Procedure<TE
return timeout != NO_TIMEOUT;
}
- /**
- * @return the timeout in msec
- */
+ /** Returns the timeout in msec */
public int getTimeout() {
return timeout;
}
@@ -665,16 +663,12 @@ public abstract class Procedure<TEnvironment> implements Comparable<Procedure<TE
// ==========================================================================
// runtime state
// ==========================================================================
- /**
- * @return the time elapsed between the last update and the start time of the procedure.
- */
+ /** Returns the time elapsed between the last update and the start time of the procedure. */
public long elapsedTime() {
return getLastUpdate() - getSubmittedTime();
}
- /**
- * @return the serialized result if any, otherwise null
- */
+ /** Returns the serialized result if any, otherwise null */
public byte[] getResult() {
return result;
}
@@ -717,9 +711,7 @@ public abstract class Procedure<TEnvironment> implements Comparable<Procedure<TE
// just because the procedure can get scheduled on different executor threads on each step.
// ==============================================================================================
- /**
- * @return true if the procedure is in a RUNNABLE state.
- */
+ /** Returns true if the procedure is in a RUNNABLE state. */
public synchronized boolean isRunnable() {
return state == ProcedureState.RUNNABLE;
}
@@ -728,16 +720,12 @@ public abstract class Procedure<TEnvironment> implements Comparable<Procedure<TE
return state == ProcedureState.INITIALIZING;
}
- /**
- * @return true if the procedure has failed. It may or may not have rolled back.
- */
+ /** Returns true if the procedure has failed. It may or may not have rolled back. */
public synchronized boolean isFailed() {
return state == ProcedureState.FAILED || state == ProcedureState.ROLLEDBACK;
}
- /**
- * @return true if the procedure is finished successfully.
- */
+ /** Returns true if the procedure is finished successfully. */
public synchronized boolean isSuccess() {
return state == ProcedureState.SUCCESS && !hasException();
}
@@ -750,9 +738,7 @@ public abstract class Procedure<TEnvironment> implements Comparable<Procedure<TE
return isSuccess() || state == ProcedureState.ROLLEDBACK;
}
- /**
- * @return true if the procedure is waiting for a child to finish or for an external event.
- */
+ /** Returns true if the procedure is waiting for a child to finish or for an external event. */
public synchronized boolean isWaiting() {
switch (state) {
case WAITING:
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureEvent.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureEvent.java
index e55f73211d7..ab1929fb8ce 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureEvent.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureEvent.java
@@ -42,7 +42,7 @@ public class ProcedureEvent<T> {
}
/**
- * @return true if event is not ready and adds procedure to suspended queue, else returns false.
+ * Returns true if event is not ready and adds procedure to suspended queue, else returns false.
*/
public synchronized boolean suspendIfNotReady(Procedure proc) {
if (!ready) {
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.java
index 530e5305ba2..4e7c1ca04a7 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.java
@@ -680,16 +680,12 @@ public class ProcedureExecutor<TEnvironment> {
return running.get();
}
- /**
- * @return the current number of worker threads.
- */
+ /** Returns the current number of worker threads. */
public int getWorkerThreadCount() {
return workerThreads.size();
}
- /**
- * @return the core pool size settings.
- */
+ /** Returns the core pool size settings. */
public int getCorePoolSize() {
return corePoolSize;
}
@@ -2004,9 +2000,7 @@ public class ProcedureExecutor<TEnvironment> {
return getName() + "(pid=" + (p == null ? Procedure.NO_PROC_ID : p.getProcId() + ")");
}
- /**
- * @return the time since the current procedure is running
- */
+ /** Returns the time since the current procedure is running */
public long getCurrentRunTime() {
return EnvironmentEdgeManager.currentTime() - executionStartTime.get();
}
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureMetrics.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureMetrics.java
index f86a2b2d00a..bc82145cfe2 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureMetrics.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureMetrics.java
@@ -33,18 +33,14 @@ import org.apache.yetus.audience.InterfaceAudience;
*/
@InterfaceAudience.Private
public interface ProcedureMetrics {
- /**
- * @return Total number of instances submitted for a type of a procedure
- */
+ /** Returns Total number of instances submitted for a type of a procedure */
Counter getSubmittedCounter();
/**
- * @return Histogram of runtimes for all successfully completed instances of a type of a procedure
+ * Returns Histogram of runtimes for all successfully completed instances of a type of a procedure
*/
Histogram getTimeHisto();
- /**
- * @return Total number of instances failed for a type of a procedure
- */
+ /** Returns Total number of instances failed for a type of a procedure */
Counter getFailedCounter();
}
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureScheduler.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureScheduler.java
index f89cac5137c..c0ef8501f4c 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureScheduler.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureScheduler.java
@@ -87,9 +87,7 @@ public interface ProcedureScheduler {
*/
void completionCleanup(Procedure proc);
- /**
- * @return true if there are procedures available to process, otherwise false.
- */
+ /** Returns true if there are procedures available to process, otherwise false. */
boolean hasRunnables();
/**
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/ProcedureStore.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/ProcedureStore.java
index 7e54cfa9a12..386886d0c8c 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/ProcedureStore.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/ProcedureStore.java
@@ -165,14 +165,10 @@ public interface ProcedureStore {
*/
void stop(boolean abort);
- /**
- * @return true if the store is running, otherwise false.
- */
+ /** Returns true if the store is running, otherwise false. */
boolean isRunning();
- /**
- * @return the number of threads/slots passed to start()
- */
+ /** Returns the number of threads/slots passed to start() */
int getNumThreads();
/**
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/BitSetNode.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/BitSetNode.java
index fdbf183b19f..e97fe9bfc84 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/BitSetNode.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/BitSetNode.java
@@ -202,9 +202,7 @@ class BitSetNode {
return (modified[wordIndex] & (1L << bitmapIndex)) != 0;
}
- /**
- * @return true, if all the procedures has been modified.
- */
+ /** Returns true, if all the procedures has been modified. */
public boolean isAllModified() {
// TODO: cache the value
for (int i = 0; i < modified.length; ++i) {
@@ -215,9 +213,7 @@ class BitSetNode {
return true;
}
- /**
- * @return all the active procedure ids in this bit set.
- */
+ /** Returns all the active procedure ids in this bit set. */
public long[] getActiveProcIds() {
List<Long> procIds = new ArrayList<>();
for (int wordIndex = 0; wordIndex < modified.length; wordIndex++) {
@@ -236,9 +232,7 @@ class BitSetNode {
return procIds.stream().mapToLong(Long::longValue).toArray();
}
- /**
- * @return true, if there are no active procedures in this BitSetNode, else false.
- */
+ /** Returns true, if there are no active procedures in this BitSetNode, else false. */
public boolean isEmpty() {
// TODO: cache the value
for (int i = 0; i < deleted.length; ++i) {
@@ -429,16 +423,12 @@ class BitSetNode {
// ========================================================================
// Helpers
// ========================================================================
- /**
- * @return upper boundary (aligned to multiple of BITS_PER_WORD) of bitmap range x belongs to.
- */
+ /** Returns upper boundary (aligned to multiple of BITS_PER_WORD) of bitmap range x belongs to. */
private static long alignUp(final long x) {
return (x + (BITS_PER_WORD - 1)) & -BITS_PER_WORD;
}
- /**
- * @return lower boundary (aligned to multiple of BITS_PER_WORD) of bitmap range x belongs to.
- */
+ /** Returns lower boundary (aligned to multiple of BITS_PER_WORD) of bitmap range x belongs to. */
private static long alignDown(final long x) {
return x & -BITS_PER_WORD;
}
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureStoreTracker.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureStoreTracker.java
index eeb5a3a827b..692333ff733 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureStoreTracker.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureStoreTracker.java
@@ -343,9 +343,7 @@ class ProcedureStoreTracker {
this.partial = isPartial;
}
- /**
- * @return true, if no procedure is active, else false.
- */
+ /** Returns true, if no procedure is active, else false. */
public boolean isEmpty() {
for (Map.Entry<Long, BitSetNode> entry : map.entrySet()) {
if (!entry.getValue().isEmpty()) {
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/util/DelayedUtil.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/util/DelayedUtil.java
index 19c4f8bcb58..32232ac18e6 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/util/DelayedUtil.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/util/DelayedUtil.java
@@ -74,9 +74,7 @@ public final class DelayedUtil {
}
};
- /**
- * @return null (if an interrupt) or an instance of E; resets interrupt on calling thread.
- */
+ /** Returns null (if an interrupt) or an instance of E; resets interrupt on calling thread. */
public static <E extends Delayed> E takeWithoutInterrupt(final DelayQueue<E> queue,
final long timeout, final TimeUnit timeUnit) {
try {
@@ -87,9 +85,7 @@ public final class DelayedUtil {
}
}
- /**
- * @return Time remaining as milliseconds.
- */
+ /** Returns Time remaining as milliseconds. */
public static long getRemainingTime(final TimeUnit resultUnit, final long timeout) {
final long currentTime = EnvironmentEdgeManager.currentTime();
if (currentTime >= timeout) {
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/ProtobufMessageHandler.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/ProtobufMessageHandler.java
index f90354b7ab7..39a7ba71dd6 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/ProtobufMessageHandler.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/ProtobufMessageHandler.java
@@ -26,9 +26,7 @@ import org.apache.yetus.audience.InterfaceAudience;
*/
@InterfaceAudience.Private
public interface ProtobufMessageHandler {
- /**
- * @return the protobuf represention of the model
- */
+ /** Returns the protobuf represention of the model */
byte[] createProtobufOutput();
/**
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/RESTServlet.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/RESTServlet.java
index 10b96ec9284..66834f36176 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/RESTServlet.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/RESTServlet.java
@@ -52,17 +52,13 @@ public class RESTServlet implements Constants {
return realUser;
}
- /**
- * @return the RESTServlet singleton instance
- */
+ /** Returns the RESTServlet singleton instance */
public synchronized static RESTServlet getInstance() {
assert (INSTANCE != null);
return INSTANCE;
}
- /**
- * @return the ConnectionCache instance
- */
+ /** Returns the ConnectionCache instance */
public ConnectionCache getConnectionCache() {
return connectionCache;
}
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java
index c10ad88b4b7..c40165c8254 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java
@@ -53,14 +53,12 @@ public class TableResource extends ResourceBase {
this.table = table;
}
- /** @return the table name */
+ /** Returns the table name */
String getName() {
return table;
}
- /**
- * @return true if the table exists n
- */
+ /** Returns true if the table exists n */
boolean exists() throws IOException {
return servlet.getAdmin().tableExists(TableName.valueOf(table));
}
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/Client.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/Client.java
index ff1fe141bb8..85cb2af86a8 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/Client.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/Client.java
@@ -212,9 +212,7 @@ public class Client {
public void shutdown() {
}
- /**
- * @return the wrapped HttpClient
- */
+ /** Returns the wrapped HttpClient */
public HttpClient getHttpClient() {
return httpClient;
}
@@ -394,9 +392,7 @@ public class Client {
}
}
- /**
- * @return the cluster definition
- */
+ /** Returns the cluster definition */
public Cluster getCluster() {
return cluster;
}
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/Cluster.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/Cluster.java
index dbb30adbc74..9070dfd2800 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/Cluster.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/Cluster.java
@@ -45,9 +45,7 @@ public class Cluster {
this.nodes.addAll(nodes);
}
- /**
- * @return true if no locations have been added, false otherwise
- */
+ /** Returns true if no locations have been added, false otherwise */
public boolean isEmpty() {
return nodes.isEmpty();
}
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/Response.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/Response.java
index cd0ac33dd79..7041d88ce30 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/Response.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/Response.java
@@ -82,9 +82,7 @@ public class Response {
this.stream = in;
}
- /**
- * @return the HTTP response code
- */
+ /** Returns the HTTP response code */
public int getCode() {
return code;
}
@@ -97,9 +95,7 @@ public class Response {
return this.stream;
}
- /**
- * @return the HTTP response headers
- */
+ /** Returns the HTTP response headers */
public Header[] getHeaders() {
return headers;
}
@@ -113,23 +109,17 @@ public class Response {
return null;
}
- /**
- * @return the value of the Location header
- */
+ /** Returns the value of the Location header */
public String getLocation() {
return getHeader("Location");
}
- /**
- * @return true if a response body was sent
- */
+ /** Returns true if a response body was sent */
public boolean hasBody() {
return body != null;
}
- /**
- * @return the HTTP response body
- */
+ /** Returns the HTTP response body */
public byte[] getBody() {
if (body == null) {
try {
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/CellModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/CellModel.java
index ae64a009289..6e2391546bb 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/CellModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/CellModel.java
@@ -119,9 +119,7 @@ public class CellModel implements ProtobufMessageHandler, Serializable {
this.value = value;
}
- /**
- * @return the column
- */
+ /** Returns the column */
public byte[] getColumn() {
return column;
}
@@ -133,16 +131,12 @@ public class CellModel implements ProtobufMessageHandler, Serializable {
this.column = column;
}
- /**
- * @return true if the timestamp property has been specified by the user
- */
+ /** Returns true if the timestamp property has been specified by the user */
public boolean hasUserTimestamp() {
return timestamp != HConstants.LATEST_TIMESTAMP;
}
- /**
- * @return the timestamp
- */
+ /** Returns the timestamp */
public long getTimestamp() {
return timestamp;
}
@@ -154,9 +148,7 @@ public class CellModel implements ProtobufMessageHandler, Serializable {
this.timestamp = timestamp;
}
- /**
- * @return the value
- */
+ /** Returns the value */
public byte[] getValue() {
return value;
}
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/CellSetModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/CellSetModel.java
index d962c71d314..7d99ea6219a 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/CellSetModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/CellSetModel.java
@@ -98,9 +98,7 @@ public class CellSetModel implements Serializable, ProtobufMessageHandler {
rows.add(row);
}
- /**
- * @return the rows
- */
+ /** Returns the rows */
public List<RowModel> getRows() {
return rows;
}
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/ColumnSchemaModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/ColumnSchemaModel.java
index c15a96d2d67..7712641e220 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/ColumnSchemaModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/ColumnSchemaModel.java
@@ -80,17 +80,13 @@ public class ColumnSchemaModel implements Serializable {
return o != null ? o.toString() : null;
}
- /**
- * @return the column name
- */
+ /** Returns the column name */
@XmlAttribute
public String getName() {
return name;
}
- /**
- * @return the map for holding unspecified (user) attributes
- */
+ /** Returns the map for holding unspecified (user) attributes */
@XmlAnyAttribute
@JsonAnyGetter
public Map<QName, Object> getAny() {
@@ -130,57 +126,43 @@ public class ColumnSchemaModel implements Serializable {
// cannot be standard bean type getters and setters, otherwise this would
// confuse JAXB
- /**
- * @return true if the BLOCKCACHE attribute is present and true
- */
+ /** Returns true if the BLOCKCACHE attribute is present and true */
public boolean __getBlockcache() {
Object o = attrs.get(BLOCKCACHE);
return o != null ? Boolean.parseBoolean(o.toString()) : HColumnDescriptor.DEFAULT_BLOCKCACHE;
}
- /**
- * @return the value of the BLOCKSIZE attribute or its default if it is unset
- */
+ /** Returns the value of the BLOCKSIZE attribute or its default if it is unset */
public int __getBlocksize() {
Object o = attrs.get(BLOCKSIZE);
return o != null ? Integer.parseInt(o.toString()) : HColumnDescriptor.DEFAULT_BLOCKSIZE;
}
- /**
- * @return the value of the BLOOMFILTER attribute or its default if unset
- */
+ /** Returns the value of the BLOOMFILTER attribute or its default if unset */
public String __getBloomfilter() {
Object o = attrs.get(BLOOMFILTER);
return o != null ? o.toString() : HColumnDescriptor.DEFAULT_BLOOMFILTER;
}
- /**
- * @return the value of the COMPRESSION attribute or its default if unset
- */
+ /** Returns the value of the COMPRESSION attribute or its default if unset */
public String __getCompression() {
Object o = attrs.get(COMPRESSION);
return o != null ? o.toString() : HColumnDescriptor.DEFAULT_COMPRESSION;
}
- /**
- * @return true if the IN_MEMORY attribute is present and true
- */
+ /** Returns true if the IN_MEMORY attribute is present and true */
public boolean __getInMemory() {
Object o = attrs.get(IN_MEMORY);
return o != null ? Boolean.parseBoolean(o.toString()) : HColumnDescriptor.DEFAULT_IN_MEMORY;
}
- /**
- * @return the value of the TTL attribute or its default if it is unset
- */
+ /** Returns the value of the TTL attribute or its default if it is unset */
public int __getTTL() {
Object o = attrs.get(TTL);
return o != null ? Integer.parseInt(o.toString()) : HColumnDescriptor.DEFAULT_TTL;
}
- /**
- * @return the value of the VERSIONS attribute or its default if it is unset
- */
+ /** Returns the value of the VERSIONS attribute or its default if it is unset */
public int __getVersions() {
Object o = attrs.get(VERSIONS);
return o != null ? Integer.parseInt(o.toString()) : HColumnDescriptor.DEFAULT_VERSIONS;
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/NamespacesInstanceModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/NamespacesInstanceModel.java
index aeeff20b041..475f9185976 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/NamespacesInstanceModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/NamespacesInstanceModel.java
@@ -103,9 +103,7 @@ public class NamespacesInstanceModel implements Serializable, ProtobufMessageHan
properties.put(key, value);
}
- /**
- * @return The map of uncategorized namespace properties.
- */
+ /** Returns The map of uncategorized namespace properties. */
public Map<String, String> getProperties() {
if (properties == null) {
properties = new HashMap<>();
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/NamespacesModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/NamespacesModel.java
index cfd94d1ebc4..3b0e2d1e58b 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/NamespacesModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/NamespacesModel.java
@@ -67,9 +67,7 @@ public class NamespacesModel implements Serializable, ProtobufMessageHandler {
}
}
- /**
- * @return all namespaces
- */
+ /** Returns all namespaces */
public List<String> getNamespaces() {
return namespaces;
}
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/RowModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/RowModel.java
index 34f065105d6..f3e892aba0c 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/RowModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/RowModel.java
@@ -112,9 +112,7 @@ public class RowModel implements ProtobufMessageHandler, Serializable {
cells.add(cell);
}
- /**
- * @return the row key
- */
+ /** Returns the row key */
public byte[] getKey() {
return key;
}
@@ -126,9 +124,7 @@ public class RowModel implements ProtobufMessageHandler, Serializable {
this.key = key;
}
- /**
- * @return the cells
- */
+ /** Returns the cells */
public List<CellModel> getCells() {
return cells;
}
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/ScannerModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/ScannerModel.java
index fc59d14d56c..2eba8d4cad8 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/ScannerModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/ScannerModel.java
@@ -644,39 +644,29 @@ public class ScannerModel implements ProtobufMessageHandler, Serializable {
labels.add(label);
}
- /**
- * @return true if a start row was specified
- */
+ /** Returns true if a start row was specified */
public boolean hasStartRow() {
return !Bytes.equals(startRow, HConstants.EMPTY_START_ROW);
}
- /**
- * @return start row
- */
+ /** Returns start row */
@XmlAttribute
public byte[] getStartRow() {
return startRow;
}
- /**
- * @return true if an end row was specified
- */
+ /** Returns true if an end row was specified */
public boolean hasEndRow() {
return !Bytes.equals(endRow, HConstants.EMPTY_END_ROW);
}
- /**
- * @return end row
- */
+ /** Returns end row */
@XmlAttribute
public byte[] getEndRow() {
return endRow;
}
- /**
- * @return list of columns of interest in column:qualifier format, or empty for all
- */
+ /** Returns list of columns of interest in column:qualifier format, or empty for all */
@XmlElement(name = "column")
public List<byte[]> getColumns() {
return columns;
@@ -687,57 +677,43 @@ public class ScannerModel implements ProtobufMessageHandler, Serializable {
return labels;
}
- /**
- * @return the number of cells to return in batch
- */
+ /** Returns the number of cells to return in batch */
@XmlAttribute
public int getBatch() {
return batch;
}
- /**
- * @return the number of rows that the scanner to fetch at once
- */
+ /** Returns the number of rows that the scanner to fetch at once */
@XmlAttribute
public int getCaching() {
return caching;
}
- /**
- * @return true if HFile blocks should be cached on the servers for this scan, false otherwise
- */
+ /** Returns true if HFile blocks should be cached on the servers for this scan, false otherwise */
@XmlAttribute
public boolean getCacheBlocks() {
return cacheBlocks;
}
- /**
- * @return the lower bound on timestamps of items of interest
- */
+ /** Returns the lower bound on timestamps of items of interest */
@XmlAttribute
public long getStartTime() {
return startTime;
}
- /**
- * @return the upper bound on timestamps of items of interest
- */
+ /** Returns the upper bound on timestamps of items of interest */
@XmlAttribute
public long getEndTime() {
return endTime;
}
- /**
- * @return maximum number of versions to return
- */
+ /** Returns maximum number of versions to return */
@XmlAttribute
public int getMaxVersions() {
return maxVersions;
}
- /**
- * @return the filter specification
- */
+ /** Returns the filter specification */
@XmlElement
public String getFilter() {
return filter;
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterStatusModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterStatusModel.java
index 46bfb9091c3..027534c2888 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterStatusModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterStatusModel.java
@@ -162,105 +162,79 @@ public class StorageClusterStatusModel implements Serializable, ProtobufMessageH
this.currentCompactedKVs = currentCompactedKVs;
}
- /**
- * @return the region name
- */
+ /** Returns the region name */
@XmlAttribute
public byte[] getName() {
return name;
}
- /**
- * @return the number of stores
- */
+ /** Returns the number of stores */
@XmlAttribute
public int getStores() {
return stores;
}
- /**
- * @return the number of store files
- */
+ /** Returns the number of store files */
@XmlAttribute
public int getStorefiles() {
return storefiles;
}
- /**
- * @return the total size of store files, in MB
- */
+ /** Returns the total size of store files, in MB */
@XmlAttribute
public int getStorefileSizeMB() {
return storefileSizeMB;
}
- /**
- * @return memstore size, in MB
- */
+ /** Returns memstore size, in MB */
@XmlAttribute
public int getMemStoreSizeMB() {
return memstoreSizeMB;
}
- /**
- * @return the total size of store file indexes, in KB
- */
+ /** Returns the total size of store file indexes, in KB */
@XmlAttribute
public long getStorefileIndexSizeKB() {
return storefileIndexSizeKB;
}
- /**
- * @return the current total read requests made to region
- */
+ /** Returns the current total read requests made to region */
@XmlAttribute
public long getReadRequestsCount() {
return readRequestsCount;
}
- /**
- * @return the current total write requests made to region
- */
+ /** Returns the current total write requests made to region */
@XmlAttribute
public long getWriteRequestsCount() {
return writeRequestsCount;
}
- /**
- * @return The current total size of root-level indexes for the region, in KB.
- */
+ /** Returns The current total size of root-level indexes for the region, in KB. */
@XmlAttribute
public int getRootIndexSizeKB() {
return rootIndexSizeKB;
}
- /**
- * @return The total size of static index, in KB
- */
+ /** Returns The total size of static index, in KB */
@XmlAttribute
public int getTotalStaticIndexSizeKB() {
return totalStaticIndexSizeKB;
}
- /**
- * @return The total size of static bloom, in KB
- */
+ /** Returns The total size of static bloom, in KB */
@XmlAttribute
public int getTotalStaticBloomSizeKB() {
return totalStaticBloomSizeKB;
}
- /**
- * @return The total number of compacting key-values
- */
+ /** Returns The total number of compacting key-values */
@XmlAttribute
public long getTotalCompactingKVs() {
return totalCompactingKVs;
}
- /**
- * @return The number of current compacted key-values
- */
+ /** Returns The number of current compacted key-values */
@XmlAttribute
public long getCurrentCompactedKVs() {
return currentCompactedKVs;
@@ -405,49 +379,37 @@ public class StorageClusterStatusModel implements Serializable, ProtobufMessageH
this.startCode = startCode;
}
- /**
- * @return the region server's name
- */
+ /** Returns the region server's name */
@XmlAttribute
public String getName() {
return name;
}
- /**
- * @return the region server's start code
- */
+ /** Returns the region server's start code */
@XmlAttribute
public long getStartCode() {
return startCode;
}
- /**
- * @return the current heap size, in MB
- */
+ /** Returns the current heap size, in MB */
@XmlAttribute
public int getHeapSizeMB() {
return heapSizeMB;
}
- /**
- * @return the maximum heap size, in MB
- */
+ /** Returns the maximum heap size, in MB */
@XmlAttribute
public int getMaxHeapSizeMB() {
return maxHeapSizeMB;
}
- /**
- * @return the list of regions served by the region server
- */
+ /** Returns the list of regions served by the region server */
@XmlElement(name = "Region")
public List<Region> getRegions() {
return regions;
}
- /**
- * @return the number of requests per second processed by the region server
- */
+ /** Returns the number of requests per second processed by the region server */
@XmlAttribute
public long getRequests() {
return requests;
@@ -547,9 +509,7 @@ public class StorageClusterStatusModel implements Serializable, ProtobufMessageH
public StorageClusterStatusModel() {
}
- /**
- * @return the list of live nodes
- */
+ /** Returns the list of live nodes */
@XmlElement(name = "Node")
@XmlElementWrapper(name = "LiveNodes")
// workaround https://github.com/FasterXML/jackson-dataformat-xml/issues/192
@@ -558,9 +518,7 @@ public class StorageClusterStatusModel implements Serializable, ProtobufMessageH
return liveNodes;
}
- /**
- * @return the list of dead nodes
- */
+ /** Returns the list of dead nodes */
@XmlElement(name = "Node")
@XmlElementWrapper(name = "DeadNodes")
// workaround https://github.com/FasterXML/jackson-dataformat-xml/issues/192
@@ -569,9 +527,7 @@ public class StorageClusterStatusModel implements Serializable, ProtobufMessageH
return deadNodes;
}
- /**
- * @return the total number of regions served by the cluster
- */
+ /** Returns the total number of regions served by the cluster */
@XmlAttribute
public int getRegions() {
return regions;
@@ -586,9 +542,7 @@ public class StorageClusterStatusModel implements Serializable, ProtobufMessageH
return requests;
}
- /**
- * @return the average load of the region servers in the cluster
- */
+ /** Returns the average load of the region servers in the cluster */
@XmlAttribute
public double getAverageLoad() {
return averageLoad;
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterVersionModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterVersionModel.java
index e23c6af8177..0125e321f75 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterVersionModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterVersionModel.java
@@ -38,9 +38,7 @@ public class StorageClusterVersionModel implements Serializable {
private String version;
- /**
- * @return the storage cluster version
- */
+ /** Returns the storage cluster version */
@XmlAttribute(name = "Version")
public String getVersion() {
return version;
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableInfoModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableInfoModel.java
index 6a73c9227a5..37f7f79f89f 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableInfoModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableInfoModel.java
@@ -80,17 +80,13 @@ public class TableInfoModel implements Serializable, ProtobufMessageHandler {
return regions.get(index);
}
- /**
- * @return the table name
- */
+ /** Returns the table name */
@XmlAttribute
public String getName() {
return name;
}
- /**
- * @return the regions
- */
+ /** Returns the regions */
@XmlElement(name = "Region")
public List<TableRegionModel> getRegions() {
return regions;
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableListModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableListModel.java
index 7118d159c2e..0b7ea10ab40 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableListModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableListModel.java
@@ -61,9 +61,7 @@ public class TableListModel implements Serializable, ProtobufMessageHandler {
return tables.get(index);
}
- /**
- * @return the tables
- */
+ /** Returns the tables */
@XmlElementRef(name = "table")
public List<TableModel> getTables() {
return tables;
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableModel.java
index 75513eadbaa..51a2bc567cd 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableModel.java
@@ -55,9 +55,7 @@ public class TableModel implements Serializable {
this.name = name;
}
- /**
- * @return the name
- */
+ /** Returns the name */
@XmlAttribute
public String getName() {
return name;
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableRegionModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableRegionModel.java
index 63d955a9929..b0d4b2f47a6 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableRegionModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableRegionModel.java
@@ -83,9 +83,7 @@ public class TableRegionModel implements Serializable {
this.location = location;
}
- /**
- * @return the region name
- */
+ /** Returns the region name */
@XmlAttribute
public String getName() {
byte[] tableNameAsBytes = Bytes.toBytes(this.table);
@@ -95,33 +93,25 @@ public class TableRegionModel implements Serializable {
return Bytes.toString(nameAsBytes);
}
- /**
- * @return the encoded region id
- */
+ /** Returns the encoded region id */
@XmlAttribute
public long getId() {
return id;
}
- /**
- * @return the start key
- */
+ /** Returns the start key */
@XmlAttribute
public byte[] getStartKey() {
return startKey;
}
- /**
- * @return the end key
- */
+ /** Returns the end key */
@XmlAttribute
public byte[] getEndKey() {
return endKey;
}
- /**
- * @return the name and port of the region server hosting the region
- */
+ /** Returns the name and port of the region server hosting the region */
@XmlAttribute
public String getLocation() {
return location;
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableSchemaModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableSchemaModel.java
index 80896afa18f..1756db20d02 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableSchemaModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableSchemaModel.java
@@ -136,26 +136,20 @@ public class TableSchemaModel implements Serializable, ProtobufMessageHandler {
return columns.get(index);
}
- /**
- * @return the table name
- */
+ /** Returns the table name */
@XmlAttribute
public String getName() {
return name;
}
- /**
- * @return the map for holding unspecified (user) attributes
- */
+ /** Returns the map for holding unspecified (user) attributes */
@XmlAnyAttribute
@JsonAnyGetter
public Map<QName, Object> getAny() {
return attrs;
}
- /**
- * @return the columns
- */
+ /** Returns the columns */
@XmlElement(name = "ColumnSchema")
public List<ColumnSchemaModel> getColumns() {
return columns;
@@ -211,25 +205,19 @@ public class TableSchemaModel implements Serializable, ProtobufMessageHandler {
// cannot be standard bean type getters and setters, otherwise this would
// confuse JAXB
- /**
- * @return true if IS_META attribute exists and is truel
- */
+ /** Returns true if IS_META attribute exists and is truel */
public boolean __getIsMeta() {
Object o = attrs.get(IS_META);
return o != null && Boolean.parseBoolean(o.toString());
}
- /**
- * @return true if IS_ROOT attribute exists and is truel
- */
+ /** Returns true if IS_ROOT attribute exists and is truel */
public boolean __getIsRoot() {
Object o = attrs.get(IS_ROOT);
return o != null && Boolean.parseBoolean(o.toString());
}
- /**
- * @return true if READONLY attribute exists and is truel
- */
+ /** Returns true if READONLY attribute exists and is truel */
public boolean __getReadOnly() {
Object o = attrs.get(READONLY);
return o != null ? Boolean.parseBoolean(o.toString()) : HTableDescriptor.DEFAULT_READONLY;
@@ -324,9 +312,7 @@ public class TableSchemaModel implements Serializable, ProtobufMessageHandler {
return this;
}
- /**
- * @return a table descriptor
- */
+ /** Returns a table descriptor */
@JsonIgnore
public HTableDescriptor getTableDescriptor() {
HTableDescriptor htd = new HTableDescriptor(TableName.valueOf(getName()));
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/VersionModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/VersionModel.java
index 41331070907..74bd722d36f 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/VersionModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/VersionModel.java
@@ -74,41 +74,31 @@ public class VersionModel implements Serializable, ProtobufMessageHandler {
if (jerseyVersion == null) jerseyVersion = "";
}
- /**
- * @return the REST gateway version
- */
+ /** Returns the REST gateway version */
@XmlAttribute(name = "REST")
public String getRESTVersion() {
return restVersion;
}
- /**
- * @return the JVM vendor and version
- */
+ /** Returns the JVM vendor and version */
@XmlAttribute(name = "JVM")
public String getJVMVersion() {
return jvmVersion;
}
- /**
- * @return the OS name, version, and hardware architecture
- */
+ /** Returns the OS name, version, and hardware architecture */
@XmlAttribute(name = "OS")
public String getOSVersion() {
return osVersion;
}
- /**
- * @return the servlet container version
- */
+ /** Returns the servlet container version */
@XmlAttribute(name = "Server")
public String getServerVersion() {
return serverVersion;
}
- /**
- * @return the version of the embedded Jersey framework
- */
+ /** Returns the version of the embedded Jersey framework */
@XmlAttribute(name = "Jersey")
public String getJerseyVersion() {
return jerseyVersion;
diff --git a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupAdminServer.java b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupAdminServer.java
index 16d562bbb9f..bf535188c5b 100644
--- a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupAdminServer.java
+++ b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupAdminServer.java
@@ -119,9 +119,7 @@ public class RSGroupAdminServer implements RSGroupAdmin {
return rsGroupInfo;
}
- /**
- * @return List of Regions associated with this <code>server</code>.
- */
+ /** Returns List of Regions associated with this <code>server</code>. */
private List<RegionInfo> getRegions(final Address server) {
LinkedList<RegionInfo> regions = new LinkedList<>();
for (Map.Entry<RegionInfo, ServerName> el : master.getAssignmentManager().getRegionStates()
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/HDFSBlocksDistribution.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/HDFSBlocksDistribution.java
index d4aa962f2b4..dcddb74b71b 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/HDFSBlocksDistribution.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/HDFSBlocksDistribution.java
@@ -71,23 +71,17 @@ public class HDFSBlocksDistribution {
this.weightForSsd += weightForSsd;
}
- /**
- * @return the host name
- */
+ /** Returns the host name */
public String getHost() {
return host;
}
- /**
- * @return the weight
- */
+ /** Returns the weight */
public long getWeight() {
return weight;
}
- /**
- * @return the weight for ssd
- */
+ /** Returns the weight for ssd */
public long getWeightForSsd() {
return weightForSsd;
}
@@ -186,9 +180,7 @@ public class HDFSBlocksDistribution {
}
}
- /**
- * @return the hosts and their weights
- */
+ /** Returns the hosts and their weights */
public Map<String, HostAndWeight> getHostAndWeights() {
return this.hostAndWeights;
}
@@ -209,9 +201,7 @@ public class HDFSBlocksDistribution {
return weight;
}
- /**
- * @return the sum of all unique blocks' weight
- */
+ /** Returns the sum of all unique blocks' weight */
public long getUniqueBlocksTotalWeight() {
return uniqueBlocksTotalWeight;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/LocalHBaseCluster.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/LocalHBaseCluster.java
index bc562b0a532..a803b6aee9d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/LocalHBaseCluster.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/LocalHBaseCluster.java
@@ -247,9 +247,7 @@ public class LocalHBaseCluster {
return regionThreads.get(serverNumber).getRegionServer();
}
- /**
- * @return Read-only list of region server threads.
- */
+ /** Returns Read-only list of region server threads. */
public List<JVMClusterUtil.RegionServerThread> getRegionServers() {
return Collections.unmodifiableList(this.regionThreads);
}
@@ -268,9 +266,7 @@ public class LocalHBaseCluster {
return liveServers;
}
- /**
- * @return the Configuration used by this LocalHBaseCluster
- */
+ /** Returns the Configuration used by this LocalHBaseCluster */
public Configuration getConfiguration() {
return this.conf;
}
@@ -306,9 +302,7 @@ public class LocalHBaseCluster {
return rst.getName();
}
- /**
- * @return the HMaster thread
- */
+ /** Returns the HMaster thread */
public HMaster getMaster(int serverNumber) {
return masterThreads.get(serverNumber).getMaster();
}
@@ -328,9 +322,7 @@ public class LocalHBaseCluster {
return null;
}
- /**
- * @return Read-only list of master threads.
- */
+ /** Returns Read-only list of master threads. */
public List<JVMClusterUtil.MasterThread> getMasters() {
return Collections.unmodifiableList(this.masterThreads);
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/MetaRegionLocationCache.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/MetaRegionLocationCache.java
index c1ebb6fb9b4..6f539b3d173 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/MetaRegionLocationCache.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/MetaRegionLocationCache.java
@@ -201,9 +201,7 @@ public class MetaRegionLocationCache extends ZKListener {
cachedMetaLocations.put(replicaId, location);
}
- /**
- * @return Optional list of HRegionLocations for meta replica(s), null if the cache is empty.
- */
+ /** Returns Optional list of HRegionLocations for meta replica(s), null if the cache is empty. */
public Optional<List<HRegionLocation>> getMetaRegionLocations() {
ConcurrentNavigableMap<Integer, HRegionLocation> snapshot =
cachedMetaLocations.tailMap(cachedMetaLocations.firstKey());
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/Server.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/Server.java
index 0476a2acb7a..69ffd64e02d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/Server.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/Server.java
@@ -58,9 +58,7 @@ public interface Server extends Abortable, Stoppable {
*/
ClusterConnection getClusterConnection();
- /**
- * @return The unique server name for this server.
- */
+ /** Returns The unique server name for this server. */
ServerName getServerName();
/**
@@ -68,14 +66,10 @@ public interface Server extends Abortable, Stoppable {
*/
CoordinatedStateManager getCoordinatedStateManager();
- /**
- * @return The {@link ChoreService} instance for this server
- */
+ /** Returns The {@link ChoreService} instance for this server */
ChoreService getChoreService();
- /**
- * @return Return the FileSystem object used (can return null!).
- */
+ /** Returns Return the FileSystem object used (can return null!). */
// TODO: Distinguish between "dataFs" and "walFs".
default FileSystem getFileSystem() {
// This default is pretty dodgy!
@@ -91,9 +85,7 @@ public interface Server extends Abortable, Stoppable {
return fs;
};
- /**
- * @return True is the server is Stopping
- */
+ /** Returns True is the server is Stopping */
// Note: This method is not part of the Stoppable Interface.
default boolean isStopping() {
return false;
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java
index 1dc17eff0d0..7ef131e9e6a 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java
@@ -35,9 +35,7 @@ public interface TableDescriptors {
return get(tableName) != null;
}
- /**
- * @return TableDescriptor for tablename
- */
+ /** Returns TableDescriptor for tablename */
TableDescriptor get(TableName tableName) throws IOException;
/**
@@ -72,8 +70,6 @@ public interface TableDescriptors {
*/
void update(TableDescriptor htd, boolean cacheOnly) throws IOException;
- /**
- * @return Instance of table descriptor or null if none found.
- */
+ /** Returns Instance of table descriptor or null if none found. */
TableDescriptor remove(TableName tablename) throws IOException;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/HFileArchiver.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/HFileArchiver.java
index 8e666308d4f..68dc87502e0 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/HFileArchiver.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/HFileArchiver.java
@@ -80,9 +80,7 @@ public class HFileArchiver {
// hidden ctor since this is just a util
}
- /**
- * @return True if the Region exits in the filesystem.
- */
+ /** Returns True if the Region exits in the filesystem. */
public static boolean exists(Configuration conf, FileSystem fs, RegionInfo info)
throws IOException {
Path rootDir = CommonFSUtils.getRootDir(conf);
@@ -703,14 +701,10 @@ public class HFileArchiver {
*/
abstract void close() throws IOException;
- /**
- * @return the name of the file (not the full fs path, just the individual file name)
- */
+ /** Returns the name of the file (not the full fs path, just the individual file name) */
abstract String getName();
- /**
- * @return the path to this file
- */
+ /** Returns the path to this file */
abstract Path getPath();
/**
@@ -722,9 +716,7 @@ public class HFileArchiver {
return CommonFSUtils.renameAndSetModifyTime(fs, p, dest);
}
- /**
- * @return the {@link FileSystem} on which this file resides
- */
+ /** Returns the {@link FileSystem} on which this file resides */
public FileSystem getFileSystem() {
return this.fs;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/example/TableHFileArchiveTracker.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/example/TableHFileArchiveTracker.java
index 1896199ede2..9ae06f2840c 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/example/TableHFileArchiveTracker.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/example/TableHFileArchiveTracker.java
@@ -222,9 +222,7 @@ public final class TableHFileArchiveTracker extends ZKListener {
return getMonitor().shouldArchiveTable(tableName);
}
- /**
- * @return the tracker for which tables should be archived.
- */
+ /** Returns the tracker for which tables should be archived. */
public final HFileArchiveTableMonitor getMonitor() {
return this.monitor;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/client/VersionInfoUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/client/VersionInfoUtil.java
index ec01bc5b51b..e9703cb8cb2 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/client/VersionInfoUtil.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/client/VersionInfoUtil.java
@@ -65,9 +65,7 @@ public final class VersionInfoUtil {
return false;
}
- /**
- * @return the versionInfo extracted from the current RpcCallContext
- */
+ /** Returns the versionInfo extracted from the current RpcCallContext */
public static HBaseProtos.VersionInfo getCurrentClientVersionInfo() {
return RpcServer.getCurrentCall().map(RpcCallContext::getClientVersionInfo).orElse(null);
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/client/locking/EntityLock.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/client/locking/EntityLock.java
index 50c6dc27657..747ccb54e83 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/client/locking/EntityLock.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/client/locking/EntityLock.java
@@ -208,9 +208,7 @@ public class EntityLock {
setDaemon(true);
}
- /**
- * @return Shuts down the thread clean and quietly.
- */
+ /** Returns Shuts down the thread clean and quietly. */
Thread shutdown() {
shutdown = true;
interrupt();
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/SplitLogManagerCoordination.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/SplitLogManagerCoordination.java
index 324a6f37b99..29aa273b2b3 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/SplitLogManagerCoordination.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/SplitLogManagerCoordination.java
@@ -56,30 +56,22 @@ public interface SplitLogManagerCoordination {
this.failedDeletions = failedDeletions;
}
- /**
- * @return the master value
- */
+ /** Returns the master value */
public MasterServices getMaster() {
return master;
}
- /**
- * @return map of tasks
- */
+ /** Returns map of tasks */
public ConcurrentMap<String, Task> getTasks() {
return tasks;
}
- /**
- * @return a set of failed deletions
- */
+ /** Returns a set of failed deletions */
public Set<String> getFailedDeletions() {
return failedDeletions;
}
- /**
- * @return server name
- */
+ /** Returns server name */
public ServerName getServerName() {
return master.getServerName();
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/SplitLogWorkerCoordination.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/SplitLogWorkerCoordination.java
index ff2fbfbe7e7..3ab8ee83531 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/SplitLogWorkerCoordination.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/SplitLogWorkerCoordination.java
@@ -64,9 +64,7 @@ public interface SplitLogWorkerCoordination {
*/
void stopProcessingTasks();
- /**
- * @return the current value of exitWorker
- */
+ /** Returns the current value of exitWorker */
boolean isStop();
/**
@@ -123,9 +121,7 @@ public interface SplitLogWorkerCoordination {
*/
interface SplitTaskDetails {
- /**
- * @return full file path in HDFS for the WAL file to be split.
- */
+ /** Returns full file path in HDFS for the WAL file to be split. */
String getWALFile();
}
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/ZkSplitLogWorkerCoordination.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/ZkSplitLogWorkerCoordination.java
index 1b255b25e17..7acb0891dbc 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/ZkSplitLogWorkerCoordination.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/ZkSplitLogWorkerCoordination.java
@@ -317,9 +317,7 @@ public class ZkSplitLogWorkerCoordination extends ZKListener implements SplitLog
server.getExecutorService().submit(hsh);
}
- /**
- * @return true if more splitters are available, otherwise false.
- */
+ /** Returns true if more splitters are available, otherwise false. */
private boolean areSplittersAvailable() {
return maxConcurrentTasks - tasksInProgress.get() > 0;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseEnvironment.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseEnvironment.java
index b8cabe8cfed..b5f0809c046 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseEnvironment.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseEnvironment.java
@@ -118,13 +118,13 @@ public class BaseEnvironment<C extends Coprocessor> implements CoprocessorEnviro
return seq;
}
- /** @return the coprocessor environment version */
+ /** Returns the coprocessor environment version */
@Override
public int getVersion() {
return Coprocessor.VERSION;
}
- /** @return the HBase release */
+ /** Returns the HBase release */
@Override
public String getHBaseVersion() {
return VersionInfo.getVersion();
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/HasMasterServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/HasMasterServices.java
index 2682b78fd51..dd8babf21dd 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/HasMasterServices.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/HasMasterServices.java
@@ -30,8 +30,6 @@ import org.apache.yetus.audience.InterfaceAudience;
@Deprecated
@InterfaceAudience.Private
public interface HasMasterServices {
- /**
- * @return An instance of RegionServerServices, an object NOT for Coprocessor consumption.
- */
+ /** Returns An instance of RegionServerServices, an object NOT for Coprocessor consumption. */
MasterServices getMasterServices();
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/HasRegionServerServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/HasRegionServerServices.java
index cef03390acb..21301efeafa 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/HasRegionServerServices.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/HasRegionServerServices.java
@@ -30,8 +30,6 @@ import org.apache.yetus.audience.InterfaceAudience;
@Deprecated
@InterfaceAudience.Private
public interface HasRegionServerServices {
- /**
- * @return An instance of RegionServerServices, an object NOT for Coprocessor consumption.
- */
+ /** Returns An instance of RegionServerServices, an object NOT for Coprocessor consumption. */
RegionServerServices getRegionServerServices();
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterCoprocessorEnvironment.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterCoprocessorEnvironment.java
index c83b9da4308..61c98346887 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterCoprocessorEnvironment.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterCoprocessorEnvironment.java
@@ -30,9 +30,7 @@ import org.apache.yetus.audience.InterfaceStability;
@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.COPROC)
@InterfaceStability.Evolving
public interface MasterCoprocessorEnvironment extends CoprocessorEnvironment<MasterCoprocessor> {
- /**
- * @return Hosting Server's ServerName
- */
+ /** Returns Hosting Server's ServerName */
ServerName getServerName();
/**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.java
index 97d572d3727..8967f596379 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.java
@@ -36,23 +36,19 @@ import org.apache.yetus.audience.InterfaceStability;
@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.COPROC)
@InterfaceStability.Evolving
public interface RegionCoprocessorEnvironment extends CoprocessorEnvironment<RegionCoprocessor> {
- /** @return the region associated with this coprocessor */
+ /** Returns the region associated with this coprocessor */
Region getRegion();
- /** @return region information for the region this coprocessor is running on */
+ /** Returns region information for the region this coprocessor is running on */
RegionInfo getRegionInfo();
- /**
- * @return Interface to Map of regions online on this RegionServer {@link #getServerName()}}.
- */
+ /** Returns Interface to Map of regions online on this RegionServer {@link #getServerName()}}. */
OnlineRegions getOnlineRegions();
- /** @return shared data between all instances of this coprocessor */
+ /** Returns shared data between all instances of this coprocessor */
ConcurrentMap<String, Object> getSharedData();
- /**
- * @return Hosting Server's ServerName
- */
+ /** Returns Hosting Server's ServerName */
ServerName getServerName();
/**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionServerCoprocessorEnvironment.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionServerCoprocessorEnvironment.java
index 4b2e8b5791b..39d186aba29 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionServerCoprocessorEnvironment.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionServerCoprocessorEnvironment.java
@@ -32,14 +32,10 @@ import org.apache.yetus.audience.InterfaceStability;
@InterfaceStability.Evolving
public interface RegionServerCoprocessorEnvironment
extends CoprocessorEnvironment<RegionServerCoprocessor> {
- /**
- * @return Hosting Server's ServerName
- */
+ /** Returns Hosting Server's ServerName */
ServerName getServerName();
- /**
- * @return Interface to Map of regions online on this RegionServer {@link #getServerName()}}.
- */
+ /** Returns Interface to Map of regions online on this RegionServer {@link #getServerName()}}. */
OnlineRegions getOnlineRegions();
/**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/WALCoprocessorEnvironment.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/WALCoprocessorEnvironment.java
index 1774481f210..2d414111db5 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/WALCoprocessorEnvironment.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/WALCoprocessorEnvironment.java
@@ -27,7 +27,7 @@ import org.apache.yetus.audience.InterfaceStability;
@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.COPROC)
@InterfaceStability.Evolving
public interface WALCoprocessorEnvironment extends CoprocessorEnvironment<WALCoprocessor> {
- /** @return reference to the region server's WAL */
+ /** Returns reference to the region server's WAL */
WAL getWAL();
/**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/EventHandler.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/EventHandler.java
index fcb60a35c4c..94418f0c381 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/EventHandler.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/EventHandler.java
@@ -134,9 +134,7 @@ public abstract class EventHandler implements Runnable, Comparable<EventHandler>
return Integer.MAX_VALUE;
}
- /**
- * @return This events' sequence id.
- */
+ /** Returns This events' sequence id. */
public long getSeqid() {
return this.seqid;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/ExecutorService.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/ExecutorService.java
index eafbad1ecba..a60a1ed7512 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/ExecutorService.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/ExecutorService.java
@@ -214,7 +214,7 @@ public class ExecutorService {
}
/**
- * @return the executor name inferred from the type and the servername on which this is running.
+ * Returns the executor name inferred from the type and the servername on which this is running.
*/
public String getName() {
return getExecutorType().getExecutorName(servername);
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/ExecutorType.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/ExecutorType.java
index a8c62818f15..320842985d4 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/ExecutorType.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/ExecutorType.java
@@ -58,9 +58,7 @@ public enum ExecutorType {
ExecutorType(int value) {
}
- /**
- * @return Conflation of the executor type and the passed {@code serverName}.
- */
+ /** Returns Conflation of the executor type and the passed {@code serverName}. */
String getExecutorName(String serverName) {
return this.toString() + "-" + serverName.replace("%", "%%");
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/favored/FavoredNodesPlan.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/favored/FavoredNodesPlan.java
index 9fbcf288e92..0f0304ab9db 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/favored/FavoredNodesPlan.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/favored/FavoredNodesPlan.java
@@ -71,9 +71,7 @@ public class FavoredNodesPlan {
return favoredNodesMap.remove(region.getRegionNameAsString());
}
- /**
- * @return the list of favored region server for this region based on the plan
- */
+ /** Returns the list of favored region server for this region based on the plan */
public List<ServerName> getFavoredNodes(RegionInfo region) {
return favoredNodesMap.get(region.getRegionNameAsString());
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/filter/FilterWrapper.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/filter/FilterWrapper.java
index f2807ea023b..c31fc25ec82 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/filter/FilterWrapper.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/filter/FilterWrapper.java
@@ -45,9 +45,7 @@ final public class FilterWrapper extends Filter {
this.filter = filter;
}
- /**
- * @return The filter serialized using pb
- */
+ /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() throws IOException {
FilterProtos.FilterWrapper.Builder builder = FilterProtos.FilterWrapper.newBuilder();
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/FSDataInputStreamWrapper.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/FSDataInputStreamWrapper.java
index 819c5651081..3ad07622a4f 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/FSDataInputStreamWrapper.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/FSDataInputStreamWrapper.java
@@ -186,9 +186,7 @@ public class FSDataInputStreamWrapper implements Closeable {
readahead = 0;
}
- /**
- * @return Whether we are presently using HBase checksum.
- */
+ /** Returns Whether we are presently using HBase checksum. */
public boolean shouldUseHBaseChecksum() {
return this.useHBaseChecksum;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/FileLink.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/FileLink.java
index 2d12fd88c11..8fbaa775391 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/FileLink.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/FileLink.java
@@ -335,9 +335,7 @@ public class FileLink {
this.locations = locations.toArray(new Path[locations.size()]);
}
- /**
- * @return the locations to look for the linked file.
- */
+ /** Returns the locations to look for the linked file. */
public Path[] getLocations() {
return locations;
}
@@ -354,9 +352,7 @@ public class FileLink {
return str.toString();
}
- /**
- * @return true if the file pointed by the link exists
- */
+ /** Returns true if the file pointed by the link exists */
public boolean exists(final FileSystem fs) throws IOException {
for (int i = 0; i < locations.length; ++i) {
if (fs.exists(locations[i])) {
@@ -366,9 +362,7 @@ public class FileLink {
return false;
}
- /**
- * @return the path of the first available link.
- */
+ /** Returns the path of the first available link. */
public Path getAvailablePath(FileSystem fs) throws IOException {
for (int i = 0; i < locations.length; ++i) {
if (fs.exists(locations[i])) {
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/HFileLink.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/HFileLink.java
index 4cf350004a3..4184d6ca21e 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/HFileLink.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/HFileLink.java
@@ -160,23 +160,17 @@ public class HFileLink extends FileLink {
return HFileLink.buildFromHFileLinkPattern(conf, createPath(table, region, family, hfile));
}
- /**
- * @return the origin path of the hfile.
- */
+ /** Returns the origin path of the hfile. */
public Path getOriginPath() {
return this.originPath;
}
- /**
- * @return the path of the archived hfile.
- */
+ /** Returns the path of the archived hfile. */
public Path getArchivePath() {
return this.archivePath;
}
- /**
- * @return the path of the mob hfiles.
- */
+ /** Returns the path of the mob hfiles. */
public Path getMobPath() {
return this.mobPath;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java
index 7cbfad4b0ac..52eaa30317a 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java
@@ -118,15 +118,11 @@ public interface BlockCache extends Iterable<CachedBlock> {
*/
long getDataBlockCount();
- /**
- * @return Iterator over the blocks in the cache.
- */
+ /** Returns Iterator over the blocks in the cache. */
@Override
Iterator<CachedBlock> iterator();
- /**
- * @return The list of sub blockcaches that make up this one; returns null if no sub caches.
- */
+ /** Returns The list of sub blockcaches that make up this one; returns null if no sub caches. */
BlockCache[] getBlockCaches();
/**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCacheKey.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCacheKey.java
index f9c52e4acbc..81da24f9341 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCacheKey.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCacheKey.java
@@ -82,9 +82,7 @@ public class BlockCacheKey implements HeapSize, java.io.Serializable {
}
// can't avoid this unfortunately
- /**
- * @return The hfileName portion of this cache key
- */
+ /** Returns The hfileName portion of this cache key */
public String getHfileName() {
return hfileName;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCacheUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCacheUtil.java
index 2a2faae3eae..1bad6c5023d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCacheUtil.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCacheUtil.java
@@ -116,9 +116,7 @@ public class BlockCacheUtil {
}
}
- /**
- * @return A JSON String of <code>filename</code> and counts of <code>blocks</code>
- */
+ /** Returns A JSON String of <code>filename</code> and counts of <code>blocks</code> */
public static String toJSON(String filename, NavigableSet<CachedBlock> blocks)
throws IOException {
CachedBlockCountsPerFile counts = new CachedBlockCountsPerFile(filename);
@@ -134,16 +132,12 @@ public class BlockCacheUtil {
return GSON.toJson(counts);
}
- /**
- * @return JSON string of <code>cbsf</code> aggregated
- */
+ /** Returns JSON string of <code>cbsf</code> aggregated */
public static String toJSON(CachedBlocksByFile cbsbf) throws IOException {
return GSON.toJson(cbsbf);
}
- /**
- * @return JSON string of <code>bc</code> content.
- */
+ /** Returns JSON string of <code>bc</code> content. */
public static String toJSON(BlockCache bc) throws IOException {
return GSON.toJson(bc);
}
@@ -318,9 +312,7 @@ public class BlockCacheUtil {
return this.cachedBlockByFile;
}
- /**
- * @return count of blocks in the cache
- */
+ /** Returns count of blocks in the cache */
public int getCount() {
return count;
}
@@ -329,16 +321,12 @@ public class BlockCacheUtil {
return dataBlockCount;
}
- /**
- * @return size of blocks in the cache
- */
+ /** Returns size of blocks in the cache */
public long getSize() {
return size;
}
- /**
- * @return Size of data.
- */
+ /** Returns Size of data. */
public long getDataSize() {
return dataSize;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheConfig.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheConfig.java
index 9e456afcebc..32c6dfd6c2a 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheConfig.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheConfig.java
@@ -255,9 +255,7 @@ public class CacheConfig {
|| (prefetchOnOpen && (category != BlockCategory.META && category != BlockCategory.UNKNOWN));
}
- /**
- * @return true if blocks in this file should be flagged as in-memory
- */
+ /** Returns true if blocks in this file should be flagged as in-memory */
public boolean isInMemory() {
return this.inMemory;
}
@@ -320,15 +318,13 @@ public class CacheConfig {
this.evictOnClose = evictOnClose;
}
- /**
- * @return true if data blocks should be compressed in the cache, false if not
- */
+ /** Returns true if data blocks should be compressed in the cache, false if not */
public boolean shouldCacheDataCompressed() {
return this.cacheDataOnRead && this.cacheDataCompressed;
}
/**
- * @return true if this {@link BlockCategory} should be compressed in blockcache, false otherwise
+ * Returns true if this {@link BlockCategory} should be compressed in blockcache, false otherwise
*/
public boolean shouldCacheCompressed(BlockCategory category) {
switch (category) {
@@ -339,23 +335,17 @@ public class CacheConfig {
}
}
- /**
- * @return true if blocks should be prefetched into the cache on open, false if not
- */
+ /** Returns true if blocks should be prefetched into the cache on open, false if not */
public boolean shouldPrefetchOnOpen() {
return this.prefetchOnOpen;
}
- /**
- * @return true if blocks should be cached while writing during compaction, false if not
- */
+ /** Returns true if blocks should be cached while writing during compaction, false if not */
public boolean shouldCacheCompactedBlocksOnWrite() {
return this.cacheCompactedDataOnWrite;
}
- /**
- * @return total file size in bytes threshold for caching while writing during compaction
- */
+ /** Returns total file size in bytes threshold for caching while writing during compaction */
public long getCacheCompactedBlocksOnWriteThreshold() {
return this.cacheCompactedDataOnWriteThreshold;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/Cacheable.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/Cacheable.java
index 0a41b53a804..43456d7857d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/Cacheable.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/Cacheable.java
@@ -50,9 +50,7 @@ public interface Cacheable extends HeapSize, HBaseReferenceCounted {
*/
CacheableDeserializer<Cacheable> getDeserializer();
- /**
- * @return the block type of this cached HFile block
- */
+ /** Returns the block type of this cached HFile block */
BlockType getBlockType();
/******************************* ReferenceCounted Interfaces ***********************************/
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java
index ba7ff84996f..89a407b3389 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java
@@ -211,7 +211,7 @@ public final class HFile {
/** Add an element to the file info map. */
void appendFileInfo(byte[] key, byte[] value) throws IOException;
- /** @return the path to this {@link HFile} */
+ /** Returns the path to this {@link HFile} */
Path getPath();
/**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java
index e9bc08f7d15..7bc2d689f87 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java
@@ -434,7 +434,7 @@ public class HFileBlock implements Cacheable {
return this;
}
- /** @return get data block encoding id that was used to encode this block */
+ /** Returns get data block encoding id that was used to encode this block */
short getDataBlockEncodingId() {
if (blockType != BlockType.ENCODED_DATA) {
throw new IllegalArgumentException("Querying encoder ID of a block " + "of type other than "
@@ -443,30 +443,22 @@ public class HFileBlock implements Cacheable {
return buf.getShort(headerSize());
}
- /**
- * @return the on-disk size of header + data part + checksum.
- */
+ /** Returns the on-disk size of header + data part + checksum. */
public int getOnDiskSizeWithHeader() {
return onDiskSizeWithoutHeader + headerSize();
}
- /**
- * @return the on-disk size of the data part + checksum (header excluded).
- */
+ /** Returns the on-disk size of the data part + checksum (header excluded). */
int getOnDiskSizeWithoutHeader() {
return onDiskSizeWithoutHeader;
}
- /**
- * @return the uncompressed size of data part (header and checksum excluded).
- */
+ /** Returns the uncompressed size of data part (header and checksum excluded). */
int getUncompressedSizeWithoutHeader() {
return uncompressedSizeWithoutHeader;
}
- /**
- * @return the offset of the previous block of the same type in the file, or -1 if unknown
- */
+ /** Returns the offset of the previous block of the same type in the file, or -1 if unknown */
long getPrevBlockOffset() {
return prevBlockOffset;
}
@@ -692,9 +684,7 @@ public class HFileBlock implements Cacheable {
return offset;
}
- /**
- * @return a byte stream reading the data + checksum of this block
- */
+ /** Returns a byte stream reading the data + checksum of this block */
DataInputStream getByteStream() {
ByteBuff dup = this.buf.duplicate();
dup.position(this.headerSize());
@@ -1081,7 +1071,7 @@ public class HFileBlock implements Cacheable {
return baosInMemory.size();
}
- /** @return true if a block is being written */
+ /** Returns true if a block is being written */
boolean isWriting() {
return state == State.WRITING;
}
@@ -1545,7 +1535,7 @@ public class HFileBlock implements Cacheable {
}
/**
- * @return Check <code>onDiskSizeWithHeaderL</code> size is healthy and then return it as an int
+ * Returns Check <code>onDiskSizeWithHeaderL</code> size is healthy and then return it as an int
*/
private static int checkAndGetSizeAsInt(final long onDiskSizeWithHeaderL, final int hdrSize)
throws IOException {
@@ -1922,7 +1912,7 @@ public class HFileBlock implements Cacheable {
return this.fileContext.getBytesPerChecksum();
}
- /** @return the size of data on disk + header. Excludes checksum. */
+ /** Returns the size of data on disk + header. Excludes checksum. */
int getOnDiskDataSizeWithHeader() {
return this.onDiskDataSizeWithHeader;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java
index 782383d697b..0e886d9873c 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java
@@ -508,9 +508,7 @@ public class HFileBlockIndex {
*/
protected int searchTreeLevel;
- /**
- * @return true if the block index is empty.
- */
+ /** Returns true if the block index is empty. */
public abstract boolean isEmpty();
/**
@@ -585,9 +583,7 @@ public class HFileBlockIndex {
return blockDataSizes[i];
}
- /**
- * @return the number of root-level blocks in this block index
- */
+ /** Returns the number of root-level blocks in this block index */
public int getRootBlockCount() {
return rootCount;
}
@@ -1111,16 +1107,12 @@ public class HFileBlockIndex {
curFirstKey = null;
}
- /**
- * @return how many block index entries there are in the root level
- */
+ /** Returns how many block index entries there are in the root level */
public final int getNumRootEntries() {
return rootChunk.getNumEntries();
}
- /**
- * @return the number of levels in this block index.
- */
+ /** Returns the number of levels in this block index. */
public int getNumLevels() {
return numLevels;
}
@@ -1462,9 +1454,7 @@ public class HFileBlockIndex {
}
}
- /**
- * @return the size of this chunk if stored in the non-root index block format
- */
+ /** Returns the size of this chunk if stored in the non-root index block format */
int getNonRootSize() {
return Bytes.SIZEOF_INT // Number of entries
+ Bytes.SIZEOF_INT * (blockKeys.size() + 1) // Secondary index
@@ -1486,16 +1476,12 @@ public class HFileBlockIndex {
}
}
- /**
- * @return the size of this chunk if stored in the root index block format
- */
+ /** Returns the size of this chunk if stored in the root index block format */
int getRootSize() {
return curTotalRootSize;
}
- /**
- * @return the number of entries in this block index chunk
- */
+ /** Returns the number of entries in this block index chunk */
public int getNumEntries() {
return blockKeys.size();
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileDataBlockEncoder.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileDataBlockEncoder.java
index cf253e20bd8..cb2d5bbcfb6 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileDataBlockEncoder.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileDataBlockEncoder.java
@@ -70,7 +70,7 @@ public interface HFileDataBlockEncoder {
*/
void saveMetadata(HFile.Writer writer) throws IOException;
- /** @return the data block encoding */
+ /** Returns the data block encoding */
DataBlockEncoding getDataBlockEncoding();
/**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.java
index c74bbb1bdb7..fc2ffaac876 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.java
@@ -214,13 +214,13 @@ public abstract class HFileReaderImpl implements HFile.Reader, Configurable {
return getLastKey().map(CellUtil::cloneRow);
}
- /** @return number of KV entries in this HFile */
+ /** Returns number of KV entries in this HFile */
@Override
public long getEntries() {
return trailer.getEntryCount();
}
- /** @return comparator */
+ /** Returns comparator */
@Override
public CellComparator getComparator() {
return this.hfileContext.getCellComparator();
@@ -961,16 +961,12 @@ public abstract class HFileReaderImpl implements HFile.Reader, Configurable {
return blockSeek(key, seekBefore);
}
- /**
- * @return True if v <= 0 or v > current block buffer limit.
- */
+ /** Returns True if v <= 0 or v > current block buffer limit. */
protected final boolean checkKeyLen(final int v) {
return v <= 0 || v > this.blockBuffer.limit();
}
- /**
- * @return True if v < 0 or v > current block buffer limit.
- */
+ /** Returns True if v < 0 or v > current block buffer limit. */
protected final boolean checkLen(final int v) {
return v < 0 || v > this.blockBuffer.limit();
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileScanner.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileScanner.java
index 3e5ada1442f..e77b133523f 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileScanner.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileScanner.java
@@ -101,9 +101,7 @@ public interface HFileScanner extends Shipper, Closeable {
*/
ByteBuffer getValue();
- /**
- * @return Instance of {@link org.apache.hadoop.hbase.Cell}.
- */
+ /** Returns Instance of {@link org.apache.hadoop.hbase.Cell}. */
Cell getCell();
/**
@@ -124,9 +122,7 @@ public interface HFileScanner extends Shipper, Closeable {
@Deprecated
String getValueString();
- /**
- * @return Reader that underlies this Scanner instance.
- */
+ /** Returns Reader that underlies this Scanner instance. */
HFile.Reader getReader();
/**
@@ -135,9 +131,7 @@ public interface HFileScanner extends Shipper, Closeable {
*/
boolean isSeeked();
- /**
- * @return the next key in the index (the key to seek to the next block)
- */
+ /** Returns the next key in the index (the key to seek to the next block) */
Cell getNextIndexedKey();
/**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterImpl.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterImpl.java
index 9170cbef4e7..a638a443fc4 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterImpl.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterImpl.java
@@ -259,9 +259,7 @@ public class HFileWriterImpl implements HFile.Writer {
}
}
- /**
- * @return Path or null if we were passed a stream rather than a Path.
- */
+ /** Returns Path or null if we were passed a stream rather than a Path. */
@Override
public Path getPath() {
return path;
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/InlineBlockWriter.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/InlineBlockWriter.java
index 040ca6b9164..6b0c913ca1b 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/InlineBlockWriter.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/InlineBlockWriter.java
@@ -58,8 +58,6 @@ public interface InlineBlockWriter {
*/
BlockType getInlineBlockType();
- /**
- * @return true if inline blocks produced by this writer should be cached
- */
+ /** Returns true if inline blocks produced by this writer should be cached */
boolean getCacheOnWrite();
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/LruCachedBlock.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/LruCachedBlock.java
index f60e4300e4e..ddeee207f1f 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/LruCachedBlock.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/LruCachedBlock.java
@@ -80,9 +80,7 @@ public class LruCachedBlock implements HeapSize, Comparable<LruCachedBlock> {
}
}
- /**
- * @return Time we were cached at in nano seconds.
- */
+ /** Returns Time we were cached at in nano seconds. */
public long getCachedTime() {
return this.cachedTime;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/LruCachedBlockQueue.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/LruCachedBlockQueue.java
index 8e45cb772d2..272b5b0b0cd 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/LruCachedBlockQueue.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/LruCachedBlockQueue.java
@@ -89,16 +89,12 @@ public class LruCachedBlockQueue implements HeapSize {
}
}
- /**
- * @return The next element in this queue, or {@code null} if the queue is empty.
- */
+ /** Returns The next element in this queue, or {@code null} if the queue is empty. */
public LruCachedBlock poll() {
return queue.poll();
}
- /**
- * @return The last element in this queue, or {@code null} if the queue is empty.
- */
+ /** Returns The last element in this queue, or {@code null} if the queue is empty. */
public LruCachedBlock pollLast() {
return queue.pollLast();
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/CachedEntryQueue.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/CachedEntryQueue.java
index daa25cee1de..446cea55f14 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/CachedEntryQueue.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/CachedEntryQueue.java
@@ -92,16 +92,12 @@ public class CachedEntryQueue {
}
}
- /**
- * @return The next element in this queue, or {@code null} if the queue is empty.
- */
+ /** Returns The next element in this queue, or {@code null} if the queue is empty. */
public Map.Entry<BlockCacheKey, BucketEntry> poll() {
return queue.poll();
}
- /**
- * @return The last element in this queue, or {@code null} if the queue is empty.
- */
+ /** Returns The last element in this queue, or {@code null} if the queue is empty. */
public Map.Entry<BlockCacheKey, BucketEntry> pollLast() {
return queue.pollLast();
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/IOEngine.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/IOEngine.java
index 62ff19878fc..a7b73a5d886 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/IOEngine.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/IOEngine.java
@@ -28,9 +28,7 @@ import org.apache.yetus.audience.InterfaceAudience;
*/
@InterfaceAudience.Private
public interface IOEngine {
- /**
- * @return true if persistent storage is supported for the cache when shutdown
- */
+ /** Returns true if persistent storage is supported for the cache when shutdown */
boolean isPersistent();
/**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/util/MemorySizeUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/util/MemorySizeUtil.java
index f5f9d760a79..85bcc643558 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/util/MemorySizeUtil.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/util/MemorySizeUtil.java
@@ -153,9 +153,7 @@ public class MemorySizeUtil {
return DEFAULT_MEMSTORE_SIZE_LOWER_LIMIT;
}
- /**
- * @return Pair of global memstore size and memory type(ie. on heap or off heap).
- */
+ /** Returns Pair of global memstore size and memory type(ie. on heap or off heap). */
public static Pair<Long, MemoryType> getGlobalMemStoreSize(Configuration conf) {
long offheapMSGlobal = conf.getLong(OFFHEAP_MEMSTORE_SIZE_KEY, 0);// Size in MBs
if (offheapMSGlobal > 0) {
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/FastPathBalancedQueueRpcExecutor.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/FastPathBalancedQueueRpcExecutor.java
index 7dcf5c1361a..440f4d3a819 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/FastPathBalancedQueueRpcExecutor.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/FastPathBalancedQueueRpcExecutor.java
@@ -73,9 +73,7 @@ public class FastPathBalancedQueueRpcExecutor extends BalancedQueueRpcExecutor {
return handler != null ? handler.loadCallRunner(callTask) : super.dispatch(callTask);
}
- /**
- * @return Pop a Handler instance if one available ready-to-go or else return null.
- */
+ /** Returns Pop a Handler instance if one available ready-to-go or else return null. */
private FastPathRpcHandler popReadyHandler() {
return this.fastPathHandlerStack.poll();
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/QueueBalancer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/QueueBalancer.java
index a13f5d85823..2111d5b78ca 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/QueueBalancer.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/QueueBalancer.java
@@ -27,8 +27,6 @@ import org.apache.yetus.audience.InterfaceStability;
@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CONFIG)
@InterfaceStability.Stable
public interface QueueBalancer {
- /**
- * @return the index of the next queue to which a request should be inserted
- */
+ /** Returns the index of the next queue to which a request should be inserted */
int getNextQueue(CallRunner callRunner);
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcCall.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcCall.java
index e12bcf6964f..59008cb08a4 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcCall.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcCall.java
@@ -36,34 +36,22 @@ import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader
@InterfaceStability.Evolving
public interface RpcCall extends RpcCallContext {
- /**
- * @return The service of this call.
- */
+ /** Returns The service of this call. */
BlockingService getService();
- /**
- * @return The service method.
- */
+ /** Returns The service method. */
MethodDescriptor getMethod();
- /**
- * @return The call parameter message.
- */
+ /** Returns The call parameter message. */
Message getParam();
- /**
- * @return The CellScanner that can carry input and result payload.
- */
+ /** Returns The CellScanner that can carry input and result payload. */
CellScanner getCellScanner();
- /**
- * @return The timestamp when the call is constructed.
- */
+ /** Returns The timestamp when the call is constructed. */
long getReceiveTime();
- /**
- * @return The time when the call starts to be executed.
- */
+ /** Returns The time when the call starts to be executed. */
long getStartTime();
/**
@@ -71,14 +59,10 @@ public interface RpcCall extends RpcCallContext {
*/
void setStartTime(long startTime);
- /**
- * @return The timeout of this call.
- */
+ /** Returns The timeout of this call. */
int getTimeout();
- /**
- * @return The Priority of this call.
- */
+ /** Returns The Priority of this call. */
int getPriority();
/**
@@ -95,14 +79,10 @@ public interface RpcCall extends RpcCallContext {
*/
long getSize();
- /**
- * @return The request header of this call.
- */
+ /** Returns The request header of this call. */
RequestHeader getHeader();
- /**
- * @return Port of remote address in this call
- */
+ /** Returns Port of remote address in this call */
int getRemotePort();
/**
@@ -125,8 +105,6 @@ public interface RpcCall extends RpcCallContext {
*/
void cleanup();
- /**
- * @return A short string format of this call without possibly lengthy params
- */
+ /** Returns A short string format of this call without possibly lengthy params */
String toShortString();
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcCallContext.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcCallContext.java
index 6f045731ecb..95843652abe 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcCallContext.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcCallContext.java
@@ -55,21 +55,15 @@ public interface RpcCallContext {
*/
Optional<User> getRequestUser();
- /**
- * @return Current request's user name or not present if none ongoing.
- */
+ /** Returns Current request's user name or not present if none ongoing. */
default Optional<String> getRequestUserName() {
return getRequestUser().map(User::getShortName);
}
- /**
- * @return Address of remote client in this call
- */
+ /** Returns Address of remote client in this call */
InetAddress getRemoteAddress();
- /**
- * @return the client version info, or null if the information is not present
- */
+ /** Returns the client version info, or null if the information is not present */
VersionInfo getClientVersionInfo();
/**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java
index fa3bc2cc63a..2ca86f57f74 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java
@@ -708,9 +708,7 @@ public abstract class RpcServer implements RpcServerInterface, ConfigurationObse
return getRequestUser().map(User::getShortName);
}
- /**
- * @return Address of remote client if a request is ongoing, else null
- */
+ /** Returns Address of remote client if a request is ongoing, else null */
public static Optional<InetAddress> getRemoteAddress() {
return getCurrentCall().map(RpcCall::getRemoteAddress);
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ActiveMasterManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ActiveMasterManager.java
index 50628f8717b..3f8d6d2a702 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ActiveMasterManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ActiveMasterManager.java
@@ -314,9 +314,7 @@ public class ActiveMasterManager extends ZKListener {
return false;
}
- /**
- * @return True if cluster has an active master.
- */
+ /** Returns True if cluster has an active master. */
boolean hasActiveMaster() {
try {
if (ZKUtil.checkExists(watcher, watcher.getZNodePaths().masterAddressZNode) >= 0) {
@@ -353,9 +351,7 @@ public class ActiveMasterManager extends ZKListener {
}
}
- /**
- * @return list of registered backup masters.
- */
+ /** Returns list of registered backup masters. */
public List<ServerName> getBackupMasters() {
return backupMasters;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
index 969fb4f5583..b29a8a76f5c 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
@@ -1811,9 +1811,7 @@ public class HMaster extends HRegionServer implements MasterServices {
}
}
- /**
- * @return Get remote side's InetAddress
- */
+ /** Returns Get remote side's InetAddress */
InetAddress getRemoteInetAddress(final int port, final long serverStartCode)
throws UnknownHostException {
// Do it out here in its own little method so can fake an address when
@@ -1831,9 +1829,7 @@ public class HMaster extends HRegionServer implements MasterServices {
return ia;
}
- /**
- * @return Maximum time we should run balancer for
- */
+ /** Returns Maximum time we should run balancer for */
private int getMaxBalancingTime() {
// if max balancing time isn't set, defaulting it to period time
int maxBalancingTime =
@@ -1842,9 +1838,7 @@ public class HMaster extends HRegionServer implements MasterServices {
return maxBalancingTime;
}
- /**
- * @return Maximum number of regions in transition
- */
+ /** Returns Maximum number of regions in transition */
private int getMaxRegionsInTransition() {
int numRegions = this.assignmentManager.getRegionStates().getRegionAssignments().size();
return Math.max((int) Math.floor(numRegions * this.maxRitPercent), 1);
@@ -2122,9 +2116,7 @@ public class HMaster extends HRegionServer implements MasterServices {
return regionNormalizerManager.normalizeRegions(targetTables, isHighPriority);
}
- /**
- * @return Client info for use as prefix on an audit log string; who did an action
- */
+ /** Returns Client info for use as prefix on an audit log string; who did an action */
@Override
public String getClientIdAuditPrefix() {
return "Client=" + RpcServer.getRequestUserName().orElse(null) + "/"
@@ -2925,9 +2917,7 @@ public class HMaster extends HRegionServer implements MasterServices {
return null;
}
- /**
- * @return cluster status
- */
+ /** Returns cluster status */
public ClusterMetrics getClusterMetrics() throws IOException {
return getClusterMetrics(EnumSet.allOf(Option.class));
}
@@ -2953,9 +2943,7 @@ public class HMaster extends HRegionServer implements MasterServices {
return activeMasterManager.getBackupMasters();
}
- /**
- * @return info port of active master or 0 if any exception occurs.
- */
+ /** Returns info port of active master or 0 if any exception occurs. */
public int getActiveMasterInfoPort() {
return activeMasterManager.getActiveMasterInfoPort();
}
@@ -2982,23 +2970,17 @@ public class HMaster extends HRegionServer implements MasterServices {
return CoprocessorHost.getLoadedCoprocessors().toString();
}
- /**
- * @return timestamp in millis when HMaster was started.
- */
+ /** Returns timestamp in millis when HMaster was started. */
public long getMasterStartTime() {
return startcode;
}
- /**
- * @return timestamp in millis when HMaster became the active master.
- */
+ /** Returns timestamp in millis when HMaster became the active master. */
public long getMasterActiveTime() {
return masterActiveTime;
}
- /**
- * @return timestamp in millis when HMaster finished becoming the active master
- */
+ /** Returns timestamp in millis when HMaster finished becoming the active master */
public long getMasterFinishedInitializationTime() {
return masterFinishedInitializationTime;
}
@@ -3030,9 +3012,7 @@ public class HMaster extends HRegionServer implements MasterServices {
assignmentManager.checkIfShouldMoveSystemRegionAsync();
}
- /**
- * @return array of coprocessor SimpleNames.
- */
+ /** Returns array of coprocessor SimpleNames. */
public String[] getMasterCoprocessors() {
Set<String> masterCoprocessors = getMasterCoprocessorHost().getCoprocessors();
return masterCoprocessors.toArray(new String[masterCoprocessors.size()]);
@@ -3305,17 +3285,13 @@ public class HMaster extends HRegionServer implements MasterServices {
return this.logCleaner;
}
- /**
- * @return the underlying snapshot manager
- */
+ /** Returns the underlying snapshot manager */
@Override
public SnapshotManager getSnapshotManager() {
return this.snapshotManager;
}
- /**
- * @return the underlying MasterProcedureManagerHost
- */
+ /** Returns the underlying MasterProcedureManagerHost */
@Override
public MasterProcedureManagerHost getMasterProcedureManagerHost() {
return mpmHost;
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java
index 5b210df4b26..7803fbd9f11 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java
@@ -185,37 +185,27 @@ public class MasterFileSystem {
return this.conf;
}
- /**
- * @return HBase root dir.
- */
+ /** Returns HBase root dir. */
public Path getRootDir() {
return this.rootdir;
}
- /**
- * @return HBase root log dir.
- */
+ /** Returns HBase root log dir. */
public Path getWALRootDir() {
return this.walRootDir;
}
- /**
- * @return the directory for a give {@code region}.
- */
+ /** Returns the directory for a give {@code region}. */
public Path getRegionDir(RegionInfo region) {
return FSUtils.getRegionDirFromRootDir(getRootDir(), region);
}
- /**
- * @return HBase temp dir.
- */
+ /** Returns HBase temp dir. */
public Path getTempDir() {
return this.tempdir;
}
- /**
- * @return The unique identifier generated for this cluster
- */
+ /** Returns The unique identifier generated for this cluster */
public ClusterId getClusterId() {
return clusterId;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java
index d48671b79f5..bc3b4af2bc3 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java
@@ -502,9 +502,7 @@ public class MasterRpcServices extends RSRpcServices
return switchBalancer(b, BalanceSwitchMode.SYNC);
}
- /**
- * @return list of blocking services and their security info classes that this server supports
- */
+ /** Returns list of blocking services and their security info classes that this server supports */
@Override
protected List<BlockingServiceAndInterface> getServices() {
List<BlockingServiceAndInterface> bssi = new ArrayList<>(5);
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java
index cfaeb139331..6251bdd8eaf 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java
@@ -61,84 +61,52 @@ import org.apache.yetus.audience.InterfaceAudience;
*/
@InterfaceAudience.Private
public interface MasterServices extends Server {
- /**
- * @return the underlying snapshot manager
- */
+ /** Returns the underlying snapshot manager */
SnapshotManager getSnapshotManager();
- /**
- * @return the underlying MasterProcedureManagerHost
- */
+ /** Returns the underlying MasterProcedureManagerHost */
MasterProcedureManagerHost getMasterProcedureManagerHost();
- /**
- * @return Master's instance of {@link ClusterSchema}
- */
+ /** Returns Master's instance of {@link ClusterSchema} */
ClusterSchema getClusterSchema();
- /**
- * @return Master's instance of the {@link AssignmentManager}
- */
+ /** Returns Master's instance of the {@link AssignmentManager} */
AssignmentManager getAssignmentManager();
- /**
- * @return Master's filesystem {@link MasterFileSystem} utility class.
- */
+ /** Returns Master's filesystem {@link MasterFileSystem} utility class. */
MasterFileSystem getMasterFileSystem();
- /**
- * @return Master's WALs {@link MasterWalManager} utility class.
- */
+ /** Returns Master's WALs {@link MasterWalManager} utility class. */
MasterWalManager getMasterWalManager();
- /**
- * @return Master's {@link ServerManager} instance.
- */
+ /** Returns Master's {@link ServerManager} instance. */
ServerManager getServerManager();
- /**
- * @return Master's instance of {@link ExecutorService}
- */
+ /** Returns Master's instance of {@link ExecutorService} */
ExecutorService getExecutorService();
- /**
- * @return Master's instance of {@link TableStateManager}
- */
+ /** Returns Master's instance of {@link TableStateManager} */
TableStateManager getTableStateManager();
- /**
- * @return Master's instance of {@link MasterCoprocessorHost}
- */
+ /** Returns Master's instance of {@link MasterCoprocessorHost} */
MasterCoprocessorHost getMasterCoprocessorHost();
- /**
- * @return Master's instance of {@link MasterQuotaManager}
- */
+ /** Returns Master's instance of {@link MasterQuotaManager} */
MasterQuotaManager getMasterQuotaManager();
- /**
- * @return Master's instance of {@link RegionNormalizerManager}
- */
+ /** Returns Master's instance of {@link RegionNormalizerManager} */
RegionNormalizerManager getRegionNormalizerManager();
- /**
- * @return Master's instance of {@link CatalogJanitor}
- */
+ /** Returns Master's instance of {@link CatalogJanitor} */
CatalogJanitor getCatalogJanitor();
- /**
- * @return Master's instance of {@link ProcedureExecutor}
- */
+ /** Returns Master's instance of {@link ProcedureExecutor} */
ProcedureExecutor<MasterProcedureEnv> getMasterProcedureExecutor();
- /**
- * @return Tripped when Master has finished initialization.
- */
+ /** Returns Tripped when Master has finished initialization. */
public ProcedureEvent<?> getInitializedEvent();
- /**
- * @return Master's instance of {@link MetricsMaster}
- */
+ /** Returns Master's instance of {@link MetricsMaster} */
MetricsMaster getMasterMetrics();
/**
@@ -259,9 +227,7 @@ public interface MasterServices extends Server {
long splitRegion(final RegionInfo regionInfo, final byte[] splitRow, final long nonceGroup,
final long nonce) throws IOException;
- /**
- * @return Return table descriptors implementation.
- */
+ /** Returns Return table descriptors implementation. */
TableDescriptors getTableDescriptors();
/**
@@ -277,14 +243,10 @@ public interface MasterServices extends Server {
*/
boolean registerService(Service instance);
- /**
- * @return true if master is the active one
- */
+ /** Returns true if master is the active one */
boolean isActiveMaster();
- /**
- * @return true if master is initialized
- */
+ /** Returns true if master is initialized */
boolean isInitialized();
/**
@@ -347,16 +309,12 @@ public interface MasterServices extends Server {
*/
public long getLastMajorCompactionTimestampForRegion(byte[] regionName) throws IOException;
- /**
- * @return load balancer
- */
+ /** Returns load balancer */
public LoadBalancer getLoadBalancer();
boolean isSplitOrMergeEnabled(MasterSwitchType switchType);
- /**
- * @return Favored Nodes Manager
- */
+ /** Returns Favored Nodes Manager */
public FavoredNodesManager getFavoredNodesManager();
/**
@@ -415,9 +373,7 @@ public interface MasterServices extends Server {
List<ReplicationPeerDescription> listReplicationPeers(String regex)
throws ReplicationException, IOException;
- /**
- * @return {@link LockManager} to lock namespaces/tables/regions.
- */
+ /** Returns {@link LockManager} to lock namespaces/tables/regions. */
LockManager getLockManager();
public String getRegionServerVersion(final ServerName sn);
@@ -430,26 +386,18 @@ public interface MasterServices extends Server {
String getClientIdAuditPrefix();
- /**
- * @return True if cluster is up; false if cluster is not up (we are shutting down).
- */
+ /** Returns True if cluster is up; false if cluster is not up (we are shutting down). */
boolean isClusterUp();
- /**
- * @return return null if current is zk-based WAL splitting
- */
+ /** Returns return null if current is zk-based WAL splitting */
default SplitWALManager getSplitWALManager() {
return null;
}
- /**
- * @return the {@link AccessChecker}
- */
+ /** Returns the {@link AccessChecker} */
AccessChecker getAccessChecker();
- /**
- * @return the {@link ZKPermissionWatcher}
- */
+ /** Returns the {@link ZKPermissionWatcher} */
ZKPermissionWatcher getZKPermissionWatcher();
/**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterWalManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterWalManager.java
index e6ab12c0250..125657db187 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterWalManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterWalManager.java
@@ -171,9 +171,7 @@ public class MasterWalManager {
p -> !p.getName().endsWith(AbstractFSWALProvider.SPLITTING_EXT));
}
- /**
- * @return listing of ServerNames found by parsing WAL directory paths in FS.
- */
+ /** Returns listing of ServerNames found by parsing WAL directory paths in FS. */
public Set<ServerName> getServerNamesFromWALDirPath(final PathFilter filter) throws IOException {
FileStatus[] walDirForServerNames = getWALDirPaths(filter);
return Stream.of(walDirForServerNames).map(s -> {
@@ -189,7 +187,7 @@ public class MasterWalManager {
}
/**
- * @return List of all RegionServer WAL dirs; i.e. this.rootDir/HConstants.HREGION_LOGDIR_NAME.
+ * Returns List of all RegionServer WAL dirs; i.e. this.rootDir/HConstants.HREGION_LOGDIR_NAME.
*/
public FileStatus[] getWALDirPaths(final PathFilter filter) throws IOException {
Path walDirPath = new Path(CommonFSUtils.getWALRootDir(conf), HConstants.HREGION_LOGDIR_NAME);
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MetricsAssignmentManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MetricsAssignmentManager.java
index 41575c5a7bd..eaf335e52ef 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MetricsAssignmentManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MetricsAssignmentManager.java
@@ -127,58 +127,42 @@ public class MetricsAssignmentManager {
assignmentManagerSource.setEmptyRegionInfoRegions(emptyRegionInfoRegions);
}
- /**
- * @return Set of common metrics for assign procedure
- */
+ /** Returns Set of common metrics for assign procedure */
public ProcedureMetrics getAssignProcMetrics() {
return assignProcMetrics;
}
- /**
- * @return Set of common metrics for unassign procedure
- */
+ /** Returns Set of common metrics for unassign procedure */
public ProcedureMetrics getUnassignProcMetrics() {
return unassignProcMetrics;
}
- /**
- * @return Set of common metrics for move procedure
- */
+ /** Returns Set of common metrics for move procedure */
public ProcedureMetrics getMoveProcMetrics() {
return moveProcMetrics;
}
- /**
- * @return Set of common metrics for reopen procedure
- */
+ /** Returns Set of common metrics for reopen procedure */
public ProcedureMetrics getReopenProcMetrics() {
return reopenProcMetrics;
}
- /**
- * @return Set of common metrics for OpenRegionProcedure
- */
+ /** Returns Set of common metrics for OpenRegionProcedure */
public ProcedureMetrics getOpenProcMetrics() {
return openProcMetrics;
}
- /**
- * @return Set of common metrics for CloseRegionProcedure
- */
+ /** Returns Set of common metrics for CloseRegionProcedure */
public ProcedureMetrics getCloseProcMetrics() {
return closeProcMetrics;
}
- /**
- * @return Set of common metrics for split procedure
- */
+ /** Returns Set of common metrics for split procedure */
public ProcedureMetrics getSplitProcMetrics() {
return splitProcMetrics;
}
- /**
- * @return Set of common metrics for merge procedure
- */
+ /** Returns Set of common metrics for merge procedure */
public ProcedureMetrics getMergeProcMetrics() {
return mergeProcMetrics;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MetricsMaster.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MetricsMaster.java
index bf6589bf4e8..671291c62e2 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MetricsMaster.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MetricsMaster.java
@@ -116,9 +116,7 @@ public class MetricsMaster {
masterQuotaSource.incrementSpaceQuotaObserverChoreTime(executionTime);
}
- /**
- * @return Set of metrics for assign procedure
- */
+ /** Returns Set of metrics for assign procedure */
public ProcedureMetrics getServerCrashProcMetrics() {
return serverCrashProcMetrics;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionPlacementMaintainer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionPlacementMaintainer.java
index 17b87848733..c9876aeb83c 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionPlacementMaintainer.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionPlacementMaintainer.java
@@ -128,9 +128,7 @@ public class RegionPlacementMaintainer {
}
}
- /**
- * @return the new RegionAssignmentSnapshot n
- */
+ /** Returns the new RegionAssignmentSnapshot n */
public SnapshotOfRegionAssignmentFromMeta getRegionAssignmentSnapshot() throws IOException {
SnapshotOfRegionAssignmentFromMeta currentAssignmentShapshot =
new SnapshotOfRegionAssignmentFromMeta(ConnectionFactory.createConnection(conf));
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
index faae49ff7e8..9471a9ff317 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
@@ -439,15 +439,13 @@ public class ServerManager {
return numServers == 0 ? 0 : (double) totalLoad / (double) numServers;
}
- /** @return the count of active regionservers */
+ /** Returns the count of active regionservers */
public int countOfRegionServers() {
// Presumes onlineServers is a concurrent map
return this.onlineServers.size();
}
- /**
- * @return Read-only map of servers to serverinfo
- */
+ /** Returns Read-only map of servers to serverinfo */
public Map<ServerName, ServerMetrics> getOnlineServers() {
// Presumption is that iterating the returned Map is OK.
synchronized (this.onlineServers) {
@@ -835,9 +833,7 @@ public class ServerManager {
return max == Integer.MAX_VALUE ? "NO_LIMIT" : Integer.toString(max);
}
- /**
- * @return A copy of the internal list of online servers.
- */
+ /** Returns A copy of the internal list of online servers. */
public List<ServerName> getOnlineServersList() {
// TODO: optimize the load balancer call so we don't need to make a new list
// TODO: FIX. THIS IS POPULAR CALL.
@@ -865,9 +861,7 @@ public class ServerManager {
return names;
}
- /**
- * @return A copy of the internal list of draining servers.
- */
+ /** Returns A copy of the internal list of draining servers. */
public List<ServerName> getDrainingServersList() {
return new ArrayList<>(this.drainingServers);
}
@@ -882,9 +876,7 @@ public class ServerManager {
UNKNOWN
}
- /**
- * @return whether the server is online, dead, or unknown.
- */
+ /** Returns whether the server is online, dead, or unknown. */
public synchronized ServerLiveState isServerKnownAndOnline(ServerName serverName) {
return onlineServers.containsKey(serverName)
? ServerLiveState.LIVE
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SnapshotSentinel.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SnapshotSentinel.java
index 0a564b9b9c6..e14027bf173 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SnapshotSentinel.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SnapshotSentinel.java
@@ -37,9 +37,7 @@ public interface SnapshotSentinel {
*/
boolean isFinished();
- /**
- * @return -1 if the snapshot is in progress, otherwise the completion timestamp.
- */
+ /** Returns -1 if the snapshot is in progress, otherwise the completion timestamp. */
long getCompletionTimestamp();
/**
@@ -48,9 +46,7 @@ public interface SnapshotSentinel {
*/
void cancel(String why);
- /**
- * @return the description of the snapshot being run
- */
+ /** Returns the description of the snapshot being run */
SnapshotDescription getSnapshot();
/**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignmentManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignmentManager.java
index 71dad35c0f3..6257edf04f7 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignmentManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignmentManager.java
@@ -2386,9 +2386,7 @@ public class AssignmentManager {
return master;
}
- /**
- * @return a snapshot of rsReports
- */
+ /** Returns a snapshot of rsReports */
public Map<ServerName, Set<byte[]>> getRSReports() {
Map<ServerName, Set<byte[]>> rsReportsSnapshot = new HashMap<>();
synchronized (rsReports) {
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.java
index d00394a6fe3..882716a17ac 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.java
@@ -745,9 +745,7 @@ public class MergeTableRegionsProcedure
}
}
- /**
- * @return The merged region. Maybe be null if called to early or we failed.
- */
+ /** Returns The merged region. Maybe be null if called to early or we failed. */
RegionInfo getMergedRegion() {
return this.mergedRegion;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/RegionStates.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/RegionStates.java
index 855731982ba..3053aece59d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/RegionStates.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/RegionStates.java
@@ -177,12 +177,12 @@ public class RegionStates {
return regions;
}
- /** @return A view of region state nodes for all the regions. */
+ /** Returns A view of region state nodes for all the regions. */
public Collection<RegionStateNode> getRegionStateNodes() {
return Collections.unmodifiableCollection(regionsMap.values());
}
- /** @return A snapshot of region state nodes for all the regions. */
+ /** Returns A snapshot of region state nodes for all the regions. */
public ArrayList<RegionState> getRegionStates() {
final ArrayList<RegionState> regions = new ArrayList<>(regionsMap.size());
for (RegionStateNode node : regionsMap.values()) {
@@ -217,9 +217,7 @@ public class RegionStates {
return !getTableRegionStates(tableName).isEmpty();
}
- /**
- * @return Return online regions of table; does not include OFFLINE or SPLITTING regions.
- */
+ /** Returns Return online regions of table; does not include OFFLINE or SPLITTING regions. */
public List<RegionInfo> getRegionsOfTable(TableName table) {
return getRegionsOfTable(table, regionNode -> !regionNode.isInState(State.OFFLINE, State.SPLIT)
&& !regionNode.getRegionInfo().isSplitParent());
@@ -352,9 +350,7 @@ public class RegionStates {
.collect(Collectors.toList());
}
- /**
- * @return Return the regions of the table and filter them.
- */
+ /** Returns Return the regions of the table and filter them. */
private List<RegionInfo> getRegionsOfTable(TableName table, Predicate<RegionStateNode> filter) {
return getTableRegionStateNodes(table).stream().filter(filter).map(n -> n.getRegionInfo())
.collect(Collectors.toList());
@@ -465,9 +461,7 @@ public class RegionStates {
return isRegionInState(regionInfo, State.OPEN);
}
- /**
- * @return True if region is offline (In OFFLINE or CLOSED state).
- */
+ /** Returns True if region is offline (In OFFLINE or CLOSED state). */
public boolean isRegionOffline(final RegionInfo regionInfo) {
return isRegionInState(regionInfo, State.OFFLINE, State.CLOSED);
}
@@ -762,9 +756,7 @@ public class RegionStates {
serverMap.remove(serverName);
}
- /**
- * @return Pertinent ServerStateNode or NULL if none found (Do not make modifications).
- */
+ /** Returns Pertinent ServerStateNode or NULL if none found (Do not make modifications). */
public ServerStateNode getServerNode(final ServerName serverName) {
return serverMap.get(serverName);
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.java
index 60c9609d3a2..59003412a8d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.java
@@ -130,7 +130,7 @@ public class UnassignProcedure extends RegionTransitionProcedure {
}
/**
- * @return If true, we will re-wake up this procedure; if false, the procedure stays suspended.
+ * Returns If true, we will re-wake up this procedure; if false, the procedure stays suspended.
*/
@Override
protected boolean remoteCallFailed(final MasterProcedureEnv env, final RegionStateNode regionNode,
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/FavoredStochasticBalancer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/FavoredStochasticBalancer.java
index d2c93b8ead1..7fc867bf1b9 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/FavoredStochasticBalancer.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/FavoredStochasticBalancer.java
@@ -83,9 +83,7 @@ public class FavoredStochasticBalancer extends StochasticLoadBalancer
return fnPickers;
}
- /**
- * @return any candidate generator in random
- */
+ /** Returns any candidate generator in random */
@Override
protected CandidateGenerator getRandomGenerator() {
return candidateGenerators.get(ThreadLocalRandom.current().nextInt(candidateGenerators.size()));
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/hbck/HbckChore.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/hbck/HbckChore.java
index 0a3d310c98c..85e978014e5 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/hbck/HbckChore.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/hbck/HbckChore.java
@@ -81,7 +81,7 @@ public class HbckChore extends ScheduledChore {
}
/**
- * @return Returns last published Report that comes of last successful execution of this chore.
+ * Returns Returns last published Report that comes of last successful execution of this chore.
*/
public HbckReport getLastReport() {
return lastReport;
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/janitor/CatalogJanitor.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/janitor/CatalogJanitor.java
index 4bc54aae969..0d96d584d7d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/janitor/CatalogJanitor.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/janitor/CatalogJanitor.java
@@ -237,9 +237,7 @@ public class CatalogJanitor extends ScheduledChore {
return visitor.getReport();
}
- /**
- * @return Returns last published Report that comes of last successful scan of hbase:meta.
- */
+ /** Returns Returns last published Report that comes of last successful scan of hbase:meta. */
public CatalogJanitorReport getLastReport() {
return this.lastReport;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/janitor/CatalogJanitorReport.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/janitor/CatalogJanitorReport.java
index 6a9eb7e1b4e..57ecb626d1e 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/janitor/CatalogJanitorReport.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/janitor/CatalogJanitorReport.java
@@ -85,9 +85,7 @@ public class CatalogJanitorReport {
return emptyRegionInfo;
}
- /**
- * @return True if an 'empty' lastReport -- no problems found.
- */
+ /** Returns True if an 'empty' lastReport -- no problems found. */
public boolean isEmpty() {
return this.holes.isEmpty() && this.overlaps.isEmpty() && this.unknownServers.isEmpty()
&& this.emptyRegionInfo.isEmpty();
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/janitor/ReportMakingVisitor.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/janitor/ReportMakingVisitor.java
index a61c9010685..20b24ada2a5 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/janitor/ReportMakingVisitor.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/janitor/ReportMakingVisitor.java
@@ -195,9 +195,7 @@ class ReportMakingVisitor implements MetaTableAccessor.CloseableVisitor {
this.report.holes.add(new Pair<>(a, b));
}
- /**
- * @return True if table is disabled or disabling; defaults false!
- */
+ /** Returns True if table is disabled or disabling; defaults false! */
boolean isTableDisabled(RegionInfo ri) {
if (ri == null) {
return false;
@@ -274,9 +272,7 @@ class ReportMakingVisitor implements MetaTableAccessor.CloseableVisitor {
}
}
- /**
- * @return True iff first row in hbase:meta or if we've broached a new table in hbase:meta
- */
+ /** Returns True iff first row in hbase:meta or if we've broached a new table in hbase:meta */
private boolean isTableTransition(RegionInfo ri) {
return this.previous == null || !this.previous.getTable().equals(ri.getTable());
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/NormalizationPlan.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/NormalizationPlan.java
index 435553eb8cf..0e0d39d10b1 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/NormalizationPlan.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/NormalizationPlan.java
@@ -32,8 +32,6 @@ public interface NormalizationPlan {
NONE
}
- /**
- * @return the type of this plan
- */
+ /** Returns the type of this plan */
PlanType getType();
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java
index c4da9c049ab..ceb8851645d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java
@@ -234,9 +234,7 @@ class SimpleRegionNormalizer implements RegionNormalizer, ConfigurationObserver
return plans;
}
- /**
- * @return size of region in MB and if region is not found than -1
- */
+ /** Returns size of region in MB and if region is not found than -1 */
private long getRegionSizeMB(RegionInfo hri) {
ServerName sn =
masterServices.getAssignmentManager().getRegionStates().getRegionServerOfRegion(hri);
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/AbstractStateMachineRegionProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/AbstractStateMachineRegionProcedure.java
index a6c2e52aef6..fe98a78b4d7 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/AbstractStateMachineRegionProcedure.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/AbstractStateMachineRegionProcedure.java
@@ -46,9 +46,7 @@ public abstract class AbstractStateMachineRegionProcedure<TState>
super();
}
- /**
- * @return The RegionInfo of the region we are operating on.
- */
+ /** Returns The RegionInfo of the region we are operating on. */
public RegionInfo getRegion() {
return this.hri;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/EnableTableProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/EnableTableProcedure.java
index b42cc1a2f6f..3466bf68843 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/EnableTableProcedure.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/EnableTableProcedure.java
@@ -342,9 +342,7 @@ public class EnableTableProcedure extends AbstractStateMachineTableProcedure<Ena
}
}
- /**
- * @return Maximum region replica id found in passed list of regions.
- */
+ /** Returns Maximum region replica id found in passed list of regions. */
private static int getMaxReplicaId(List<RegionInfo> regions) {
int max = 0;
for (RegionInfo regionInfo : regions) {
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.java
index 8691675972e..dff3de7b5be 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.java
@@ -267,9 +267,7 @@ public class ServerCrashProcedure extends
return Flow.HAS_MORE_STATE;
}
- /**
- * @return List of Regions on crashed server.
- */
+ /** Returns List of Regions on crashed server. */
List<RegionInfo> getRegionsOnCrashedServer(MasterProcedureEnv env) {
return env.getMasterServices().getAssignmentManager().getRegionsOnServer(serverName);
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ServerProcedureInterface.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ServerProcedureInterface.java
index 6578754a0db..1d7dcb77e09 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ServerProcedureInterface.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ServerProcedureInterface.java
@@ -58,14 +58,10 @@ public interface ServerProcedureInterface {
VERIFY_SNAPSHOT
}
- /**
- * @return Name of this server instance.
- */
+ /** Returns Name of this server instance. */
ServerName getServerName();
- /**
- * @return True if this server has an hbase:meta table region.
- */
+ /** Returns True if this server has an hbase:meta table region. */
boolean hasMetaTableRegion();
/**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/TableProcedureInterface.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/TableProcedureInterface.java
index 1a362e108f8..4b3ae5ae138 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/TableProcedureInterface.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/TableProcedureInterface.java
@@ -44,9 +44,7 @@ public interface TableProcedureInterface {
MERGED_REGIONS_GC/* region operations */
};
- /**
- * @return the name of the table the procedure is operating on
- */
+ /** Returns the name of the table the procedure is operating on */
TableName getTableName();
/**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.java
index 5d9701d9bcc..e9a33f6cb28 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.java
@@ -349,9 +349,7 @@ public class SnapshotFileCache implements Stoppable {
this.files = files;
}
- /**
- * @return the hfiles in the snapshot when <tt>this</tt> was made.
- */
+ /** Returns the hfiles in the snapshot when <tt>this</tt> was made. */
public Collection<String> getFiles() {
return this.files;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotManager.java
index 259987ca66d..56c4af7a85e 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotManager.java
@@ -824,9 +824,7 @@ public class SnapshotManager extends MasterProcedureManager implements Stoppable
}
}
- /**
- * @return distributed commit coordinator for all running snapshots
- */
+ /** Returns distributed commit coordinator for all running snapshots */
ProcedureCoordinator getCoordinator() {
return coordinator;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/zksyncer/ClientZKSyncer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/zksyncer/ClientZKSyncer.java
index a9881ef9b64..ed5b78e211d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/zksyncer/ClientZKSyncer.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/zksyncer/ClientZKSyncer.java
@@ -271,9 +271,7 @@ public abstract class ClientZKSyncer extends ZKListener {
*/
protected abstract boolean validate(String path);
- /**
- * @return the zk path(s) to watch
- */
+ /** Returns the zk path(s) to watch */
protected abstract Set<String> getPathsToWatch();
protected final void refreshWatchingList() {
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/Procedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/Procedure.java
index cd8ae58e53a..8eb477fb846 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/Procedure.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/Procedure.java
@@ -161,7 +161,7 @@ public class Procedure implements Callable<Void>, ForeignExceptionListener {
}
/**
- * @return String of the procedure members both trying to enter the barrier and already in barrier
+ * Returns String of the procedure members both trying to enter the barrier and already in barrier
*/
public String getStatus() {
String waiting, done;
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinator.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinator.java
index 2e7d7e207ea..908419a42b4 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinator.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinator.java
@@ -289,9 +289,7 @@ public class ProcedureCoordinator {
proc.barrierReleasedByMember(member, dataFromMember);
}
- /**
- * @return the rpcs implementation for all current procedures
- */
+ /** Returns the rpcs implementation for all current procedures */
ProcedureCoordinatorRpcs getRpcs() {
return rpcs;
}
@@ -306,9 +304,7 @@ public class ProcedureCoordinator {
return procedures.get(name);
}
- /**
- * @return Return set of all procedure names.
- */
+ /** Returns Return set of all procedure names. */
public Set<String> getProcedureNames() {
return new HashSet<>(procedures.keySet());
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/MasterQuotaManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/MasterQuotaManager.java
index 1132c4fcfe7..644ae6c887c 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/MasterQuotaManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/MasterQuotaManager.java
@@ -523,9 +523,7 @@ public class MasterQuotaManager implements RegionStateListener {
}
}
- /**
- * @return cached region count, or -1 if quota manager is disabled or table status not found
- */
+ /** Returns cached region count, or -1 if quota manager is disabled or table status not found */
public int getRegionCountOfTable(TableName tName) throws IOException {
if (initialized) {
return namespaceQuotaManager.getRegionCountOfTable(tName);
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/OperationQuota.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/OperationQuota.java
index 7ba88e51fd4..aaae64b6184 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/OperationQuota.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/OperationQuota.java
@@ -67,6 +67,6 @@ public interface OperationQuota {
*/
void addMutation(Mutation mutation);
- /** @return the number of bytes available to read to avoid exceeding the quota */
+ /** Returns the number of bytes available to read to avoid exceeding the quota */
long getReadAvailable();
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaLimiter.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaLimiter.java
index bbe2161a327..14326e4e0d2 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaLimiter.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaLimiter.java
@@ -70,12 +70,12 @@ public interface QuotaLimiter {
*/
void consumeRead(long size, long capacityUnit);
- /** @return true if the limiter is a noop */
+ /** Returns true if the limiter is a noop */
boolean isBypass();
- /** @return the number of bytes available to read to avoid exceeding the quota */
+ /** Returns the number of bytes available to read to avoid exceeding the quota */
long getReadAvailable();
- /** @return the number of bytes available to write to avoid exceeding the quota */
+ /** Returns the number of bytes available to write to avoid exceeding the quota */
long getWriteAvailable();
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaState.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaState.java
index 6f7b351ef09..7c9445e1558 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaState.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaState.java
@@ -69,9 +69,7 @@ public class QuotaState {
return builder.toString();
}
- /**
- * @return true if there is no quota information associated to this object
- */
+ /** Returns true if there is no quota information associated to this object */
public synchronized boolean isBypass() {
return globalLimiter == NoopQuotaLimiter.get();
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/RateLimiter.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/RateLimiter.java
index 9e258e70951..b83e45e9407 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/RateLimiter.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/RateLimiter.java
@@ -199,15 +199,13 @@ public abstract class RateLimiter {
}
}
- /**
- * @return estimate of the ms required to wait before being able to provide 1 resource.
- */
+ /** Returns estimate of the ms required to wait before being able to provide 1 resource. */
public long waitInterval() {
return waitInterval(1);
}
/**
- * @return estimate of the ms required to wait before being able to provide "amount" resources.
+ * Returns estimate of the ms required to wait before being able to provide "amount" resources.
*/
public synchronized long waitInterval(final long amount) {
// TODO Handle over quota?
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/RegionServerSpaceQuotaManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/RegionServerSpaceQuotaManager.java
index ea194da1430..0233e33312f 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/RegionServerSpaceQuotaManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/RegionServerSpaceQuotaManager.java
@@ -108,9 +108,7 @@ public class RegionServerSpaceQuotaManager {
started = false;
}
- /**
- * @return if the {@code Chore} has been started.
- */
+ /** Returns if the {@code Chore} has been started. */
public boolean isStarted() {
return started;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/UserQuotaState.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/UserQuotaState.java
index 4759672570e..6078c687d3d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/UserQuotaState.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/UserQuotaState.java
@@ -83,9 +83,7 @@ public class UserQuotaState extends QuotaState {
return builder.toString();
}
- /**
- * @return true if there is no quota information associated to this object
- */
+ /** Returns true if there is no quota information associated to this object */
@Override
public synchronized boolean isBypass() {
return !bypassGlobals && getGlobalLimiterWithoutUpdatingLastQuery() == NoopQuotaLimiter.get()
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMemStore.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMemStore.java
index af954474014..179c617462f 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMemStore.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMemStore.java
@@ -218,9 +218,7 @@ public abstract class AbstractMemStore implements MemStore {
}
}
- /**
- * @return Oldest timestamp of all the Cells in the MemStore
- */
+ /** Returns Oldest timestamp of all the Cells in the MemStore */
@Override
public long timeOfOldestEdit() {
return timeOfOldestEdit;
@@ -356,7 +354,7 @@ public abstract class AbstractMemStore implements MemStore {
}
/**
- * @return The total size of cells in this memstore. We will not consider cells in the snapshot
+ * Returns The total size of cells in this memstore. We will not consider cells in the snapshot
*/
protected abstract long keySize();
@@ -378,9 +376,7 @@ public abstract class AbstractMemStore implements MemStore {
return snapshot;
}
- /**
- * @return an ordered list of segments from most recent to oldest in memstore
- */
+ /** Returns an ordered list of segments from most recent to oldest in memstore */
protected abstract List<Segment> getSegments() throws IOException;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ChangedReadersObserver.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ChangedReadersObserver.java
index a7a3dc800ce..d45a8046873 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ChangedReadersObserver.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ChangedReadersObserver.java
@@ -27,9 +27,7 @@ import org.apache.yetus.audience.InterfaceAudience;
@InterfaceAudience.Private
public interface ChangedReadersObserver {
- /**
- * @return the read point of the current scan
- */
+ /** Returns the read point of the current scan */
long getReadPoint();
/**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Chunk.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Chunk.java
index eb6282bf163..5e37ce4ee72 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Chunk.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Chunk.java
@@ -171,9 +171,7 @@ public abstract class Chunk {
}
}
- /**
- * @return This chunk's backing data.
- */
+ /** Returns This chunk's backing data. */
ByteBuffer getData() {
return this.data;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplit.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplit.java
index 6719db72c8c..8a6e806292a 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplit.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplit.java
@@ -497,9 +497,7 @@ public class CompactSplit implements CompactionRequester, PropagatingConfigurati
return (regionSplitLimit > server.getNumberOfOnlineRegions());
}
- /**
- * @return the regionSplitLimit
- */
+ /** Returns the regionSplitLimit */
public int getRegionSplitLimit() {
return this.regionSplitLimit;
}
@@ -856,16 +854,12 @@ public class CompactSplit implements CompactionRequester, PropagatingConfigurati
this.conf.set(HBASE_REGION_SERVER_ENABLE_COMPACTION, String.valueOf(compactionsEnabled));
}
- /**
- * @return the longCompactions thread pool executor
- */
+ /** Returns the longCompactions thread pool executor */
ThreadPoolExecutor getLongCompactions() {
return longCompactions;
}
- /**
- * @return the shortCompactions thread pool executor
- */
+ /** Returns the shortCompactions thread pool executor */
ThreadPoolExecutor getShortCompactions() {
return shortCompactions;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompositeImmutableSegment.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompositeImmutableSegment.java
index 57cc0c6f21a..f955eb5d582 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompositeImmutableSegment.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompositeImmutableSegment.java
@@ -58,9 +58,7 @@ public class CompositeImmutableSegment extends ImmutableSegment {
return segments.size();
}
- /**
- * @return whether the segment has any cells
- */
+ /** Returns whether the segment has any cells */
@Override
public boolean isEmpty() {
for (ImmutableSegment s : segments) {
@@ -69,9 +67,7 @@ public class CompositeImmutableSegment extends ImmutableSegment {
return true;
}
- /**
- * @return number of cells in segment
- */
+ /** Returns number of cells in segment */
@Override
public int getCellsCount() {
int result = 0;
@@ -160,17 +156,13 @@ public class CompositeImmutableSegment extends ImmutableSegment {
return false;
}
- /**
- * @return Sum of all cell sizes.
- */
+ /** Returns Sum of all cell sizes. */
@Override
public long getDataSize() {
return this.keySize;
}
- /**
- * @return The heap size of this segment.
- */
+ /** Returns The heap size of this segment. */
@Override
public long getHeapSize() {
long result = 0;
@@ -224,9 +216,7 @@ public class CompositeImmutableSegment extends ImmutableSegment {
throw new IllegalStateException("Not supported by CompositeImmutableScanner");
}
- /**
- * @return a set of all cells in the segment
- */
+ /** Returns a set of all cells in the segment */
@Override
protected CellSet getCellSet() {
throw new IllegalStateException("Not supported by CompositeImmutableScanner");
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ConstantSizeRegionSplitPolicy.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ConstantSizeRegionSplitPolicy.java
index e0d7627a929..be8fbb1b320 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ConstantSizeRegionSplitPolicy.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ConstantSizeRegionSplitPolicy.java
@@ -90,9 +90,7 @@ public class ConstantSizeRegionSplitPolicy extends RegionSplitPolicy {
return this.jitterRate > 0;
}
- /**
- * @return true if region size exceed the sizeToCheck
- */
+ /** Returns true if region size exceed the sizeToCheck */
protected final boolean isExceedSize(long sizeToCheck) {
if (overallHRegionFiles) {
long sumSize = 0;
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/FlushNonSloppyStoresFirstPolicy.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/FlushNonSloppyStoresFirstPolicy.java
index 75277fdcab6..b130373ab74 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/FlushNonSloppyStoresFirstPolicy.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/FlushNonSloppyStoresFirstPolicy.java
@@ -33,9 +33,7 @@ public class FlushNonSloppyStoresFirstPolicy extends FlushLargeStoresPolicy {
private Collection<HStore> regularStores = new HashSet<>();
private Collection<HStore> sloppyStores = new HashSet<>();
- /**
- * @return the stores need to be flushed.
- */
+ /** Returns the stores need to be flushed. */
@Override
public Collection<HStore> selectStoresToFlush() {
Collection<HStore> specificStoresToFlush = new HashSet<>();
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/FlushPolicy.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/FlushPolicy.java
index fb9f8c8cb7a..401f79ce383 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/FlushPolicy.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/FlushPolicy.java
@@ -40,8 +40,6 @@ public abstract class FlushPolicy extends Configured {
this.region = region;
}
- /**
- * @return the stores need to be flushed.
- */
+ /** Returns the stores need to be flushed. */
public abstract Collection<HStore> selectStoresToFlush();
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
index 7ee26d2ec92..89117fc79a0 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
@@ -622,9 +622,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
private boolean failedSanityCheck = false;
private boolean wrongFamily = false;
- /**
- * @return If a {@link WrongRegionException} has been observed.
- */
+ /** Returns If a {@link WrongRegionException} has been observed. */
boolean hasSeenWrongRegion() {
return wrongRegion;
}
@@ -636,9 +634,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
wrongRegion = true;
}
- /**
- * @return If a {@link FailedSanityCheckException} has been observed.
- */
+ /** Returns If a {@link FailedSanityCheckException} has been observed. */
boolean hasSeenFailedSanityCheck() {
return failedSanityCheck;
}
@@ -650,9 +646,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
failedSanityCheck = true;
}
- /**
- * @return If a {@link NoSuchColumnFamilyException} has been observed.
- */
+ /** Returns If a {@link NoSuchColumnFamilyException} has been observed. */
boolean hasSeenNoSuchFamily() {
return wrongFamily;
}
@@ -1174,9 +1168,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
}
}
- /**
- * @return Map of StoreFiles by column family
- */
+ /** Returns Map of StoreFiles by column family */
private NavigableMap<byte[], List<Path>> getStoreFiles() {
NavigableMap<byte[], List<Path>> allStoreFiles = new TreeMap<>(Bytes.BYTES_COMPARATOR);
for (HStore store : stores.values()) {
@@ -1219,9 +1211,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
}
}
- /**
- * @return True if this region has references.
- */
+ /** Returns True if this region has references. */
public boolean hasReferences() {
return stores.values().stream().anyMatch(HStore::hasReferences);
}
@@ -1344,9 +1334,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
return this.fs.getRegionInfo();
}
- /**
- * @return Instance of {@link RegionServerServices} used by this HRegion. Can be null.
- */
+ /** Returns Instance of {@link RegionServerServices} used by this HRegion. Can be null. */
RegionServerServices getRegionServerServices() {
return this.rsServices;
}
@@ -1381,7 +1369,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
return memStoreSizing.getOffHeapSize();
}
- /** @return store services for this region, to access services required by store level needs */
+ /** Returns store services for this region, to access services required by store level needs */
public RegionServicesForStores getRegionServicesForStores() {
return regionServicesForStores;
}
@@ -1472,9 +1460,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
return maxFlushedSeqId;
}
- /**
- * @return readpoint considering given IsolationLevel. Pass {@code null} for default
- */
+ /** Returns readpoint considering given IsolationLevel. Pass {@code null} for default */
public long getReadPoint(IsolationLevel isolationLevel) {
if (isolationLevel != null && isolationLevel == IsolationLevel.READ_UNCOMMITTED) {
// This scan can read even uncommitted transactions
@@ -1955,9 +1941,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
});
}
- /**
- * @return True if its worth doing a flush before we put up the close flag.
- */
+ /** Returns True if its worth doing a flush before we put up the close flag. */
private boolean worthPreFlushing() {
return this.memStoreSizing.getDataSize()
> this.conf.getLong("hbase.hregion.preclose.flush.size", 1024 * 1024 * 5);
@@ -1978,7 +1962,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
htableDescriptor = desc;
}
- /** @return WAL in use for this region */
+ /** Returns WAL in use for this region */
public WAL getWAL() {
return this.wal;
}
@@ -2005,9 +1989,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
this.mobFileCache = mobFileCache;
}
- /**
- * @return split policy for this region.
- */
+ /** Returns split policy for this region. */
RegionSplitPolicy getSplitPolicy() {
return this.splitPolicy;
}
@@ -2023,23 +2005,23 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
return this.baseConf;
}
- /** @return {@link FileSystem} being used by this region */
+ /** Returns {@link FileSystem} being used by this region */
public FileSystem getFilesystem() {
return fs.getFileSystem();
}
- /** @return the {@link HRegionFileSystem} used by this region */
+ /** Returns the {@link HRegionFileSystem} used by this region */
public HRegionFileSystem getRegionFileSystem() {
return this.fs;
}
- /** @return the WAL {@link HRegionFileSystem} used by this region */
+ /** Returns the WAL {@link HRegionFileSystem} used by this region */
HRegionWALFileSystem getRegionWALFileSystem() throws IOException {
return new HRegionWALFileSystem(conf, getWalFileSystem(),
CommonFSUtils.getWALTableDir(conf, htableDescriptor.getTableName()), fs.getRegionInfo());
}
- /** @return the WAL {@link FileSystem} being used by this region */
+ /** Returns the WAL {@link FileSystem} being used by this region */
FileSystem getWalFileSystem() throws IOException {
if (walFS == null) {
walFS = CommonFSUtils.getWALFileSystem(conf);
@@ -2336,13 +2318,13 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
CANNOT_FLUSH
}
- /** @return the detailed result code */
+ /** Returns the detailed result code */
Result getResult();
- /** @return true if the memstores were flushed, else false */
+ /** Returns true if the memstores were flushed, else false */
boolean isFlushSucceeded();
- /** @return True if the flush requested a compaction, else false */
+ /** Returns True if the flush requested a compaction, else false */
boolean isCompactionNeeded();
}
@@ -2792,9 +2774,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
}
}
- /**
- * @return True if passed Set is all families in the region.
- */
+ /** Returns True if passed Set is all families in the region. */
private boolean isAllFamilies(Collection<HStore> families) {
return families == null || this.stores.size() == families.size();
}
@@ -4010,9 +3990,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
}
}
- /**
- * @return Get the long out of the passed in Cell
- */
+ /** Returns Get the long out of the passed in Cell */
private static long getLongValue(final Cell cell) throws DoNotRetryIOException {
int len = cell.getValueLength();
if (len != Bytes.SIZEOF_LONG) {
@@ -7486,9 +7464,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
});
}
- /**
- * @return statistics about the current load of the region
- */
+ /** Returns statistics about the current load of the region */
public ClientProtos.RegionLoadStats getLoadStatistics() {
if (!regionStatsEnabled) {
return null;
@@ -7779,9 +7755,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
SequenceId.NO_SEQUENCE_ID);
}
- /**
- * @return writeEntry associated with this append
- */
+ /** Returns writeEntry associated with this append */
private WriteEntry doWALAppend(WALEdit walEdit, Durability durability, List<UUID> clusterIds,
long now, long nonceGroup, long nonce, long origLogSeqNum) throws IOException {
Preconditions.checkArgument(walEdit != null && !walEdit.isEmpty(), "WALEdit is null or empty!");
@@ -8002,9 +7976,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
}
}
- /**
- * @return The priority that this region should have in the compaction queue
- */
+ /** Returns The priority that this region should have in the compaction queue */
public int getCompactPriority() {
if (conf.getBoolean(SPLIT_IGNORE_BLOCKING_ENABLED_KEY, false) && checkSplit().isPresent()) {
// if a region should split, split it before compact
@@ -8014,7 +7986,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
.orElse(Store.NO_PRIORITY);
}
- /** @return the coprocessor host */
+ /** Returns the coprocessor host */
public RegionCoprocessorHost getCoprocessorHost() {
return coprocessorHost;
}
@@ -8243,7 +8215,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
return regionDurability.ordinal() > Durability.ASYNC_WAL.ordinal();
}
- /** @return the latest sequence number that was read from storage when this region was opened */
+ /** Returns the latest sequence number that was read from storage when this region was opened */
public long getOpenSeqNum() {
return this.openSeqNum;
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionFileSystem.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionFileSystem.java
index 663e2b73ae2..2d2965d7237 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionFileSystem.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionFileSystem.java
@@ -124,12 +124,12 @@ public class HRegionFileSystem {
conf.getInt("hdfs.client.sleep.before.retries", DEFAULT_BASE_SLEEP_BEFORE_RETRIES);
}
- /** @return the underlying {@link FileSystem} */
+ /** Returns the underlying {@link FileSystem} */
public FileSystem getFileSystem() {
return this.fs;
}
- /** @return the {@link RegionInfo} that describe this on-disk region view */
+ /** Returns the {@link RegionInfo} that describe this on-disk region view */
public RegionInfo getRegionInfo() {
return this.regionInfo;
}
@@ -138,12 +138,12 @@ public class HRegionFileSystem {
return this.regionInfoForFs;
}
- /** @return {@link Path} to the region's root directory. */
+ /** Returns {@link Path} to the region's root directory. */
public Path getTableDir() {
return this.tableDir;
}
- /** @return {@link Path} to the region directory. */
+ /** Returns {@link Path} to the region directory. */
public Path getRegionDir() {
return regionDir;
}
@@ -151,7 +151,7 @@ public class HRegionFileSystem {
// ===========================================================================
// Temp Helpers
// ===========================================================================
- /** @return {@link Path} to the region's temp directory, used for file creations */
+ /** Returns {@link Path} to the region's temp directory, used for file creations */
public Path getTempDir() {
return new Path(getRegionDir(), REGION_TEMP_DIR);
}
@@ -379,9 +379,7 @@ public class HRegionFileSystem {
return false;
}
- /**
- * @return the set of families present on disk n
- */
+ /** Returns the set of families present on disk n */
public Collection<String> getFamilies() throws IOException {
FileStatus[] fds =
CommonFSUtils.listStatus(fs, getRegionDir(), new FSUtils.FamilyDirFilter(fs));
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java
index 4e74034e51b..5408685311e 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java
@@ -1020,9 +1020,7 @@ public class HRegionServer extends Thread
}
}
- /**
- * @return True if the cluster is up.
- */
+ /** Returns True if the cluster is up. */
@Override
public boolean isClusterUp() {
return this.masterless
@@ -1314,9 +1312,7 @@ public class HRegionServer extends Thread
return allUserRegionsOffline;
}
- /**
- * @return Current write count for all online regions.
- */
+ /** Returns Current write count for all online regions. */
private long getWriteRequestCount() {
long writeCount = 0;
for (Map.Entry<String, HRegion> e : this.onlineRegions.entrySet()) {
@@ -2066,9 +2062,7 @@ public class HRegionServer extends Thread
}
}
- /**
- * @return Master address tracker instance.
- */
+ /** Returns Master address tracker instance. */
public MasterAddressTracker getMasterAddressTracker() {
return this.masterAddressTracker;
}
@@ -2821,17 +2815,13 @@ public class HRegionServer extends Thread
}
}
- /**
- * @return Return the object that implements the replication source executorService.
- */
+ /** Returns Return the object that implements the replication source executorService. */
@InterfaceAudience.Private
public ReplicationSourceService getReplicationSourceService() {
return replicationSourceHandler;
}
... 3617 lines suppressed ...