You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by zh...@apache.org on 2022/07/19 02:10:35 UTC

[hbase] branch branch-2 updated: HBASE-27220 Apply the spotless format change in HBASE-27208 to our code base

This is an automated email from the ASF dual-hosted git repository.

zhangduo pushed a commit to branch branch-2
in repository https://gitbox.apache.org/repos/asf/hbase.git


The following commit(s) were added to refs/heads/branch-2 by this push:
     new 99f2ab5aa8f HBASE-27220 Apply the spotless format change in HBASE-27208 to our code base
99f2ab5aa8f is described below

commit 99f2ab5aa8f01c056d8f0902d0c67fae0e8b104e
Author: Duo Zhang <zh...@apache.org>
AuthorDate: Tue Jul 19 10:00:23 2022 +0800

    HBASE-27220 Apply the spotless format change in HBASE-27208 to our code base
    
    Signed-off-by: Andrew Purtell <ap...@apache.org>
---
 .../hadoop/hbase/io/asyncfs/AsyncFSOutput.java     |   4 +-
 .../hadoop/hbase/util/RecoverLeaseFSUtils.java     |   4 +-
 .../java/org/apache/hadoop/hbase/ClusterId.java    |   8 +-
 .../org/apache/hadoop/hbase/ClusterMetrics.java    |  28 ++---
 .../org/apache/hadoop/hbase/ClusterStatus.java     |   8 +-
 .../hadoop/hbase/CoprocessorEnvironment.java       |  14 ++-
 .../apache/hadoop/hbase/HBaseServerException.java  |   4 +-
 .../org/apache/hadoop/hbase/HColumnDescriptor.java |  12 +--
 .../java/org/apache/hadoop/hbase/HRegionInfo.java  |  48 +++------
 .../org/apache/hadoop/hbase/HTableDescriptor.java  |  12 +--
 .../org/apache/hadoop/hbase/MetaTableAccessor.java |  16 +--
 .../java/org/apache/hadoop/hbase/RegionLoad.java   |   8 +-
 .../org/apache/hadoop/hbase/RegionMetrics.java     |  92 ++++------------
 .../org/apache/hadoop/hbase/ServerMetrics.java     |  40 ++-----
 .../main/java/org/apache/hadoop/hbase/Size.java    |   4 +-
 .../java/org/apache/hadoop/hbase/UserMetrics.java  |  24 ++---
 .../java/org/apache/hadoop/hbase/client/Admin.java |   8 +-
 .../org/apache/hadoop/hbase/client/Append.java     |   4 +-
 .../org/apache/hadoop/hbase/client/AsyncAdmin.java |  24 ++---
 .../apache/hadoop/hbase/client/CheckAndMutate.java |  36 ++-----
 .../hadoop/hbase/client/CheckAndMutateResult.java  |   8 +-
 .../hadoop/hbase/client/ClientIdGenerator.java     |   4 +-
 .../hadoop/hbase/client/ClusterConnection.java     |  30 ++----
 .../hbase/client/ColumnFamilyDescriptor.java       |  96 +++++------------
 .../org/apache/hadoop/hbase/client/Connection.java |   8 +-
 .../hbase/client/ConnectionRegistryFactory.java    |   4 +-
 .../hadoop/hbase/client/CoprocessorDescriptor.java |  14 +--
 .../org/apache/hadoop/hbase/client/HBaseAdmin.java |  26 ++---
 .../org/apache/hadoop/hbase/client/HTable.java     |   4 +-
 .../hadoop/hbase/client/HTableMultiplexer.java     |   4 +-
 .../org/apache/hadoop/hbase/client/Increment.java  |   4 +-
 .../apache/hadoop/hbase/client/MultiResponse.java  |   8 +-
 .../hadoop/hbase/client/MutableRegionInfo.java     |  20 ++--
 .../org/apache/hadoop/hbase/client/Mutation.java   |  24 ++---
 .../apache/hadoop/hbase/client/NonceGenerator.java |   4 +-
 .../java/org/apache/hadoop/hbase/client/Query.java |  12 +--
 .../hbase/client/RegionAdminServiceCallable.java   |   4 +-
 .../org/apache/hadoop/hbase/client/RegionInfo.java |  66 ++++--------
 .../hadoop/hbase/client/RegionReplicaUtil.java     |   4 +-
 .../hadoop/hbase/client/RegionServerCallable.java  |   8 +-
 .../hadoop/hbase/client/RequestController.java     |   8 +-
 .../org/apache/hadoop/hbase/client/Result.java     |   4 +-
 .../apache/hadoop/hbase/client/ResultScanner.java  |   4 +-
 .../java/org/apache/hadoop/hbase/client/Row.java   |   4 +-
 .../org/apache/hadoop/hbase/client/RowAccess.java  |   8 +-
 .../apache/hadoop/hbase/client/RowMutations.java   |   4 +-
 .../hadoop/hbase/client/RpcRetryingCallerImpl.java |   4 +-
 .../java/org/apache/hadoop/hbase/client/Scan.java  |  68 +++---------
 .../hadoop/hbase/client/ScannerCallable.java       |   4 +-
 .../hadoop/hbase/client/TableDescriptor.java       |  12 +--
 .../hbase/client/TableDescriptorBuilder.java       |  20 +---
 .../org/apache/hadoop/hbase/client/TableState.java |  28 ++---
 .../hbase/client/backoff/ClientBackoffPolicy.java  |   4 +-
 .../exceptions/PreemptiveFastFailException.java    |  20 +---
 .../hadoop/hbase/filter/BigDecimalComparator.java  |   4 +-
 .../hadoop/hbase/filter/BinaryComparator.java      |   4 +-
 .../hbase/filter/BinaryComponentComparator.java    |   4 +-
 .../hbase/filter/BinaryPrefixComparator.java       |   4 +-
 .../apache/hadoop/hbase/filter/BitComparator.java  |   8 +-
 .../hadoop/hbase/filter/ColumnCountGetFilter.java  |   4 +-
 .../hbase/filter/ColumnPaginationFilter.java       |   4 +-
 .../hadoop/hbase/filter/ColumnPrefixFilter.java    |   4 +-
 .../hadoop/hbase/filter/ColumnRangeFilter.java     |  28 ++---
 .../hadoop/hbase/filter/ColumnValueFilter.java     |  16 +--
 .../apache/hadoop/hbase/filter/CompareFilter.java  |   8 +-
 .../hadoop/hbase/filter/DependentColumnFilter.java |  16 +--
 .../apache/hadoop/hbase/filter/FamilyFilter.java   |   4 +-
 .../org/apache/hadoop/hbase/filter/FilterList.java |   4 +-
 .../hadoop/hbase/filter/FirstKeyOnlyFilter.java    |   8 +-
 .../FirstKeyValueMatchingQualifiersFilter.java     |   4 +-
 .../apache/hadoop/hbase/filter/FuzzyRowFilter.java |   4 +-
 .../hadoop/hbase/filter/InclusiveStopFilter.java   |   4 +-
 .../apache/hadoop/hbase/filter/KeyOnlyFilter.java  |   4 +-
 .../apache/hadoop/hbase/filter/LongComparator.java |   4 +-
 .../hadoop/hbase/filter/MultiRowRangeFilter.java   |  12 +--
 .../hbase/filter/MultipleColumnPrefixFilter.java   |   4 +-
 .../apache/hadoop/hbase/filter/NullComparator.java |   4 +-
 .../org/apache/hadoop/hbase/filter/PageFilter.java |   4 +-
 .../apache/hadoop/hbase/filter/PrefixFilter.java   |   4 +-
 .../hadoop/hbase/filter/QualifierFilter.java       |   4 +-
 .../hadoop/hbase/filter/RandomRowFilter.java       |   8 +-
 .../hadoop/hbase/filter/RegexStringComparator.java |   4 +-
 .../org/apache/hadoop/hbase/filter/RowFilter.java  |   4 +-
 .../filter/SingleColumnValueExcludeFilter.java     |   4 +-
 .../hbase/filter/SingleColumnValueFilter.java      |  16 +--
 .../org/apache/hadoop/hbase/filter/SkipFilter.java |   4 +-
 .../hadoop/hbase/filter/SubstringComparator.java   |   4 +-
 .../hadoop/hbase/filter/TimestampsFilter.java      |   8 +-
 .../apache/hadoop/hbase/filter/ValueFilter.java    |   4 +-
 .../hadoop/hbase/filter/WhileMatchFilter.java      |   4 +-
 .../hadoop/hbase/ipc/HBaseRpcController.java       |  12 +--
 .../hadoop/hbase/ipc/HBaseRpcControllerImpl.java   |   4 +-
 .../java/org/apache/hadoop/hbase/ipc/IPCUtil.java  |   8 +-
 .../hbase/ipc/RemoteWithExtrasException.java       |  16 +--
 .../hadoop/hbase/protobuf/ProtobufMagic.java       |   4 +-
 .../apache/hadoop/hbase/protobuf/ProtobufUtil.java |   4 +-
 .../apache/hadoop/hbase/quotas/QuotaFilter.java    |  12 +--
 .../hadoop/hbase/quotas/SpaceQuotaSnapshot.java    |   4 +-
 .../hbase/quotas/SpaceQuotaSnapshotView.java       |   4 +-
 .../hadoop/hbase/security/access/Permission.java   |   2 +-
 .../hbase/security/visibility/CellVisibility.java  |   4 +-
 .../hadoop/hbase/shaded/protobuf/ProtobufUtil.java |   4 +-
 .../hbase/shaded/protobuf/RequestConverter.java    |   4 +-
 .../hbase/snapshot/HBaseSnapshotException.java     |   4 +-
 .../apache/hadoop/hbase/zookeeper/ZNodePaths.java  |  12 +--
 .../org/apache/hadoop/hbase/ArrayBackedTag.java    |  16 +--
 .../hadoop/hbase/ByteBufferExtendedCell.java       |  40 ++-----
 .../main/java/org/apache/hadoop/hbase/Cell.java    |  36 ++-----
 .../org/apache/hadoop/hbase/CellScannable.java     |   4 +-
 .../java/org/apache/hadoop/hbase/CellScanner.java  |   4 +-
 .../java/org/apache/hadoop/hbase/CellUtil.java     |   8 +-
 .../java/org/apache/hadoop/hbase/ChoreService.java |  20 +---
 .../java/org/apache/hadoop/hbase/ExtendedCell.java |  12 +--
 .../apache/hadoop/hbase/HBaseConfiguration.java    |   4 +-
 .../java/org/apache/hadoop/hbase/KeyValue.java     |  94 +++++------------
 .../org/apache/hadoop/hbase/PrivateCellUtil.java   |  16 +--
 .../main/java/org/apache/hadoop/hbase/RawCell.java |   4 +-
 .../apache/hadoop/hbase/RawCellBuilderFactory.java |   4 +-
 .../org/apache/hadoop/hbase/ScheduledChore.java    |  24 ++---
 .../java/org/apache/hadoop/hbase/ServerName.java   |   4 +-
 .../java/org/apache/hadoop/hbase/Stoppable.java    |   4 +-
 .../java/org/apache/hadoop/hbase/TableName.java    |   8 +-
 .../src/main/java/org/apache/hadoop/hbase/Tag.java |  16 +--
 .../main/java/org/apache/hadoop/hbase/TagUtil.java |   8 +-
 .../org/apache/hadoop/hbase/codec/CellCodec.java   |   4 +-
 .../hadoop/hbase/codec/CellCodecWithTags.java      |   4 +-
 .../hadoop/hbase/conf/ConfigurationManager.java    |   4 +-
 .../hadoop/hbase/filter/ByteArrayComparable.java   |   4 +-
 .../hadoop/hbase/io/ByteArrayOutputStream.java     |   8 +-
 .../hadoop/hbase/io/ImmutableBytesWritable.java    |   4 +-
 .../java/org/apache/hadoop/hbase/io/TimeRange.java |   8 +-
 .../io/encoding/BufferedDataBlockEncoder.java      |   4 +-
 .../hadoop/hbase/io/encoding/DataBlockEncoder.java |   4 +-
 .../hbase/io/encoding/DataBlockEncoding.java       |   8 +-
 .../hadoop/hbase/io/encoding/EncodedDataBlock.java |   2 +-
 .../io/encoding/HFileBlockDecodingContext.java     |   4 +-
 .../io/encoding/HFileBlockEncodingContext.java     |  16 +--
 .../apache/hadoop/hbase/io/hfile/BlockType.java    |   4 +-
 .../apache/hadoop/hbase/io/hfile/HFileContext.java |   4 +-
 .../java/org/apache/hadoop/hbase/nio/ByteBuff.java |  20 +---
 .../org/apache/hadoop/hbase/nio/MultiByteBuff.java |  12 +--
 .../org/apache/hadoop/hbase/security/User.java     |   4 +-
 .../apache/hadoop/hbase/security/UserProvider.java |   4 +-
 .../hadoop/hbase/util/AbstractByteRange.java       |   4 +-
 .../org/apache/hadoop/hbase/util/ByteRange.java    |   4 +-
 .../java/org/apache/hadoop/hbase/util/Bytes.java   |   4 +-
 .../apache/hadoop/hbase/util/ExceptionUtil.java    |   8 +-
 .../java/org/apache/hadoop/hbase/util/HashKey.java |   4 +-
 .../java/org/apache/hadoop/hbase/util/Sleeper.java |   4 +-
 .../hadoop/hbase/util/WindowMovingAverage.java     |   8 +-
 .../hadoop/hbase/HBaseCommonTestingUtility.java    |   8 +-
 .../apache/hadoop/hbase/net/BoundSocketMaker.java  |   4 +-
 .../master/MetricsAssignmentManagerSource.java     |  32 ++----
 .../hadoop/hbase/master/MetricsMasterSource.java   |   4 +-
 .../regionserver/MetricsRegionServerWrapper.java   |  46 +++-----
 .../hbase/regionserver/MetricsRegionWrapper.java   |  32 ++----
 .../regionserver/MetricsTableWrapperAggregate.java |  24 ++---
 .../metrics2/lib/DynamicMetricsRegistry.java       |   4 +-
 .../org/apache/hadoop/hbase/util/JSONBean.java     |   4 +-
 .../hadoop/hbase/IntegrationTestMobCompaction.java |   4 +-
 .../org/apache/hadoop/hbase/mapred/TableSplit.java |  10 +-
 .../apache/hadoop/hbase/mapreduce/CellCreator.java |   4 +-
 .../apache/hadoop/hbase/PerformanceEvaluation.java |   8 +-
 .../mapreduce/TestRoundRobinTableInputFormat.java  |   4 +-
 .../hadoop/hbase/procedure2/LockAndQueue.java      |  12 +--
 .../apache/hadoop/hbase/procedure2/Procedure.java  |  28 ++---
 .../hadoop/hbase/procedure2/ProcedureEvent.java    |   2 +-
 .../hadoop/hbase/procedure2/ProcedureExecutor.java |  12 +--
 .../hadoop/hbase/procedure2/ProcedureMetrics.java  |  10 +-
 .../hbase/procedure2/ProcedureScheduler.java       |   4 +-
 .../hbase/procedure2/store/ProcedureStore.java     |   8 +-
 .../hbase/procedure2/store/wal/BitSetNode.java     |  20 +---
 .../store/wal/ProcedureStoreTracker.java           |   4 +-
 .../hadoop/hbase/procedure2/util/DelayedUtil.java  |   8 +-
 .../hadoop/hbase/rest/ProtobufMessageHandler.java  |   4 +-
 .../org/apache/hadoop/hbase/rest/RESTServlet.java  |   8 +-
 .../apache/hadoop/hbase/rest/TableResource.java    |   6 +-
 .../apache/hadoop/hbase/rest/client/Client.java    |   8 +-
 .../apache/hadoop/hbase/rest/client/Cluster.java   |   4 +-
 .../apache/hadoop/hbase/rest/client/Response.java  |  20 +---
 .../apache/hadoop/hbase/rest/model/CellModel.java  |  16 +--
 .../hadoop/hbase/rest/model/CellSetModel.java      |   4 +-
 .../hadoop/hbase/rest/model/ColumnSchemaModel.java |  36 ++-----
 .../hbase/rest/model/NamespacesInstanceModel.java  |   4 +-
 .../hadoop/hbase/rest/model/NamespacesModel.java   |   4 +-
 .../apache/hadoop/hbase/rest/model/RowModel.java   |   8 +-
 .../hadoop/hbase/rest/model/ScannerModel.java      |  48 +++------
 .../rest/model/StorageClusterStatusModel.java      |  92 ++++------------
 .../rest/model/StorageClusterVersionModel.java     |   4 +-
 .../hadoop/hbase/rest/model/TableInfoModel.java    |   8 +-
 .../hadoop/hbase/rest/model/TableListModel.java    |   4 +-
 .../apache/hadoop/hbase/rest/model/TableModel.java |   4 +-
 .../hadoop/hbase/rest/model/TableRegionModel.java  |  20 +---
 .../hadoop/hbase/rest/model/TableSchemaModel.java  |  28 ++---
 .../hadoop/hbase/rest/model/VersionModel.java      |  20 +---
 .../hadoop/hbase/rsgroup/RSGroupAdminServer.java   |   4 +-
 .../hadoop/hbase/HDFSBlocksDistribution.java       |  20 +---
 .../org/apache/hadoop/hbase/LocalHBaseCluster.java |  16 +--
 .../hadoop/hbase/MetaRegionLocationCache.java      |   4 +-
 .../main/java/org/apache/hadoop/hbase/Server.java  |  16 +--
 .../org/apache/hadoop/hbase/TableDescriptors.java  |   8 +-
 .../apache/hadoop/hbase/backup/HFileArchiver.java  |  16 +--
 .../backup/example/TableHFileArchiveTracker.java   |   4 +-
 .../hadoop/hbase/client/VersionInfoUtil.java       |   4 +-
 .../hadoop/hbase/client/locking/EntityLock.java    |   4 +-
 .../coordination/SplitLogManagerCoordination.java  |  16 +--
 .../coordination/SplitLogWorkerCoordination.java   |   8 +-
 .../coordination/ZkSplitLogWorkerCoordination.java |   4 +-
 .../hadoop/hbase/coprocessor/BaseEnvironment.java  |   4 +-
 .../hbase/coprocessor/HasMasterServices.java       |   4 +-
 .../hbase/coprocessor/HasRegionServerServices.java |   4 +-
 .../coprocessor/MasterCoprocessorEnvironment.java  |   4 +-
 .../coprocessor/RegionCoprocessorEnvironment.java  |  14 +--
 .../RegionServerCoprocessorEnvironment.java        |   8 +-
 .../coprocessor/WALCoprocessorEnvironment.java     |   2 +-
 .../apache/hadoop/hbase/executor/EventHandler.java |   4 +-
 .../hadoop/hbase/executor/ExecutorService.java     |   2 +-
 .../apache/hadoop/hbase/executor/ExecutorType.java |   4 +-
 .../hadoop/hbase/favored/FavoredNodesPlan.java     |   4 +-
 .../apache/hadoop/hbase/filter/FilterWrapper.java  |   4 +-
 .../hadoop/hbase/io/FSDataInputStreamWrapper.java  |   4 +-
 .../java/org/apache/hadoop/hbase/io/FileLink.java  |  12 +--
 .../java/org/apache/hadoop/hbase/io/HFileLink.java |  12 +--
 .../apache/hadoop/hbase/io/hfile/BlockCache.java   |   8 +-
 .../hadoop/hbase/io/hfile/BlockCacheKey.java       |   4 +-
 .../hadoop/hbase/io/hfile/BlockCacheUtil.java      |  24 ++---
 .../apache/hadoop/hbase/io/hfile/CacheConfig.java  |  22 ++--
 .../apache/hadoop/hbase/io/hfile/Cacheable.java    |   4 +-
 .../org/apache/hadoop/hbase/io/hfile/HFile.java    |   2 +-
 .../apache/hadoop/hbase/io/hfile/HFileBlock.java   |  28 ++---
 .../hadoop/hbase/io/hfile/HFileBlockIndex.java     |  28 ++---
 .../hbase/io/hfile/HFileDataBlockEncoder.java      |   2 +-
 .../hadoop/hbase/io/hfile/HFileReaderImpl.java     |  12 +--
 .../apache/hadoop/hbase/io/hfile/HFileScanner.java |  12 +--
 .../hadoop/hbase/io/hfile/HFileWriterImpl.java     |   4 +-
 .../hadoop/hbase/io/hfile/InlineBlockWriter.java   |   4 +-
 .../hadoop/hbase/io/hfile/LruCachedBlock.java      |   4 +-
 .../hadoop/hbase/io/hfile/LruCachedBlockQueue.java |   8 +-
 .../hbase/io/hfile/bucket/CachedEntryQueue.java    |   8 +-
 .../hadoop/hbase/io/hfile/bucket/IOEngine.java     |   4 +-
 .../hadoop/hbase/io/util/MemorySizeUtil.java       |   4 +-
 .../ipc/FastPathBalancedQueueRpcExecutor.java      |   4 +-
 .../org/apache/hadoop/hbase/ipc/QueueBalancer.java |   4 +-
 .../java/org/apache/hadoop/hbase/ipc/RpcCall.java  |  44 ++------
 .../apache/hadoop/hbase/ipc/RpcCallContext.java    |  12 +--
 .../org/apache/hadoop/hbase/ipc/RpcServer.java     |   4 +-
 .../hadoop/hbase/master/ActiveMasterManager.java   |   8 +-
 .../org/apache/hadoop/hbase/master/HMaster.java    |  48 +++------
 .../hadoop/hbase/master/MasterFileSystem.java      |  20 +---
 .../hadoop/hbase/master/MasterRpcServices.java     |   4 +-
 .../apache/hadoop/hbase/master/MasterServices.java | 104 +++++-------------
 .../hadoop/hbase/master/MasterWalManager.java      |   6 +-
 .../hbase/master/MetricsAssignmentManager.java     |  32 ++----
 .../apache/hadoop/hbase/master/MetricsMaster.java  |   4 +-
 .../hbase/master/RegionPlacementMaintainer.java    |   4 +-
 .../apache/hadoop/hbase/master/ServerManager.java  |  18 +---
 .../hadoop/hbase/master/SnapshotSentinel.java      |   8 +-
 .../hbase/master/assignment/AssignmentManager.java |   4 +-
 .../assignment/MergeTableRegionsProcedure.java     |   4 +-
 .../hbase/master/assignment/RegionStates.java      |  20 ++--
 .../hbase/master/assignment/UnassignProcedure.java |   2 +-
 .../master/balancer/FavoredStochasticBalancer.java |   4 +-
 .../apache/hadoop/hbase/master/hbck/HbckChore.java |   2 +-
 .../hbase/master/janitor/CatalogJanitor.java       |   4 +-
 .../hbase/master/janitor/CatalogJanitorReport.java |   4 +-
 .../hbase/master/janitor/ReportMakingVisitor.java  |   8 +-
 .../hbase/master/normalizer/NormalizationPlan.java |   4 +-
 .../master/normalizer/SimpleRegionNormalizer.java  |   4 +-
 .../AbstractStateMachineRegionProcedure.java       |   4 +-
 .../master/procedure/EnableTableProcedure.java     |   4 +-
 .../master/procedure/ServerCrashProcedure.java     |   4 +-
 .../master/procedure/ServerProcedureInterface.java |   8 +-
 .../master/procedure/TableProcedureInterface.java  |   4 +-
 .../hbase/master/snapshot/SnapshotFileCache.java   |   4 +-
 .../hbase/master/snapshot/SnapshotManager.java     |   4 +-
 .../hbase/master/zksyncer/ClientZKSyncer.java      |   4 +-
 .../apache/hadoop/hbase/procedure/Procedure.java   |   2 +-
 .../hbase/procedure/ProcedureCoordinator.java      |   8 +-
 .../hadoop/hbase/quotas/MasterQuotaManager.java    |   4 +-
 .../apache/hadoop/hbase/quotas/OperationQuota.java |   2 +-
 .../apache/hadoop/hbase/quotas/QuotaLimiter.java   |   6 +-
 .../org/apache/hadoop/hbase/quotas/QuotaState.java |   4 +-
 .../apache/hadoop/hbase/quotas/RateLimiter.java    |   6 +-
 .../quotas/RegionServerSpaceQuotaManager.java      |   4 +-
 .../apache/hadoop/hbase/quotas/UserQuotaState.java |   4 +-
 .../hbase/regionserver/AbstractMemStore.java       |  10 +-
 .../hbase/regionserver/ChangedReadersObserver.java |   4 +-
 .../apache/hadoop/hbase/regionserver/Chunk.java    |   4 +-
 .../hadoop/hbase/regionserver/CompactSplit.java    |  12 +--
 .../regionserver/CompositeImmutableSegment.java    |  20 +---
 .../ConstantSizeRegionSplitPolicy.java             |   4 +-
 .../FlushNonSloppyStoresFirstPolicy.java           |   4 +-
 .../hadoop/hbase/regionserver/FlushPolicy.java     |   4 +-
 .../apache/hadoop/hbase/regionserver/HRegion.java  |  78 +++++---------
 .../hbase/regionserver/HRegionFileSystem.java      |  14 ++-
 .../hadoop/hbase/regionserver/HRegionServer.java   |  60 +++--------
 .../apache/hadoop/hbase/regionserver/HStore.java   |  20 +---
 .../hadoop/hbase/regionserver/HStoreFile.java      |   4 +-
 .../hbase/regionserver/HeapMemoryManager.java      |   4 +-
 .../IncreasingToUpperBoundRegionSplitPolicy.java   |   4 +-
 .../hadoop/hbase/regionserver/KeyValueHeap.java    |   4 +-
 .../hadoop/hbase/regionserver/KeyValueScanner.java |   4 +-
 .../hadoop/hbase/regionserver/LeaseManager.java    |   6 +-
 .../apache/hadoop/hbase/regionserver/MemStore.java |   4 +-
 .../hadoop/hbase/regionserver/MemStoreFlusher.java |  16 +--
 .../hadoop/hbase/regionserver/MemStoreSizing.java  |   8 +-
 .../hbase/regionserver/MemStoreSnapshot.java       |  16 +--
 .../hbase/regionserver/MetricsUserAggregate.java   |   4 +-
 .../regionserver/MiniBatchOperationInProgress.java |   4 +-
 .../hadoop/hbase/regionserver/RSRpcServices.java   |  16 +--
 .../apache/hadoop/hbase/regionserver/Region.java   |  40 ++++---
 .../hbase/regionserver/RegionCoprocessorHost.java  |   2 +-
 .../hadoop/hbase/regionserver/RegionScanner.java   |   8 +-
 .../hbase/regionserver/RegionScannerImpl.java      |  12 +--
 .../hbase/regionserver/RegionServerAccounting.java |  16 +--
 .../hbase/regionserver/RegionServerServices.java   |  68 +++---------
 .../hbase/regionserver/RegionSplitPolicy.java      |   8 +-
 .../hadoop/hbase/regionserver/RowProcessor.java    |   8 +-
 .../hadoop/hbase/regionserver/ScannerContext.java  |  12 +--
 .../apache/hadoop/hbase/regionserver/Segment.java  |   8 +-
 .../hadoop/hbase/regionserver/SegmentScanner.java  |   4 +-
 .../hadoop/hbase/regionserver/SplitLogWorker.java  |   4 +-
 .../apache/hadoop/hbase/regionserver/Store.java    | 116 ++++++---------------
 .../hbase/regionserver/StoreConfigInformation.java |   8 +-
 .../hadoop/hbase/regionserver/StoreEngine.java     |  16 +--
 .../hadoop/hbase/regionserver/StoreFile.java       |  36 ++-----
 .../hadoop/hbase/regionserver/StoreFileInfo.java   |  18 ++--
 .../hbase/regionserver/StoreFileManager.java       |   4 +-
 .../hbase/regionserver/StoreFlushContext.java      |   4 +-
 .../hadoop/hbase/regionserver/StoreScanner.java    |   8 +-
 .../hbase/regionserver/StripeStoreFileManager.java |   4 +-
 .../hbase/regionserver/TimeRangeTracker.java       |  12 +--
 .../compactions/CompactionConfiguration.java       |  28 ++---
 .../regionserver/compactions/CompactionPolicy.java |   4 +-
 .../compactions/CompactionProgress.java            |  12 +--
 .../compactions/CompactionRequest.java             |  28 ++---
 .../compactions/DateTieredCompactionPolicy.java    |   2 +-
 .../regionserver/compactions/OffPeakHours.java     |   8 +-
 .../compactions/StripeCompactionPolicy.java        |  16 +--
 .../regionserver/handler/OpenRegionHandler.java    |   8 +-
 .../regionserver/querymatcher/ColumnCount.java     |  12 +--
 .../regionserver/querymatcher/ColumnTracker.java   |   4 +-
 .../regionserver/querymatcher/DeleteTracker.java   |   4 +-
 .../querymatcher/ScanQueryMatcher.java             |  16 +--
 .../hbase/regionserver/wal/AbstractFSWAL.java      |  14 +--
 .../hadoop/hbase/regionserver/wal/FSHLog.java      |  16 +--
 .../hadoop/hbase/regionserver/wal/FSWALEntry.java  |   8 +-
 .../hadoop/hbase/regionserver/wal/ReaderBase.java  |  16 +--
 .../hbase/regionserver/wal/RingBufferTruck.java    |   4 +-
 .../hbase/replication/ReplicationEndpoint.java     |   4 +-
 .../regionserver/ReplicationSource.java            |   4 +-
 .../regionserver/ReplicationSourceInterface.java   |  32 ++----
 .../regionserver/ReplicationSourceWALReader.java   |   4 +-
 .../replication/regionserver/WALEntryBatch.java    |  36 ++-----
 .../replication/regionserver/WALEntryStream.java   |  12 +--
 .../hbase/security/access/AccessControlFilter.java |   4 +-
 .../hadoop/hbase/security/access/AuthResult.java   |   4 +-
 .../hbase/security/token/FsDelegationToken.java    |   8 +-
 .../security/visibility/VisibilityLabelsCache.java |   4 +-
 .../hbase/snapshot/RestoreSnapshotHelper.java      |  20 +---
 .../apache/hadoop/hbase/snapshot/SnapshotInfo.java |  40 +++----
 .../org/apache/hadoop/hbase/tool/CanaryTool.java   |   4 +-
 .../apache/hadoop/hbase/util/BloomFilterBase.java  |  12 +--
 .../hadoop/hbase/util/BloomFilterFactory.java      |  20 ++--
 .../apache/hadoop/hbase/util/ConnectionCache.java  |   4 +-
 .../hadoop/hbase/util/DirectMemoryUtils.java       |   8 +-
 .../java/org/apache/hadoop/hbase/util/FSUtils.java |   8 +-
 .../org/apache/hadoop/hbase/util/HBaseFsck.java    |  16 +--
 .../apache/hadoop/hbase/util/HbckTableInfo.java    |   4 +-
 .../apache/hadoop/hbase/util/JVMClusterUtil.java   |   4 +-
 .../apache/hadoop/hbase/util/RegionSplitter.java   |   4 +-
 .../hadoop/hbase/util/RollingStatCalculator.java   |   8 +-
 .../hadoop/hbase/util/ServerRegionReplicaUtil.java |  16 +--
 .../hadoop/hbase/util/YammerHistogramUtils.java    |   6 +-
 .../hbase/util/hbck/HFileCorruptionChecker.java    |  32 ++----
 .../hadoop/hbase/wal/AbstractFSWALProvider.java    |   4 +-
 .../wal/AbstractRecoveredEditsOutputSink.java      |   4 +-
 .../apache/hadoop/hbase/wal/AbstractWALRoller.java |   4 +-
 .../org/apache/hadoop/hbase/wal/EntryBuffers.java  |   4 +-
 .../org/apache/hadoop/hbase/wal/OutputSink.java    |  12 +--
 .../main/java/org/apache/hadoop/hbase/wal/WAL.java |   4 +-
 .../java/org/apache/hadoop/hbase/wal/WALEdit.java  |   4 +-
 .../java/org/apache/hadoop/hbase/wal/WALKey.java   |  20 +---
 .../org/apache/hadoop/hbase/wal/WALKeyImpl.java    |  16 ++-
 .../org/apache/hadoop/hbase/wal/WALProvider.java   |   4 +-
 .../org/apache/hadoop/hbase/MiniClusterRule.java   |   4 +-
 .../org/apache/hadoop/hbase/MiniHBaseCluster.java  |  16 +--
 .../hadoop/hbase/StartMiniClusterOption.java       |   4 +-
 .../hbase/io/hfile/TestHFileDataBlockEncoder.java  |   4 +-
 .../assignment/TestRaceBetweenSCPAndDTP.java       |   4 +-
 .../hadoop/hbase/master/procedure/TestHBCKSCP.java |   4 +-
 .../hadoop/hbase/regionserver/TestHRegion.java     |   4 +-
 .../hbase/regionserver/TestRecoveredEdits.java     |   4 +-
 .../regionserver/wal/TestLogRollingNoCluster.java  |   4 +-
 .../TestMetaRegionReplicaReplicationEndpoint.java  |   8 +-
 .../regionserver/TestReplicationSourceManager.java |   4 +-
 .../hadoop/hbase/util/HFileArchiveTestingUtil.java |   4 +-
 .../hadoop/hbase/util/MultiThreadedAction.java     |   2 +-
 .../hbase/util/test/LoadTestDataGenerator.java     |   4 +-
 .../hadoop/hbase/client/AbstractTestShell.java     |   8 +-
 .../org/apache/hadoop/hbase/thrift/ImplType.java   |   4 +-
 .../hbase/thrift/ThriftHBaseServiceHandler.java    |   2 +-
 .../hadoop/hbase/thrift/TestThriftServer.java      |   4 +-
 .../hadoop/hbase/thrift/ThriftServerRunner.java    |   4 +-
 .../hbase/zookeeper/ClusterStatusTracker.java      |   4 +-
 .../hbase/zookeeper/MiniZooKeeperCluster.java      |   4 +-
 .../apache/hadoop/hbase/zookeeper/ZKListener.java  |   4 +-
 .../org/apache/hadoop/hbase/zookeeper/ZKUtil.java  |  20 +---
 .../apache/hadoop/hbase/zookeeper/ZKWatcher.java   |   4 +-
 .../apache/hadoop/hbase/HBaseZKTestingUtility.java |   4 +-
 410 files changed, 1260 insertions(+), 3464 deletions(-)

diff --git a/hbase-asyncfs/src/main/java/org/apache/hadoop/hbase/io/asyncfs/AsyncFSOutput.java b/hbase-asyncfs/src/main/java/org/apache/hadoop/hbase/io/asyncfs/AsyncFSOutput.java
index b88b32bdb81..cd186f904a9 100644
--- a/hbase-asyncfs/src/main/java/org/apache/hadoop/hbase/io/asyncfs/AsyncFSOutput.java
+++ b/hbase-asyncfs/src/main/java/org/apache/hadoop/hbase/io/asyncfs/AsyncFSOutput.java
@@ -89,8 +89,6 @@ public interface AsyncFSOutput extends Closeable {
   @Override
   void close() throws IOException;
 
-  /**
-   * @return byteSize success synced to underlying filesystem.
-   */
+  /** Returns byteSize success synced to underlying filesystem. */
   long getSyncedLength();
 }
diff --git a/hbase-asyncfs/src/main/java/org/apache/hadoop/hbase/util/RecoverLeaseFSUtils.java b/hbase-asyncfs/src/main/java/org/apache/hadoop/hbase/util/RecoverLeaseFSUtils.java
index e4a410aa9c3..0014185b85c 100644
--- a/hbase-asyncfs/src/main/java/org/apache/hadoop/hbase/util/RecoverLeaseFSUtils.java
+++ b/hbase-asyncfs/src/main/java/org/apache/hadoop/hbase/util/RecoverLeaseFSUtils.java
@@ -186,9 +186,7 @@ public final class RecoverLeaseFSUtils {
     return recovered;
   }
 
-  /**
-   * @return Detail to append to any log message around lease recovering.
-   */
+  /** Returns Detail to append to any log message around lease recovering. */
   private static String getLogMessageDetail(final int nbAttempt, final Path p,
     final long startWaiting) {
     return "attempt=" + nbAttempt + " on file=" + p + " after "
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterId.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterId.java
index e769e80847f..9b7a5de19bd 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterId.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterId.java
@@ -45,9 +45,7 @@ public class ClusterId {
     this.id = uuid;
   }
 
-  /**
-   * @return The clusterid serialized using pb w/ pb magic prefix
-   */
+  /** Returns The clusterid serialized using pb w/ pb magic prefix */
   public byte[] toByteArray() {
     return ProtobufUtil.prependPBMagic(convert().toByteArray());
   }
@@ -74,9 +72,7 @@ public class ClusterId {
     }
   }
 
-  /**
-   * @return A pb instance to represent this instance.
-   */
+  /** Returns A pb instance to represent this instance. */
   public ClusterIdProtos.ClusterId convert() {
     ClusterIdProtos.ClusterId.Builder builder = ClusterIdProtos.ClusterId.newBuilder();
     return builder.setClusterId(this.id).build();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterMetrics.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterMetrics.java
index 769d48496af..8be97fa1fda 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterMetrics.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterMetrics.java
@@ -69,33 +69,23 @@ import org.apache.yetus.audience.InterfaceAudience;
 @InterfaceAudience.Public
 public interface ClusterMetrics {
 
-  /**
-   * @return the HBase version string as reported by the HMaster
-   */
+  /** Returns the HBase version string as reported by the HMaster */
   @Nullable
   String getHBaseVersion();
 
-  /**
-   * @return the names of region servers on the dead list
-   */
+  /** Returns the names of region servers on the dead list */
   List<ServerName> getDeadServerNames();
 
-  /**
-   * @return the names of region servers on the live list
-   */
+  /** Returns the names of region servers on the live list */
   Map<ServerName, ServerMetrics> getLiveServerMetrics();
 
-  /**
-   * @return the number of regions deployed on the cluster
-   */
+  /** Returns the number of regions deployed on the cluster */
   default int getRegionCount() {
     return getLiveServerMetrics().entrySet().stream()
       .mapToInt(v -> v.getValue().getRegionMetrics().size()).sum();
   }
 
-  /**
-   * @return the number of requests since last report
-   */
+  /** Returns the number of requests since last report */
   default long getRequestCount() {
     return getLiveServerMetrics().entrySet().stream()
       .flatMap(v -> v.getValue().getRegionMetrics().values().stream())
@@ -109,9 +99,7 @@ public interface ClusterMetrics {
   @Nullable
   ServerName getMasterName();
 
-  /**
-   * @return the names of backup masters
-   */
+  /** Returns the names of backup masters */
   List<ServerName> getBackupMasterNames();
 
   @InterfaceAudience.Private
@@ -142,9 +130,7 @@ public interface ClusterMetrics {
 
   List<ServerName> getServersName();
 
-  /**
-   * @return the average cluster load
-   */
+  /** Returns the average cluster load */
   default double getAverageLoad() {
     int serverSize = getLiveServerMetrics().size();
     if (serverSize == 0) {
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterStatus.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterStatus.java
index b202f0ec43e..e218437694b 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterStatus.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterStatus.java
@@ -107,9 +107,7 @@ public class ClusterStatus implements ClusterMetrics {
     this.metrics = metrics;
   }
 
-  /**
-   * @return the names of region servers on the dead list
-   */
+  /** Returns the names of region servers on the dead list */
   @Override
   public List<ServerName> getDeadServerNames() {
     return metrics.getDeadServerNames();
@@ -187,9 +185,7 @@ public class ClusterStatus implements ClusterMetrics {
     return metrics.getRegionStatesInTransition();
   }
 
-  /**
-   * @return the HBase version string as reported by the HMaster
-   */
+  /** Returns the HBase version string as reported by the HMaster */
   public String getHBaseVersion() {
     return metrics.getHBaseVersion();
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/CoprocessorEnvironment.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/CoprocessorEnvironment.java
index edbc5f479d6..94909f1c14e 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/CoprocessorEnvironment.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/CoprocessorEnvironment.java
@@ -28,19 +28,19 @@ import org.apache.yetus.audience.InterfaceStability;
 @InterfaceStability.Evolving
 public interface CoprocessorEnvironment<C extends Coprocessor> {
 
-  /** @return the Coprocessor interface version */
+  /** Returns the Coprocessor interface version */
   int getVersion();
 
-  /** @return the HBase version as a string (e.g. "0.21.0") */
+  /** Returns the HBase version as a string (e.g. "0.21.0") */
   String getHBaseVersion();
 
-  /** @return the loaded coprocessor instance */
+  /** Returns the loaded coprocessor instance */
   C getInstance();
 
-  /** @return the priority assigned to the loaded coprocessor */
+  /** Returns the priority assigned to the loaded coprocessor */
   int getPriority();
 
-  /** @return the load sequence number */
+  /** Returns the load sequence number */
   int getLoadSequence();
 
   /**
@@ -49,8 +49,6 @@ public interface CoprocessorEnvironment<C extends Coprocessor> {
    */
   Configuration getConfiguration();
 
-  /**
-   * @return the classloader for the loaded coprocessor instance
-   */
+  /** Returns the classloader for the loaded coprocessor instance */
   ClassLoader getClassLoader();
 }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HBaseServerException.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HBaseServerException.java
index 2e4ebbd0baa..3484995c1bf 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HBaseServerException.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HBaseServerException.java
@@ -63,9 +63,7 @@ public class HBaseServerException extends HBaseIOException {
     this.serverOverloaded = serverOverloaded;
   }
 
-  /**
-   * @return True if server was considered overloaded when exception was thrown
-   */
+  /** Returns True if server was considered overloaded when exception was thrown */
   public boolean isServerOverloaded() {
     return serverOverloaded;
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java
index 4694afbb0b1..d55b417d482 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java
@@ -193,17 +193,13 @@ public class HColumnDescriptor implements ColumnFamilyDescriptor, Comparable<HCo
     return ColumnFamilyDescriptorBuilder.isLegalColumnFamilyName(b);
   }
 
-  /**
-   * @return Name of this column family
-   */
+  /** Returns Name of this column family */
   @Override
   public byte[] getName() {
     return delegatee.getName();
   }
 
-  /**
-   * @return The name string of this column family
-   */
+  /** Returns The name string of this column family */
   @Override
   public String getNameAsString() {
     return delegatee.getNameAsString();
@@ -650,9 +646,7 @@ public class HColumnDescriptor implements ColumnFamilyDescriptor, Comparable<HCo
     return delegatee.toString();
   }
 
-  /**
-   * @return Column family descriptor with only the customized attributes.
-   */
+  /** Returns Column family descriptor with only the customized attributes. */
   @Override
   public String toStringCustomizedValues() {
     return delegatee.toStringCustomizedValues();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java
index 1769704a992..5e9abd31a7e 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java
@@ -104,7 +104,7 @@ public class HRegionInfo implements RegionInfo {
   }
 
   /**
-   * @return Return a short, printable name for this region (usually encoded name) for us logging.
+   * Returns Return a short, printable name for this region (usually encoded name) for us logging.
    */
   @Override
   public String getShortNameToLog() {
@@ -425,7 +425,7 @@ public class HRegionInfo implements RegionInfo {
     return RegionInfo.isEncodedRegionName(regionName);
   }
 
-  /** @return the regionId */
+  /** Returns the regionId */
   @Override
   public long getRegionId() {
     return regionId;
@@ -440,9 +440,7 @@ public class HRegionInfo implements RegionInfo {
     return regionName;
   }
 
-  /**
-   * @return Region name as a String for use in logging, etc.
-   */
+  /** Returns Region name as a String for use in logging, etc. */
   @Override
   public String getRegionNameAsString() {
     if (RegionInfo.hasEncodedName(this.regionName)) {
@@ -456,9 +454,7 @@ public class HRegionInfo implements RegionInfo {
     return Bytes.toStringBinary(this.regionName) + "." + this.getEncodedName();
   }
 
-  /**
-   * @return the encoded region name
-   */
+  /** Returns the encoded region name */
   @Override
   public synchronized String getEncodedName() {
     if (this.encodedName == null) {
@@ -475,17 +471,13 @@ public class HRegionInfo implements RegionInfo {
     return this.encodedNameAsBytes;
   }
 
-  /**
-   * @return the startKey
-   */
+  /** Returns the startKey */
   @Override
   public byte[] getStartKey() {
     return startKey;
   }
 
-  /**
-   * @return the endKey
-   */
+  /** Returns the endKey */
   @Override
   public byte[] getEndKey() {
     return endKey;
@@ -524,40 +516,30 @@ public class HRegionInfo implements RegionInfo {
     return firstKeyInRange && lastKeyInRange;
   }
 
-  /**
-   * @return true if the given row falls in this region.
-   */
+  /** Returns true if the given row falls in this region. */
   @Override
   public boolean containsRow(byte[] row) {
     return Bytes.compareTo(row, startKey) >= 0
       && (Bytes.compareTo(row, endKey) < 0 || Bytes.equals(endKey, HConstants.EMPTY_BYTE_ARRAY));
   }
 
-  /**
-   * @return true if this region is from hbase:meta
-   */
+  /** Returns true if this region is from hbase:meta */
   public boolean isMetaTable() {
     return isMetaRegion();
   }
 
-  /**
-   * @return true if this region is a meta region
-   */
+  /** Returns true if this region is a meta region */
   @Override
   public boolean isMetaRegion() {
     return tableName.equals(HRegionInfo.FIRST_META_REGIONINFO.getTable());
   }
 
-  /**
-   * @return true if this region is from a system table
-   */
+  /** Returns true if this region is from a system table */
   public boolean isSystemTable() {
     return tableName.isSystemTable();
   }
 
-  /**
-   * @return true if has been split and has daughters.
-   */
+  /** Returns true if has been split and has daughters. */
   @Override
   public boolean isSplit() {
     return this.split;
@@ -570,9 +552,7 @@ public class HRegionInfo implements RegionInfo {
     this.split = split;
   }
 
-  /**
-   * @return true if this region is offline.
-   */
+  /** Returns true if this region is offline. */
   @Override
   public boolean isOffline() {
     return this.offLine;
@@ -587,9 +567,7 @@ public class HRegionInfo implements RegionInfo {
     this.offLine = offLine;
   }
 
-  /**
-   * @return true if this is a split parent region.
-   */
+  /** Returns true if this is a split parent region. */
   @Override
   public boolean isSplitParent() {
     if (!isSplit()) return false;
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
index bdaad3beafe..6c8ea810d03 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
@@ -160,9 +160,7 @@ public class HTableDescriptor implements TableDescriptor, Comparable<HTableDescr
     return delegatee.isMetaTable();
   }
 
-  /**
-   * @return Getter for fetching an unmodifiable map.
-   */
+  /** Returns Getter for fetching an unmodifiable map. */
   @Override
   public Map<Bytes, Bytes> getValues() {
     return delegatee.getValues();
@@ -525,9 +523,7 @@ public class HTableDescriptor implements TableDescriptor, Comparable<HTableDescr
     return delegatee.toStringCustomizedValues();
   }
 
-  /**
-   * @return map of all table attributes formatted into string.
-   */
+  /** Returns map of all table attributes formatted into string. */
   public String toStringTableAttributes() {
     return delegatee.toStringTableAttributes();
   }
@@ -611,9 +607,7 @@ public class HTableDescriptor implements TableDescriptor, Comparable<HTableDescr
     return hasRegionMemStoreReplication();
   }
 
-  /**
-   * @return true if the read-replicas memstore replication is enabled.
-   */
+  /** Returns true if the read-replicas memstore replication is enabled. */
   @Override
   public boolean hasRegionMemStoreReplication() {
     return delegatee.hasRegionMemStoreReplication();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java
index 543cabde279..01b5f49a205 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java
@@ -304,9 +304,7 @@ public class MetaTableAccessor {
       regionInfo.getReplicaId());
   }
 
-  /**
-   * @return Return the {@link HConstants#CATALOG_FAMILY} row from hbase:meta.
-   */
+  /** Returns Return the {@link HConstants#CATALOG_FAMILY} row from hbase:meta. */
   public static Result getCatalogFamilyRow(Connection connection, RegionInfo ri)
     throws IOException {
     Get get = new Get(getMetaKeyForRegion(ri));
@@ -433,9 +431,7 @@ public class MetaTableAccessor {
     return false;
   }
 
-  /**
-   * @return True if the column in <code>cell</code> matches the regex 'info:merge.*'.
-   */
+  /** Returns True if the column in <code>cell</code> matches the regex 'info:merge.*'. */
   private static boolean isMergeQualifierPrefix(Cell cell) {
     // Check to see if has family and that qualifier starts with the merge qualifier 'merge'
     return CellUtil.matchingFamily(cell, HConstants.CATALOG_FAMILY)
@@ -825,9 +821,7 @@ public class MetaTableAccessor {
     }
   }
 
-  /**
-   * @return Get closest metatable region row to passed <code>row</code>
-   */
+  /** Returns Get closest metatable region row to passed <code>row</code> */
   @NonNull
   private static RegionInfo getClosestRegionInfo(Connection connection,
     @NonNull final TableName tableName, @NonNull final byte[] row) throws IOException {
@@ -1228,9 +1222,7 @@ public class MetaTableAccessor {
 
     abstract void add(Result r);
 
-    /**
-     * @return Collected results; wait till visits complete to collect all possible results
-     */
+    /** Returns Collected results; wait till visits complete to collect all possible results */
     List<T> getResults() {
       return this.results;
     }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/RegionLoad.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/RegionLoad.java
index 6184940313d..0a762bf78a6 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/RegionLoad.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/RegionLoad.java
@@ -332,9 +332,7 @@ public class RegionLoad implements RegionMetrics {
     return (int) metrics.getUncompressedStoreFileSize().get(Size.Unit.KILOBYTE);
   }
 
-  /**
-   * @return the data locality of region in the regionserver.
-   */
+  /** Returns the data locality of region in the regionserver. */
   @Override
   public float getDataLocality() {
     return metrics.getDataLocality();
@@ -355,9 +353,7 @@ public class RegionLoad implements RegionMetrics {
     return metrics.getLastMajorCompactionTimestamp();
   }
 
-  /**
-   * @return the reference count for the stores of this region
-   */
+  /** Returns the reference count for the stores of this region */
   public int getStoreRefCount() {
     return metrics.getStoreRefCount();
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/RegionMetrics.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/RegionMetrics.java
index 73d036e7370..645a31a8552 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/RegionMetrics.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/RegionMetrics.java
@@ -28,58 +28,38 @@ import org.apache.yetus.audience.InterfaceAudience;
 @InterfaceAudience.Public
 public interface RegionMetrics {
 
-  /**
-   * @return the region name
-   */
+  /** Returns the region name */
   byte[] getRegionName();
 
-  /**
-   * @return the number of stores
-   */
+  /** Returns the number of stores */
   int getStoreCount();
 
-  /**
-   * @return the number of storefiles
-   */
+  /** Returns the number of storefiles */
   int getStoreFileCount();
 
-  /**
-   * @return the total size of the storefiles
-   */
+  /** Returns the total size of the storefiles */
   Size getStoreFileSize();
 
-  /**
-   * @return the memstore size
-   */
+  /** Returns the memstore size */
   Size getMemStoreSize();
 
-  /**
-   * @return the number of read requests made to region
-   */
+  /** Returns the number of read requests made to region */
   long getReadRequestCount();
 
-  /**
-   * @return the number of write requests made to region
-   */
+  /** Returns the number of write requests made to region */
   long getWriteRequestCount();
 
-  /**
-   * @return the number of write requests and read requests made to region
-   */
+  /** Returns the number of write requests and read requests made to region */
   default long getRequestCount() {
     return getReadRequestCount() + getWriteRequestCount();
   }
 
-  /**
-   * @return the region name as a string
-   */
+  /** Returns the region name as a string */
   default String getNameAsString() {
     return Bytes.toStringBinary(getRegionName());
   }
 
-  /**
-   * @return the number of filtered read requests made to region
-   */
+  /** Returns the number of filtered read requests made to region */
   long getFilteredReadRequestCount();
 
   /**
@@ -90,29 +70,19 @@ public interface RegionMetrics {
    */
   Size getStoreFileIndexSize();
 
-  /**
-   * @return The current total size of root-level indexes for the region
-   */
+  /** Returns The current total size of root-level indexes for the region */
   Size getStoreFileRootLevelIndexSize();
 
-  /**
-   * @return The total size of all index blocks, not just the root level
-   */
+  /** Returns The total size of all index blocks, not just the root level */
   Size getStoreFileUncompressedDataIndexSize();
 
-  /**
-   * @return The total size of all Bloom filter blocks, not just loaded into the block cache
-   */
+  /** Returns The total size of all Bloom filter blocks, not just loaded into the block cache */
   Size getBloomFilterSize();
 
-  /**
-   * @return the total number of cells in current compaction
-   */
+  /** Returns the total number of cells in current compaction */
   long getCompactingCellCount();
 
-  /**
-   * @return the number of already compacted kvs in current compaction
-   */
+  /** Returns the number of already compacted kvs in current compaction */
   long getCompactedCellCount();
 
   /**
@@ -121,29 +91,19 @@ public interface RegionMetrics {
    */
   long getCompletedSequenceId();
 
-  /**
-   * @return completed sequence id per store.
-   */
+  /** Returns completed sequence id per store. */
   Map<byte[], Long> getStoreSequenceId();
 
-  /**
-   * @return the uncompressed size of the storefiles
-   */
+  /** Returns the uncompressed size of the storefiles */
   Size getUncompressedStoreFileSize();
 
-  /**
-   * @return the data locality of region in the regionserver.
-   */
+  /** Returns the data locality of region in the regionserver. */
   float getDataLocality();
 
-  /**
-   * @return the timestamp of the oldest hfile for any store of this region.
-   */
+  /** Returns the timestamp of the oldest hfile for any store of this region. */
   long getLastMajorCompactionTimestamp();
 
-  /**
-   * @return the reference count for the stores of this region
-   */
+  /** Returns the reference count for the stores of this region */
   int getStoreRefCount();
 
   /**
@@ -158,9 +118,7 @@ public interface RegionMetrics {
    */
   float getDataLocalityForSsd();
 
-  /**
-   * @return the data at local weight of this region in the regionserver
-   */
+  /** Returns the data at local weight of this region in the regionserver */
   long getBlocksLocalWeight();
 
   /**
@@ -169,13 +127,9 @@ public interface RegionMetrics {
    */
   long getBlocksLocalWithSsdWeight();
 
-  /**
-   * @return the block total weight of this region
-   */
+  /** Returns the block total weight of this region */
   long getBlocksTotalWeight();
 
-  /**
-   * @return the compaction state of this region
-   */
+  /** Returns the compaction state of this region */
   CompactionState getCompactionState();
 }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerMetrics.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerMetrics.java
index 46678ac2460..5cf190c04f9 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerMetrics.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerMetrics.java
@@ -33,38 +33,26 @@ public interface ServerMetrics {
 
   ServerName getServerName();
 
-  /**
-   * @return the version number of a regionserver.
-   */
+  /** Returns the version number of a regionserver. */
   default int getVersionNumber() {
     return 0;
   }
 
-  /**
-   * @return the string type version of a regionserver.
-   */
+  /** Returns the string type version of a regionserver. */
   default String getVersion() {
     return "0.0.0";
   }
 
-  /**
-   * @return the number of requests per second.
-   */
+  /** Returns the number of requests per second. */
   long getRequestCountPerSecond();
 
-  /**
-   * @return total Number of requests from the start of the region server.
-   */
+  /** Returns total Number of requests from the start of the region server. */
   long getRequestCount();
 
-  /**
-   * @return the amount of used heap
-   */
+  /** Returns the amount of used heap */
   Size getUsedHeapSize();
 
-  /**
-   * @return the maximum allowable size of the heap
-   */
+  /** Returns the maximum allowable size of the heap */
   Size getMaxHeapSize();
 
   int getInfoServerPort();
@@ -87,14 +75,10 @@ public interface ServerMetrics {
   @Nullable
   ReplicationLoadSink getReplicationLoadSink();
 
-  /**
-   * @return region load metrics
-   */
+  /** Returns region load metrics */
   Map<byte[], RegionMetrics> getRegionMetrics();
 
-  /**
-   * @return metrics per user
-   */
+  /** Returns metrics per user */
   Map<byte[], UserMetrics> getUserMetrics();
 
   /**
@@ -103,14 +87,10 @@ public interface ServerMetrics {
    */
   Set<String> getCoprocessorNames();
 
-  /**
-   * @return the timestamp (server side) of generating this metrics
-   */
+  /** Returns the timestamp (server side) of generating this metrics */
   long getReportTimestamp();
 
-  /**
-   * @return the last timestamp (server side) of generating this metrics
-   */
+  /** Returns the last timestamp (server side) of generating this metrics */
   long getLastReportTimestamp();
 
   /**
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/Size.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/Size.java
index c248849e363..99062635f77 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/Size.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/Size.java
@@ -69,9 +69,7 @@ public final class Size implements Comparable<Size> {
     this.unit = Preconditions.checkNotNull(unit);
   }
 
-  /**
-   * @return size unit
-   */
+  /** Returns size unit */
   public Unit getUnit() {
     return unit;
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/UserMetrics.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/UserMetrics.java
index 2710aa9be27..05108c70e74 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/UserMetrics.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/UserMetrics.java
@@ -40,19 +40,13 @@ public interface UserMetrics {
     long getFilteredReadRequestsCount();
   }
 
-  /**
-   * @return the user name
-   */
+  /** Returns the user name */
   byte[] getUserName();
 
-  /**
-   * @return the number of read requests made by user
-   */
+  /** Returns the number of read requests made by user */
   long getReadRequestCount();
 
-  /**
-   * @return the number of write requests made by user
-   */
+  /** Returns the number of write requests made by user */
   long getWriteRequestCount();
 
   /**
@@ -63,20 +57,14 @@ public interface UserMetrics {
     return getReadRequestCount() + getWriteRequestCount();
   }
 
-  /**
-   * @return the user name as a string
-   */
+  /** Returns the user name as a string */
   default String getNameAsString() {
     return Bytes.toStringBinary(getUserName());
   }
 
-  /**
-   * @return metrics per client(hostname)
-   */
+  /** Returns metrics per client(hostname) */
   Map<String, ClientMetrics> getClientMetrics();
 
-  /**
-   * @return count of filtered read requests for a user
-   */
+  /** Returns count of filtered read requests for a user */
   long getFilteredReadRequests();
 }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java
index a0505463f72..6a0913c23c7 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java
@@ -112,9 +112,7 @@ public interface Admin extends Abortable, Closeable {
   @Override
   boolean isAborted();
 
-  /**
-   * @return Connection used by this object.
-   */
+  /** Returns Connection used by this object. */
   Connection getConnection();
 
   /**
@@ -1710,9 +1708,7 @@ public interface Admin extends Abortable, Closeable {
   List<RegionMetrics> getRegionMetrics(ServerName serverName, TableName tableName)
     throws IOException;
 
-  /**
-   * @return Configuration used by the instance.
-   */
+  /** Returns Configuration used by the instance. */
   Configuration getConfiguration();
 
   /**
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Append.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Append.java
index c83bc76f56b..bf1a29196db 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Append.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Append.java
@@ -90,9 +90,7 @@ public class Append extends Mutation {
     return this;
   }
 
-  /**
-   * @return current setting for returnResults
-   */
+  /** Returns current setting for returnResults */
   // This method makes public the superclasses's protected method.
   @Override
   public boolean isReturnResults() {
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncAdmin.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncAdmin.java
index a47ca261e03..2d6d499042e 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncAdmin.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncAdmin.java
@@ -1051,34 +1051,24 @@ public interface AsyncAdmin {
   CompletableFuture<Void> recommissionRegionServer(ServerName server,
     List<byte[]> encodedRegionNames);
 
-  /**
-   * @return cluster status wrapped by {@link CompletableFuture}
-   */
+  /** Returns cluster status wrapped by {@link CompletableFuture} */
   CompletableFuture<ClusterMetrics> getClusterMetrics();
 
-  /**
-   * @return cluster status wrapped by {@link CompletableFuture}
-   */
+  /** Returns cluster status wrapped by {@link CompletableFuture} */
   CompletableFuture<ClusterMetrics> getClusterMetrics(EnumSet<Option> options);
 
-  /**
-   * @return current master server name wrapped by {@link CompletableFuture}
-   */
+  /** Returns current master server name wrapped by {@link CompletableFuture} */
   default CompletableFuture<ServerName> getMaster() {
     return getClusterMetrics(EnumSet.of(Option.MASTER)).thenApply(ClusterMetrics::getMasterName);
   }
 
-  /**
-   * @return current backup master list wrapped by {@link CompletableFuture}
-   */
+  /** Returns current backup master list wrapped by {@link CompletableFuture} */
   default CompletableFuture<Collection<ServerName>> getBackupMasters() {
     return getClusterMetrics(EnumSet.of(Option.BACKUP_MASTERS))
       .thenApply(ClusterMetrics::getBackupMasterNames);
   }
 
-  /**
-   * @return current live region servers list wrapped by {@link CompletableFuture}
-   */
+  /** Returns current live region servers list wrapped by {@link CompletableFuture} */
   default CompletableFuture<Collection<ServerName>> getRegionServers() {
     return getClusterMetrics(EnumSet.of(Option.SERVERS_NAME))
       .thenApply(ClusterMetrics::getServersName);
@@ -1110,9 +1100,7 @@ public interface AsyncAdmin {
     return future;
   }
 
-  /**
-   * @return a list of master coprocessors wrapped by {@link CompletableFuture}
-   */
+  /** Returns a list of master coprocessors wrapped by {@link CompletableFuture} */
   default CompletableFuture<List<String>> getMasterCoprocessorNames() {
     return getClusterMetrics(EnumSet.of(Option.MASTER_COPROCESSORS))
       .thenApply(ClusterMetrics::getMasterCoprocessorNames);
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CheckAndMutate.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CheckAndMutate.java
index 9202e310e7f..e8949ee65ec 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CheckAndMutate.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CheckAndMutate.java
@@ -250,9 +250,7 @@ public final class CheckAndMutate implements Row {
     this.action = action;
   }
 
-  /**
-   * @return the row
-   */
+  /** Returns the row */
   @Override
   public byte[] getRow() {
     return row;
@@ -281,58 +279,42 @@ public final class CheckAndMutate implements Row {
     return Bytes.hashCode(this.getRow());
   }
 
-  /**
-   * @return the family to check
-   */
+  /** Returns the family to check */
   public byte[] getFamily() {
     return family;
   }
 
-  /**
-   * @return the qualifier to check
-   */
+  /** Returns the qualifier to check */
   public byte[] getQualifier() {
     return qualifier;
   }
 
-  /**
-   * @return the comparison operator
-   */
+  /** Returns the comparison operator */
   public CompareOperator getCompareOp() {
     return op;
   }
 
-  /**
-   * @return the expected value
-   */
+  /** Returns the expected value */
   public byte[] getValue() {
     return value;
   }
 
-  /**
-   * @return the filter to check
-   */
+  /** Returns the filter to check */
   public Filter getFilter() {
     return filter;
   }
 
-  /**
-   * @return whether this has a filter or not
-   */
+  /** Returns whether this has a filter or not */
   public boolean hasFilter() {
     return filter != null;
   }
 
-  /**
-   * @return the time range to check
-   */
+  /** Returns the time range to check */
   public TimeRange getTimeRange() {
     return timeRange;
   }
 
-  /**
-   * @return the action done if check succeeds
-   */
+  /** Returns the action done if check succeeds */
   public Row getAction() {
     return action;
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CheckAndMutateResult.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CheckAndMutateResult.java
index 88b438a74a0..8ecb49e3d5f 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CheckAndMutateResult.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CheckAndMutateResult.java
@@ -32,16 +32,12 @@ public class CheckAndMutateResult {
     this.result = result;
   }
 
-  /**
-   * @return Whether the CheckAndMutate operation is successful or not
-   */
+  /** Returns Whether the CheckAndMutate operation is successful or not */
   public boolean isSuccess() {
     return success;
   }
 
-  /**
-   * @return It is used only for CheckAndMutate operations with Increment/Append. Otherwise null
-   */
+  /** Returns It is used only for CheckAndMutate operations with Increment/Append. Otherwise null */
   public Result getResult() {
     return result;
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientIdGenerator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientIdGenerator.java
index 758cf508578..f5bedbe88d2 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientIdGenerator.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientIdGenerator.java
@@ -58,9 +58,7 @@ final class ClientIdGenerator {
     return id;
   }
 
-  /**
-   * @return PID of the current process, if it can be extracted from JVM name, or null.
-   */
+  /** Returns PID of the current process, if it can be extracted from JVM name, or null. */
   public static Long getPid() {
     String name = ManagementFactory.getRuntimeMXBean().getName();
     String[] nameParts = name.split("@");
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClusterConnection.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClusterConnection.java
index c7fde3d9214..849745b89eb 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClusterConnection.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClusterConnection.java
@@ -246,13 +246,11 @@ public interface ClusterConnection extends Connection {
   void clearCaches(final ServerName sn);
 
   /**
-   * @return Nonce generator for this ClusterConnection; may be null if disabled in configuration.
+   * Returns Nonce generator for this ClusterConnection; may be null if disabled in configuration.
    */
   NonceGenerator getNonceGenerator();
 
-  /**
-   * @return Default AsyncProcess associated with this connection.
-   */
+  /** Returns Default AsyncProcess associated with this connection. */
   AsyncProcess getAsyncProcess();
 
   /**
@@ -263,34 +261,22 @@ public interface ClusterConnection extends Connection {
    */
   RpcRetryingCallerFactory getNewRpcRetryingCallerFactory(Configuration conf);
 
-  /**
-   * @return Connection's RpcRetryingCallerFactory instance
-   */
+  /** Returns Connection's RpcRetryingCallerFactory instance */
   RpcRetryingCallerFactory getRpcRetryingCallerFactory();
 
-  /**
-   * @return Connection's RpcControllerFactory instance
-   */
+  /** Returns Connection's RpcControllerFactory instance */
   RpcControllerFactory getRpcControllerFactory();
 
-  /**
-   * @return a ConnectionConfiguration object holding parsed configuration values
-   */
+  /** Returns a ConnectionConfiguration object holding parsed configuration values */
   ConnectionConfiguration getConnectionConfiguration();
 
-  /**
-   * @return the current statistics tracker associated with this connection
-   */
+  /** Returns the current statistics tracker associated with this connection */
   ServerStatisticTracker getStatisticsTracker();
 
-  /**
-   * @return the configured client backoff policy
-   */
+  /** Returns the configured client backoff policy */
   ClientBackoffPolicy getBackoffPolicy();
 
-  /**
-   * @return the MetricsConnection instance associated with this connection.
-   */
+  /** Returns the MetricsConnection instance associated with this connection. */
   MetricsConnection getConnectionMetrics();
 
   /**
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ColumnFamilyDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ColumnFamilyDescriptor.java
index 5f11f8d258a..b3d8e4fd600 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ColumnFamilyDescriptor.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ColumnFamilyDescriptor.java
@@ -77,39 +77,25 @@ public interface ColumnFamilyDescriptor {
       return lcf.getConfiguration().hashCode() - rcf.getConfiguration().hashCode();
     };
 
-  /**
-   * @return The storefile/hfile blocksize for this column family.
-   */
+  /** Returns The storefile/hfile blocksize for this column family. */
   int getBlocksize();
 
-  /**
-   * @return bloom filter type used for new StoreFiles in ColumnFamily
-   */
+  /** Returns bloom filter type used for new StoreFiles in ColumnFamily */
   BloomType getBloomFilterType();
 
-  /**
-   * @return Compression type setting.
-   */
+  /** Returns Compression type setting. */
   Compression.Algorithm getCompactionCompressionType();
 
-  /**
-   * @return Compression type setting for major compactions.
-   */
+  /** Returns Compression type setting for major compactions. */
   Compression.Algorithm getMajorCompactionCompressionType();
 
-  /**
-   * @return Compression type setting for minor compactions.
-   */
+  /** Returns Compression type setting for minor compactions. */
   Compression.Algorithm getMinorCompactionCompressionType();
 
-  /**
-   * @return Compression type setting.
-   */
+  /** Returns Compression type setting. */
   Compression.Algorithm getCompressionType();
 
-  /**
-   * @return an unmodifiable map.
-   */
+  /** Returns an unmodifiable map. */
   Map<String, String> getConfiguration();
 
   /**
@@ -118,24 +104,16 @@ public interface ColumnFamilyDescriptor {
    */
   String getConfigurationValue(String key);
 
-  /**
-   * @return replication factor set for this CF
-   */
+  /** Returns replication factor set for this CF */
   short getDFSReplication();
 
-  /**
-   * @return the data block encoding algorithm used in block cache and optionally on disk
-   */
+  /** Returns the data block encoding algorithm used in block cache and optionally on disk */
   DataBlockEncoding getDataBlockEncoding();
 
-  /**
-   * @return Return the raw crypto key attribute for the family, or null if not set
-   */
+  /** Returns Return the raw crypto key attribute for the family, or null if not set */
   byte[] getEncryptionKey();
 
-  /**
-   * @return Return the encryption algorithm in use by this family
-   */
+  /** Returns Return the encryption algorithm in use by this family */
   String getEncryptionType();
 
   /**
@@ -144,19 +122,13 @@ public interface ColumnFamilyDescriptor {
    */
   MemoryCompactionPolicy getInMemoryCompaction();
 
-  /**
-   * @return return the KeepDeletedCells
-   */
+  /** Returns return the KeepDeletedCells */
   KeepDeletedCells getKeepDeletedCells();
 
-  /**
-   * @return maximum number of versions
-   */
+  /** Returns maximum number of versions */
   int getMaxVersions();
 
-  /**
-   * @return The minimum number of versions to keep.
-   */
+  /** Returns The minimum number of versions to keep. */
   int getMinVersions();
 
   /**
@@ -171,19 +143,13 @@ public interface ColumnFamilyDescriptor {
    */
   long getMobThreshold();
 
-  /**
-   * @return a copy of Name of this column family
-   */
+  /** Returns a copy of Name of this column family */
   byte[] getName();
 
-  /**
-   * @return Name of this column family
-   */
+  /** Returns Name of this column family */
   String getNameAsString();
 
-  /**
-   * @return the scope tag
-   */
+  /** Returns the scope tag */
   int getScope();
 
   /**
@@ -193,9 +159,7 @@ public interface ColumnFamilyDescriptor {
    */
   String getStoragePolicy();
 
-  /**
-   * @return Time-to-live of cell contents, in seconds.
-   */
+  /** Returns Time-to-live of cell contents, in seconds. */
   int getTimeToLive();
 
   /**
@@ -228,19 +192,13 @@ public interface ColumnFamilyDescriptor {
    */
   boolean isBlockCacheEnabled();
 
-  /**
-   * @return true if we should cache bloomfilter blocks on write
-   */
+  /** Returns true if we should cache bloomfilter blocks on write */
   boolean isCacheBloomsOnWrite();
 
-  /**
-   * @return true if we should cache data blocks on write
-   */
+  /** Returns true if we should cache data blocks on write */
   boolean isCacheDataOnWrite();
 
-  /**
-   * @return true if we should cache index blocks on write
-   */
+  /** Returns true if we should cache index blocks on write */
   boolean isCacheIndexesOnWrite();
 
   /**
@@ -249,9 +207,7 @@ public interface ColumnFamilyDescriptor {
    */
   boolean isCompressTags();
 
-  /**
-   * @return true if we should evict cached blocks from the blockcache on close
-   */
+  /** Returns true if we should evict cached blocks from the blockcache on close */
   boolean isEvictBlocksOnClose();
 
   /**
@@ -266,14 +222,10 @@ public interface ColumnFamilyDescriptor {
    */
   boolean isMobEnabled();
 
-  /**
-   * @return true if we should prefetch blocks into the blockcache on open
-   */
+  /** Returns true if we should prefetch blocks into the blockcache on open */
   boolean isPrefetchBlocksOnOpen();
 
-  /**
-   * @return Column family descriptor with only the customized attributes.
-   */
+  /** Returns Column family descriptor with only the customized attributes. */
   String toStringCustomizedValues();
 
   /**
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Connection.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Connection.java
index 4a8e47567ea..7db973d3f28 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Connection.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Connection.java
@@ -58,9 +58,7 @@ public interface Connection extends Abortable, Closeable {
    * general methods.
    */
 
-  /**
-   * @return Configuration instance being used by this Connection instance.
-   */
+  /** Returns Configuration instance being used by this Connection instance. */
   Configuration getConfiguration();
 
   /**
@@ -173,9 +171,7 @@ public interface Connection extends Abortable, Closeable {
    */
   TableBuilder getTableBuilder(TableName tableName, ExecutorService pool);
 
-  /**
-   * @return the cluster ID unique to this HBase cluster.
-   */
+  /** Returns the cluster ID unique to this HBase cluster. */
   String getClusterId();
 
   /**
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionRegistryFactory.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionRegistryFactory.java
index faaf61a4f52..bba75aabfc1 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionRegistryFactory.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionRegistryFactory.java
@@ -32,9 +32,7 @@ final class ConnectionRegistryFactory {
   private ConnectionRegistryFactory() {
   }
 
-  /**
-   * @return The connection registry implementation to use.
-   */
+  /** Returns The connection registry implementation to use. */
   static ConnectionRegistry getRegistry(Configuration conf) {
     Class<? extends ConnectionRegistry> clazz =
       conf.getClass(CLIENT_CONNECTION_REGISTRY_IMPL_CONF_KEY, ZKConnectionRegistry.class,
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CoprocessorDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CoprocessorDescriptor.java
index 3331c810700..40cb5abfa96 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CoprocessorDescriptor.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CoprocessorDescriptor.java
@@ -28,23 +28,17 @@ import org.apache.yetus.audience.InterfaceAudience;
  */
 @InterfaceAudience.Public
 public interface CoprocessorDescriptor {
-  /**
-   * @return the name of the class or interface represented by this object.
-   */
+  /** Returns the name of the class or interface represented by this object. */
   String getClassName();
 
   /**
-   * @return Path of the jar file. If it's null, the class will be loaded from default classloader.
+   * Returns Path of the jar file. If it's null, the class will be loaded from default classloader.
    */
   Optional<String> getJarPath();
 
-  /**
-   * @return The order to execute this coprocessor
-   */
+  /** Returns The order to execute this coprocessor */
   int getPriority();
 
-  /**
-   * @return Arbitrary key-value parameter pairs passed into the coprocessor.
-   */
+  /** Returns Arbitrary key-value parameter pairs passed into the coprocessor. */
   Map<String, String> getProperties();
 }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java
index ca2cf11d3b2..90b499e416c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java
@@ -483,7 +483,7 @@ public class HBaseAdmin implements Admin {
     }
   }
 
-  /** @return Connection used by this object. */
+  /** Returns Connection used by this object. */
   @Override
   public Connection getConnection() {
     return connection;
@@ -3531,28 +3531,20 @@ public class HBaseAdmin implements Admin {
       return getDescription();
     }
 
-    /**
-     * @return the table name
-     */
+    /** Returns the table name */
     protected TableName getTableName() {
       return tableName;
     }
 
-    /**
-     * @return the table descriptor
-     */
+    /** Returns the table descriptor */
     protected TableDescriptor getTableDescriptor() throws IOException {
       return getAdmin().getDescriptor(getTableName());
     }
 
-    /**
-     * @return the operation type like CREATE, DELETE, DISABLE etc.
-     */
+    /** Returns the operation type like CREATE, DELETE, DISABLE etc. */
     public abstract String getOperationType();
 
-    /**
-     * @return a description of the operation
-     */
+    /** Returns a description of the operation */
     protected String getDescription() {
       return "Operation: " + getOperationType() + ", " + "Table Name: "
         + tableName.getNameWithNamespaceInclAsString() + ", procId: " + procId;
@@ -3696,16 +3688,12 @@ public class HBaseAdmin implements Admin {
       this.namespaceName = namespaceName;
     }
 
-    /**
-     * @return the namespace name
-     */
+    /** Returns the namespace name */
     protected String getNamespaceName() {
       return namespaceName;
     }
 
-    /**
-     * @return the operation type like CREATE_NAMESPACE, DELETE_NAMESPACE, etc.
-     */
+    /** Returns the operation type like CREATE_NAMESPACE, DELETE_NAMESPACE, etc. */
     public abstract String getOperationType();
 
     @Override
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java
index fcc928b5c63..acb5c9fda5d 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java
@@ -204,9 +204,7 @@ public class HTable implements Table {
     this.locator = new HRegionLocator(tableName, connection);
   }
 
-  /**
-   * @return maxKeyValueSize from configuration.
-   */
+  /** Returns maxKeyValueSize from configuration. */
   public static int getMaxKeyValueSize(Configuration conf) {
     return conf.getInt(ConnectionConfiguration.MAX_KEYVALUE_SIZE_KEY, -1);
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableMultiplexer.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableMultiplexer.java
index fac13a831a6..7404d460b55 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableMultiplexer.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableMultiplexer.java
@@ -216,9 +216,7 @@ public class HTableMultiplexer {
     return put(TableName.valueOf(tableName), put);
   }
 
-  /**
-   * @return the current HTableMultiplexerStatus
-   */
+  /** Returns the current HTableMultiplexerStatus */
   public HTableMultiplexerStatus getHTableMultiplexerStatus() {
     return new HTableMultiplexerStatus(serverToFlushWorkerMap);
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Increment.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Increment.java
index 2a9bec60329..0efd71e69d2 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Increment.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Increment.java
@@ -162,9 +162,7 @@ public class Increment extends Mutation {
     return this;
   }
 
-  /**
-   * @return current setting for returnResults
-   */
+  /** Returns current setting for returnResults */
   // This method makes public the superclasses's protected method.
   @Override
   public boolean isReturnResults() {
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/MultiResponse.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/MultiResponse.java
index 97a01e1bb6e..b768b76fdb6 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/MultiResponse.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/MultiResponse.java
@@ -44,9 +44,7 @@ public class MultiResponse extends AbstractResponse {
     super();
   }
 
-  /**
-   * @return Number of pairs in this container
-   */
+  /** Returns Number of pairs in this container */
   public int size() {
     int size = 0;
     for (RegionResult result : results.values()) {
@@ -68,9 +66,7 @@ public class MultiResponse extends AbstractResponse {
     exceptions.put(regionName, ie);
   }
 
-  /**
-   * @return the exception for the region, if any. Null otherwise.
-   */
+  /** Returns the exception for the region, if any. Null otherwise. */
   public Throwable getException(byte[] regionName) {
     return exceptions.get(regionName);
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/MutableRegionInfo.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/MutableRegionInfo.java
index 22114f8f624..7627f8dc0e5 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/MutableRegionInfo.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/MutableRegionInfo.java
@@ -127,14 +127,14 @@ class MutableRegionInfo implements RegionInfo {
   }
 
   /**
-   * @return Return a short, printable name for this region (usually encoded name) for us logging.
+   * Returns Return a short, printable name for this region (usually encoded name) for us logging.
    */
   @Override
   public String getShortNameToLog() {
     return RegionInfo.prettyPrint(this.getEncodedName());
   }
 
-  /** @return the regionId */
+  /** Returns the regionId */
   @Override
   public long getRegionId() {
     return regionId;
@@ -149,15 +149,13 @@ class MutableRegionInfo implements RegionInfo {
     return regionName;
   }
 
-  /**
-   * @return Region name as a String for use in logging, etc.
-   */
+  /** Returns Region name as a String for use in logging, etc. */
   @Override
   public String getRegionNameAsString() {
     return RegionInfo.getRegionNameAsString(this, this.regionName);
   }
 
-  /** @return the encoded region name */
+  /** Returns the encoded region name */
   @Override
   public String getEncodedName() {
     return this.encodedName;
@@ -168,13 +166,13 @@ class MutableRegionInfo implements RegionInfo {
     return this.encodedNameAsBytes;
   }
 
-  /** @return the startKey */
+  /** Returns the startKey */
   @Override
   public byte[] getStartKey() {
     return startKey;
   }
 
-  /** @return the endKey */
+  /** Returns the endKey */
   @Override
   public byte[] getEndKey() {
     return endKey;
@@ -219,15 +217,13 @@ class MutableRegionInfo implements RegionInfo {
         || Bytes.equals(endKey, HConstants.EMPTY_BYTE_ARRAY));
   }
 
-  /** @return true if this region is a meta region */
+  /** Returns true if this region is a meta region */
   @Override
   public boolean isMetaRegion() {
     return tableName.equals(TableName.META_TABLE_NAME);
   }
 
-  /**
-   * @return True if has been split and has daughters.
-   */
+  /** Returns True if has been split and has daughters. */
   @Override
   public boolean isSplit() {
     return this.split;
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Mutation.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Mutation.java
index 3c22b053a0d..408b02e44fa 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Mutation.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Mutation.java
@@ -354,9 +354,7 @@ public abstract class Mutation extends OperationWithAttributes
     return this;
   }
 
-  /**
-   * @return the set of clusterIds that have consumed the mutation
-   */
+  /** Returns the set of clusterIds that have consumed the mutation */
   public List<UUID> getClusterIds() {
     List<UUID> clusterIds = new ArrayList<>();
     byte[] bytes = getAttribute(CONSUMED_CLUSTER_IDS);
@@ -379,9 +377,7 @@ public abstract class Mutation extends OperationWithAttributes
     return this;
   }
 
-  /**
-   * @return CellVisibility associated with cells in this Mutation. n
-   */
+  /** Returns CellVisibility associated with cells in this Mutation. n */
   public CellVisibility getCellVisibility() throws DeserializationException {
     byte[] cellVisibilityBytes = this.getAttribute(VisibilityConstants.VISIBILITY_LABELS_ATTR_KEY);
     if (cellVisibilityBytes == null) return null;
@@ -437,16 +433,12 @@ public abstract class Mutation extends OperationWithAttributes
     return size;
   }
 
-  /**
-   * @return the number of different families
-   */
+  /** Returns the number of different families */
   public int numFamilies() {
     return getFamilyCellMap().size();
   }
 
-  /**
-   * @return Calculate what Mutation adds to class heap size.
-   */
+  /** Returns Calculate what Mutation adds to class heap size. */
   @Override
   public long heapSize() {
     long heapsize = MUTATION_OVERHEAD;
@@ -475,9 +467,7 @@ public abstract class Mutation extends OperationWithAttributes
     return ClassSize.align(heapsize);
   }
 
-  /**
-   * @return The serialized ACL for this operation, or null if none
-   */
+  /** Returns The serialized ACL for this operation, or null if none */
   public byte[] getACL() {
     return getAttribute(AccessControlConstants.OP_ATTRIBUTE_ACL);
   }
@@ -527,9 +517,7 @@ public abstract class Mutation extends OperationWithAttributes
     return this;
   }
 
-  /**
-   * @return current value for returnResults
-   */
+  /** Returns current value for returnResults */
   // Used by Increment and Append only.
   @InterfaceAudience.Private
   protected boolean isReturnResults() {
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/NonceGenerator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/NonceGenerator.java
index 3020be22105..83556e7e91a 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/NonceGenerator.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/NonceGenerator.java
@@ -29,9 +29,9 @@ public interface NonceGenerator {
 
   static final String CLIENT_NONCES_ENABLED_KEY = "hbase.client.nonces.enabled";
 
-  /** @return the nonce group (client ID) of this client manager. */
+  /** Returns the nonce group (client ID) of this client manager. */
   long getNonceGroup();
 
-  /** @return New nonce. */
+  /** Returns New nonce. */
   long newNonce();
 }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Query.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Query.java
index 661065d0949..d752693dabc 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Query.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Query.java
@@ -75,18 +75,14 @@ public abstract class Query extends OperationWithAttributes {
     return this;
   }
 
-  /**
-   * @return The authorizations this Query is associated with. n
-   */
+  /** Returns The authorizations this Query is associated with. n */
   public Authorizations getAuthorizations() throws DeserializationException {
     byte[] authorizationsBytes = this.getAttribute(VisibilityConstants.VISIBILITY_LABELS_ATTR_KEY);
     if (authorizationsBytes == null) return null;
     return ProtobufUtil.toAuthorizations(authorizationsBytes);
   }
 
-  /**
-   * @return The serialized ACL for this operation, or null if none
-   */
+  /** Returns The serialized ACL for this operation, or null if none */
   public byte[] getACL() {
     return getAttribute(AccessControlConstants.OP_ATTRIBUTE_ACL);
   }
@@ -220,9 +216,7 @@ public abstract class Query extends OperationWithAttributes {
     return this;
   }
 
-  /**
-   * @return A map of column families to time ranges
-   */
+  /** Returns A map of column families to time ranges */
   public Map<byte[], TimeRange> getColumnFamilyTimeRange() {
     return this.colFamTimeRangeMap;
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionAdminServiceCallable.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionAdminServiceCallable.java
index 301c4315d25..8b2ed345306 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionAdminServiceCallable.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionAdminServiceCallable.java
@@ -114,9 +114,7 @@ public abstract class RegionAdminServiceCallable<T> implements RetryingCallable<
     }
   }
 
-  /**
-   * @return {@link Connection} instance used by this Callable.
-   */
+  /** Returns {@link Connection} instance used by this Callable. */
   Connection getConnection() {
     return this.connection;
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfo.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfo.java
index b9a5969b4d2..de3f53707b8 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfo.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfo.java
@@ -152,13 +152,11 @@ public interface RegionInfo extends Comparable<RegionInfo> {
   };
 
   /**
-   * @return Return a short, printable name for this region (usually encoded name) for us logging.
+   * Returns Return a short, printable name for this region (usually encoded name) for us logging.
    */
   String getShortNameToLog();
 
-  /**
-   * @return the regionId.
-   */
+  /** Returns the regionId. */
   long getRegionId();
 
   /**
@@ -167,44 +165,28 @@ public interface RegionInfo extends Comparable<RegionInfo> {
    */
   byte[] getRegionName();
 
-  /**
-   * @return Region name as a String for use in logging, etc.
-   */
+  /** Returns Region name as a String for use in logging, etc. */
   String getRegionNameAsString();
 
-  /**
-   * @return the encoded region name.
-   */
+  /** Returns the encoded region name. */
   String getEncodedName();
 
-  /**
-   * @return the encoded region name as an array of bytes.
-   */
+  /** Returns the encoded region name as an array of bytes. */
   byte[] getEncodedNameAsBytes();
 
-  /**
-   * @return the startKey.
-   */
+  /** Returns the startKey. */
   byte[] getStartKey();
 
-  /**
-   * @return the endKey.
-   */
+  /** Returns the endKey. */
   byte[] getEndKey();
 
-  /**
-   * @return current table name of the region
-   */
+  /** Returns current table name of the region */
   TableName getTable();
 
-  /**
-   * @return returns region replica id
-   */
+  /** Returns returns region replica id */
   int getReplicaId();
 
-  /**
-   * @return True if has been split and has daughters.
-   */
+  /** Returns True if has been split and has daughters. */
   boolean isSplit();
 
   /**
@@ -223,9 +205,7 @@ public interface RegionInfo extends Comparable<RegionInfo> {
   @Deprecated
   boolean isSplitParent();
 
-  /**
-   * @return true if this region is a meta region.
-   */
+  /** Returns true if this region is a meta region. */
   boolean isMetaRegion();
 
   /**
@@ -236,9 +216,7 @@ public interface RegionInfo extends Comparable<RegionInfo> {
    */
   boolean containsRange(byte[] rangeStartKey, byte[] rangeEndKey);
 
-  /**
-   * @return true if the given row falls in this region.
-   */
+  /** Returns true if the given row falls in this region. */
   boolean containsRow(byte[] row);
 
   /**
@@ -253,9 +231,7 @@ public interface RegionInfo extends Comparable<RegionInfo> {
       && (regionName[regionName.length - 1] == RegionInfo.ENC_SEPARATOR);
   }
 
-  /**
-   * @return the encodedName
-   */
+  /** Returns the encodedName */
   @InterfaceAudience.Private
   static String encodeRegionName(final byte[] regionName) {
     String encodedName;
@@ -371,7 +347,7 @@ public interface RegionInfo extends Comparable<RegionInfo> {
   }
 
   /**
-   * @return A deserialized {@link RegionInfo} or null if we failed deserialize or passed bytes null
+   * Returns A deserialized {@link RegionInfo} or null if we failed deserialize or passed bytes null
    */
   @InterfaceAudience.Private
   static RegionInfo parseFromOrNull(final byte[] bytes) {
@@ -380,7 +356,7 @@ public interface RegionInfo extends Comparable<RegionInfo> {
   }
 
   /**
-   * @return A deserialized {@link RegionInfo} or null if we failed deserialize or passed bytes null
+   * Returns A deserialized {@link RegionInfo} or null if we failed deserialize or passed bytes null
    */
   @InterfaceAudience.Private
   static RegionInfo parseFromOrNull(final byte[] bytes, int offset, int len) {
@@ -778,16 +754,12 @@ public interface RegionInfo extends Comparable<RegionInfo> {
     return ris;
   }
 
-  /**
-   * @return True if this is first Region in Table
-   */
+  /** Returns True if this is first Region in Table */
   default boolean isFirst() {
     return Bytes.equals(getStartKey(), HConstants.EMPTY_START_ROW);
   }
 
-  /**
-   * @return True if this is last Region in Table
-   */
+  /** Returns True if this is last Region in Table */
   default boolean isLast() {
     return Bytes.equals(getEndKey(), HConstants.EMPTY_END_ROW);
   }
@@ -809,9 +781,7 @@ public interface RegionInfo extends Comparable<RegionInfo> {
     return getTable().equals(other.getTable()) && areAdjacent(this, other);
   }
 
-  /**
-   * @return True if RegionInfo is degenerate... if startKey > endKey.
-   */
+  /** Returns True if RegionInfo is degenerate... if startKey > endKey. */
   default boolean isDegenerate() {
     return !isLast() && Bytes.compareTo(getStartKey(), getEndKey()) > 0;
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionReplicaUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionReplicaUtil.java
index ea022820950..4475a01c0f9 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionReplicaUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionReplicaUtil.java
@@ -73,12 +73,12 @@ public class RegionReplicaUtil {
     return getRegionInfoForReplica(regionInfo, DEFAULT_REPLICA_ID);
   }
 
-  /** @return true if this replicaId corresponds to default replica for the region */
+  /** Returns true if this replicaId corresponds to default replica for the region */
   public static boolean isDefaultReplica(int replicaId) {
     return DEFAULT_REPLICA_ID == replicaId;
   }
 
-  /** @return true if this region is a default replica for the region */
+  /** Returns true if this region is a default replica for the region */
   public static boolean isDefaultReplica(RegionInfo hri) {
     return hri.getReplicaId() == DEFAULT_REPLICA_ID;
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionServerCallable.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionServerCallable.java
index 8bdc1543d74..abe4058b99d 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionServerCallable.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionServerCallable.java
@@ -155,9 +155,7 @@ public abstract class RegionServerCallable<T, S> implements RetryingCallable<T>
     }
   }
 
-  /**
-   * @return {@link ClusterConnection} instance used by this Callable.
-   */
+  /** Returns {@link ClusterConnection} instance used by this Callable. */
   protected ClusterConnection getConnection() {
     return (ClusterConnection) this.connection;
   }
@@ -200,9 +198,7 @@ public abstract class RegionServerCallable<T, S> implements RetryingCallable<T>
     return ConnectionUtils.getPauseTime(pause, tries);
   }
 
-  /**
-   * @return the HRegionInfo for the current region
-   */
+  /** Returns the HRegionInfo for the current region */
   public HRegionInfo getHRegionInfo() {
     if (this.location == null) {
       return null;
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RequestController.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RequestController.java
index 66d864be7d4..325286fd94c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RequestController.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RequestController.java
@@ -67,9 +67,7 @@ public interface RequestController {
     void reset() throws InterruptedIOException;
   }
 
-  /**
-   * @return A new checker for evaluating a batch rows.
-   */
+  /** Returns A new checker for evaluating a batch rows. */
   Checker newChecker();
 
   /**
@@ -86,9 +84,7 @@ public interface RequestController {
    */
   void decTaskCounters(Collection<byte[]> regions, ServerName sn);
 
-  /**
-   * @return The number of running task.
-   */
+  /** Returns The number of running task. */
   long getNumberOfTasksInProgress();
 
   /**
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Result.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Result.java
index 75f6cb95868..2ae13c6ce1c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Result.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Result.java
@@ -671,9 +671,7 @@ public class Result implements CellScannable, CellScanner {
     return this.cells == null || this.cells.length == 0;
   }
 
-  /**
-   * @return the size of the underlying Cell []
-   */
+  /** Returns the size of the underlying Cell [] */
   public int size() {
     return this.cells == null ? 0 : this.cells.length;
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ResultScanner.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ResultScanner.java
index 09f4cd92c7a..89b0a6d2efa 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ResultScanner.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ResultScanner.java
@@ -114,8 +114,6 @@ public interface ResultScanner extends Closeable, Iterable<Result> {
    */
   boolean renewLease();
 
-  /**
-   * @return the scan metrics, or {@code null} if we do not enable metrics.
-   */
+  /** Returns the scan metrics, or {@code null} if we do not enable metrics. */
   ScanMetrics getScanMetrics();
 }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Row.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Row.java
index ad1089f2aad..683304ff758 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Row.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Row.java
@@ -28,9 +28,7 @@ import org.apache.yetus.audience.InterfaceAudience;
 public interface Row extends Comparable<Row> {
   Comparator<Row> COMPARATOR = (v1, v2) -> Bytes.compareTo(v1.getRow(), v2.getRow());
 
-  /**
-   * @return The row.
-   */
+  /** Returns The row. */
   byte[] getRow();
 
   /**
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RowAccess.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RowAccess.java
index ba613bb1773..cd1f4756079 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RowAccess.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RowAccess.java
@@ -26,13 +26,9 @@ import org.apache.yetus.audience.InterfaceAudience;
  */
 @InterfaceAudience.Private
 public interface RowAccess<T> extends Iterable<T> {
-  /**
-   * @return true if there are no elements.
-   */
+  /** Returns true if there are no elements. */
   boolean isEmpty();
 
-  /**
-   * @return the number of elements in this list.
-   */
+  /** Returns the number of elements in this list. */
   int size();
 }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RowMutations.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RowMutations.java
index 6b2535989bc..1d9b94a54cc 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RowMutations.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RowMutations.java
@@ -155,9 +155,7 @@ public class RowMutations implements Row {
     return row;
   }
 
-  /**
-   * @return An unmodifiable list of the current mutations.
-   */
+  /** Returns An unmodifiable list of the current mutations. */
   public List<Mutation> getMutations() {
     return Collections.unmodifiableList(mutations);
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerImpl.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerImpl.java
index c36a02a885d..2c82cdd54bc 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerImpl.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerImpl.java
@@ -176,9 +176,7 @@ public class RpcRetryingCallerImpl<T> implements RpcRetryingCaller<T> {
     }
   }
 
-  /**
-   * @return Calculate how long a single call took
-   */
+  /** Returns Calculate how long a single call took */
   private long singleCallDuration(final long expectedSleep) {
     return (EnvironmentEdgeManager.currentTime() - tracker.getStartTime()) + expectedSleep;
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Scan.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Scan.java
index 677ddc0bb3e..d417d8e8453 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Scan.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Scan.java
@@ -687,9 +687,7 @@ public class Scan extends Query {
     return this;
   }
 
-  /**
-   * @return the maximum result size in bytes. See {@link #setMaxResultSize(long)}
-   */
+  /** Returns the maximum result size in bytes. See {@link #setMaxResultSize(long)} */
   public long getMaxResultSize() {
     return maxResultSize;
   }
@@ -727,9 +725,7 @@ public class Scan extends Query {
     return this.familyMap;
   }
 
-  /**
-   * @return the number of families in familyMap
-   */
+  /** Returns the number of families in familyMap */
   public int numFamilies() {
     if (hasFamilies()) {
       return this.familyMap.size();
@@ -737,16 +733,12 @@ public class Scan extends Query {
     return 0;
   }
 
-  /**
-   * @return true if familyMap is non empty, false otherwise
-   */
+  /** Returns true if familyMap is non empty, false otherwise */
   public boolean hasFamilies() {
     return !this.familyMap.isEmpty();
   }
 
-  /**
-   * @return the keys of the familyMap
-   */
+  /** Returns the keys of the familyMap */
   public byte[][] getFamilies() {
     if (hasFamilies()) {
       return this.familyMap.keySet().toArray(new byte[0][0]);
@@ -754,51 +746,37 @@ public class Scan extends Query {
     return null;
   }
 
-  /**
-   * @return the startrow
-   */
+  /** Returns the startrow */
   public byte[] getStartRow() {
     return this.startRow;
   }
 
-  /**
-   * @return if we should include start row when scan
-   */
+  /** Returns if we should include start row when scan */
   public boolean includeStartRow() {
     return includeStartRow;
   }
 
-  /**
-   * @return the stoprow
-   */
+  /** Returns the stoprow */
   public byte[] getStopRow() {
     return this.stopRow;
   }
 
-  /**
-   * @return if we should include stop row when scan
-   */
+  /** Returns if we should include stop row when scan */
   public boolean includeStopRow() {
     return includeStopRow;
   }
 
-  /**
-   * @return the max number of versions to fetch
-   */
+  /** Returns the max number of versions to fetch */
   public int getMaxVersions() {
     return this.maxVersions;
   }
 
-  /**
-   * @return maximum number of values to return for a single call to next()
-   */
+  /** Returns maximum number of values to return for a single call to next() */
   public int getBatch() {
     return this.batch;
   }
 
-  /**
-   * @return maximum number of values to return per row per CF
-   */
+  /** Returns maximum number of values to return per row per CF */
   public int getMaxResultsPerColumnFamily() {
     return this.storeLimit;
   }
@@ -811,9 +789,7 @@ public class Scan extends Query {
     return this.storeOffset;
   }
 
-  /**
-   * @return caching the number of rows fetched when calling next on a scanner
-   */
+  /** Returns caching the number of rows fetched when calling next on a scanner */
   public int getCaching() {
     return this.caching;
   }
@@ -833,9 +809,7 @@ public class Scan extends Query {
     return filter;
   }
 
-  /**
-   * @return true is a filter has been specified, false if not
-   */
+  /** Returns true is a filter has been specified, false if not */
   public boolean hasFilter() {
     return filter != null;
   }
@@ -996,9 +970,7 @@ public class Scan extends Query {
     return this;
   }
 
-  /**
-   * @return True if this Scan is in "raw" mode.
-   */
+  /** Returns True if this Scan is in "raw" mode. */
   public boolean isRaw() {
     byte[] attr = getAttribute(RAW_ATTR);
     return attr == null ? false : Bytes.toBoolean(attr);
@@ -1098,9 +1070,7 @@ public class Scan extends Query {
     return this;
   }
 
-  /**
-   * @return True if collection of scan metrics is enabled. For advanced users.
-   */
+  /** Returns True if collection of scan metrics is enabled. For advanced users. */
   public boolean isScanMetricsEnabled() {
     byte[] attr = getAttribute(Scan.SCAN_ATTRIBUTES_METRICS_ENABLE);
     return attr == null ? false : Bytes.toBoolean(attr);
@@ -1129,9 +1099,7 @@ public class Scan extends Query {
     return this;
   }
 
-  /**
-   * @return the limit of rows for this scan
-   */
+  /** Returns the limit of rows for this scan */
   public int getLimit() {
     return limit;
   }
@@ -1163,9 +1131,7 @@ public class Scan extends Query {
     PREAD
   }
 
-  /**
-   * @return the read type for this scan
-   */
+  /** Returns the read type for this scan */
   public ReadType getReadType() {
     return readType;
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ScannerCallable.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ScannerCallable.java
index 06130f5a889..9c4f6649b19 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ScannerCallable.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ScannerCallable.java
@@ -396,9 +396,7 @@ public class ScannerCallable extends ClientServiceCallable<Result[]> {
     this.renew = val;
   }
 
-  /**
-   * @return the HRegionInfo for the current region
-   */
+  /** Returns the HRegionInfo for the current region */
   @Override
   public HRegionInfo getHRegionInfo() {
     if (!instantiated) {
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableDescriptor.java
index 85f412e2df2..ff97f419bac 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableDescriptor.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableDescriptor.java
@@ -153,9 +153,7 @@ public interface TableDescriptor {
   // those tables with the highest priority (From Yi Liang over on HBASE-18109).
   int getPriority();
 
-  /**
-   * @return Returns the configured replicas per region
-   */
+  /** Returns Returns the configured replicas per region */
   int getRegionReplication();
 
   /**
@@ -200,9 +198,7 @@ public interface TableDescriptor {
    */
   String getValue(String key);
 
-  /**
-   * @return Getter for fetching an unmodifiable map.
-   */
+  /** Returns Getter for fetching an unmodifiable map. */
   Map<Bytes, Bytes> getValues();
 
   /**
@@ -219,9 +215,7 @@ public interface TableDescriptor {
    */
   boolean hasColumnFamily(final byte[] name);
 
-  /**
-   * @return true if the read-replicas memstore replication is enabled.
-   */
+  /** Returns true if the read-replicas memstore replication is enabled. */
   boolean hasRegionMemStoreReplication();
 
   /**
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableDescriptorBuilder.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableDescriptorBuilder.java
index a005a752527..1d2b69c7146 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableDescriptorBuilder.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableDescriptorBuilder.java
@@ -1110,9 +1110,7 @@ public class TableDescriptorBuilder {
       return families.containsKey(familyName);
     }
 
-    /**
-     * @return Name of this table and then a map of all of the column family descriptors.
-     */
+    /** Returns Name of this table and then a map of all of the column family descriptors. */
     @Override
     public String toString() {
       StringBuilder s = new StringBuilder();
@@ -1135,9 +1133,7 @@ public class TableDescriptorBuilder {
       return s.toString();
     }
 
-    /**
-     * @return map of all table attributes formatted into string.
-     */
+    /** Returns map of all table attributes formatted into string. */
     public String toStringTableAttributes() {
       return getValues(true).toString();
     }
@@ -1247,9 +1243,7 @@ public class TableDescriptorBuilder {
       return false;
     }
 
-    /**
-     * @return hash code
-     */
+    /** Returns hash code */
     @Override
     public int hashCode() {
       int result = this.name.hashCode();
@@ -1297,9 +1291,7 @@ public class TableDescriptorBuilder {
       return setValue(REGION_REPLICATION_KEY, Integer.toString(regionReplication));
     }
 
-    /**
-     * @return true if the read-replicas memstore replication is enabled.
-     */
+    /** Returns true if the read-replicas memstore replication is enabled. */
     @Override
     public boolean hasRegionMemStoreReplication() {
       return getOrDefault(REGION_MEMSTORE_REPLICATION_KEY, Boolean::valueOf,
@@ -1550,9 +1542,7 @@ public class TableDescriptorBuilder {
       return getOrDefault(OWNER_KEY, Function.identity(), null);
     }
 
-    /**
-     * @return the bytes in pb format
-     */
+    /** Returns the bytes in pb format */
     private byte[] toByteArray() {
       return ProtobufUtil.prependPBMagic(ProtobufUtil.toTableSchema(this).toByteArray());
     }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableState.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableState.java
index ffd8cf8409d..6ecd97a75c9 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableState.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableState.java
@@ -94,44 +94,32 @@ public class TableState {
   private final TableName tableName;
   private final State state;
 
-  /**
-   * @return True if table is {@link State#ENABLED}.
-   */
+  /** Returns True if table is {@link State#ENABLED}. */
   public boolean isEnabled() {
     return isInStates(State.ENABLED);
   }
 
-  /**
-   * @return True if table is {@link State#ENABLING}.
-   */
+  /** Returns True if table is {@link State#ENABLING}. */
   public boolean isEnabling() {
     return isInStates(State.ENABLING);
   }
 
-  /**
-   * @return True if {@link State#ENABLED} or {@link State#ENABLING}
-   */
+  /** Returns True if {@link State#ENABLED} or {@link State#ENABLING} */
   public boolean isEnabledOrEnabling() {
     return isInStates(State.ENABLED, State.ENABLING);
   }
 
-  /**
-   * @return True if table is disabled.
-   */
+  /** Returns True if table is disabled. */
   public boolean isDisabled() {
     return isInStates(State.DISABLED);
   }
 
-  /**
-   * @return True if table is disabling.
-   */
+  /** Returns True if table is disabling. */
   public boolean isDisabling() {
     return isInStates(State.DISABLING);
   }
 
-  /**
-   * @return True if {@link State#DISABLED} or {@link State#DISABLED}
-   */
+  /** Returns True if {@link State#DISABLED} or {@link State#DISABLED} */
   public boolean isDisabledOrDisabling() {
     return isInStates(State.DISABLED, State.DISABLING);
   }
@@ -146,9 +134,7 @@ public class TableState {
     this.state = state;
   }
 
-  /**
-   * @return table state
-   */
+  /** Returns table state */
   public State getState() {
     return state;
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/backoff/ClientBackoffPolicy.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/backoff/ClientBackoffPolicy.java
index ee66bcada1b..1799c27ce2c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/backoff/ClientBackoffPolicy.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/backoff/ClientBackoffPolicy.java
@@ -32,8 +32,6 @@ public interface ClientBackoffPolicy {
 
   public static final String BACKOFF_POLICY_CLASS = "hbase.client.statistics.backoff-policy";
 
-  /**
-   * @return the number of ms to wait on the client based on the
-   */
+  /** Returns the number of ms to wait on the client based on the */
   public long getBackoffTime(ServerName serverName, byte[] region, ServerStatistics stats);
 }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/exceptions/PreemptiveFastFailException.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/exceptions/PreemptiveFastFailException.java
index 6d28f3288fd..356e90b2992 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/exceptions/PreemptiveFastFailException.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/exceptions/PreemptiveFastFailException.java
@@ -67,37 +67,27 @@ public class PreemptiveFastFailException extends ConnectException {
     this.guaranteedClientSideOnly = guaranteedClientSideOnly;
   }
 
-  /**
-   * @return time of the fist failure
-   */
+  /** Returns time of the fist failure */
   public long getFirstFailureAt() {
     return timeOfFirstFailureMilliSec;
   }
 
-  /**
-   * @return time of the latest attempt
-   */
+  /** Returns time of the latest attempt */
   public long getLastAttemptAt() {
     return timeOfLatestAttemptMilliSec;
   }
 
-  /**
-   * @return failure count
-   */
+  /** Returns failure count */
   public long getFailureCount() {
     return failureCount;
   }
 
-  /**
-   * @return true if operation was attempted by server, false otherwise.
-   */
+  /** Returns true if operation was attempted by server, false otherwise. */
   public boolean wasOperationAttemptedByServer() {
     return false;
   }
 
-  /**
-   * @return true if we know no mutation made it to the server, false otherwise.
-   */
+  /** Returns true if we know no mutation made it to the server, false otherwise. */
   public boolean isGuaranteedClientSideOnly() {
     return guaranteedClientSideOnly;
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BigDecimalComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BigDecimalComparator.java
index bfd285975ff..669a591a961 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BigDecimalComparator.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BigDecimalComparator.java
@@ -72,9 +72,7 @@ public class BigDecimalComparator extends ByteArrayComparable {
     return this.bigDecimal.compareTo(that);
   }
 
-  /**
-   * @return The comparator serialized using pb
-   */
+  /** Returns The comparator serialized using pb */
   @Override
   public byte[] toByteArray() {
     ComparatorProtos.BigDecimalComparator.Builder builder =
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryComparator.java
index 0c8274a8611..7aff979f4c2 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryComparator.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryComparator.java
@@ -54,9 +54,7 @@ public class BinaryComparator extends org.apache.hadoop.hbase.filter.ByteArrayCo
     return ByteBufferUtils.compareTo(this.value, 0, this.value.length, value, offset, length);
   }
 
-  /**
-   * @return The comparator serialized using pb
-   */
+  /** Returns The comparator serialized using pb */
   @Override
   public byte[] toByteArray() {
     ComparatorProtos.BinaryComparator.Builder builder =
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryComponentComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryComponentComparator.java
index c479d757d1c..88622ed2eba 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryComponentComparator.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryComponentComparator.java
@@ -73,9 +73,7 @@ public class BinaryComponentComparator extends ByteArrayComparable {
     return result;
   }
 
-  /**
-   * @return The comparator serialized using pb
-   */
+  /** Returns The comparator serialized using pb */
   @Override
   public byte[] toByteArray() {
     ComparatorProtos.BinaryComponentComparator.Builder builder =
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryPrefixComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryPrefixComparator.java
index f97fd070be6..fcf447dcb0c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryPrefixComparator.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryPrefixComparator.java
@@ -58,9 +58,7 @@ public class BinaryPrefixComparator extends ByteArrayComparable {
     return ByteBufferUtils.compareTo(this.value, 0, this.value.length, value, offset, length);
   }
 
-  /**
-   * @return The comparator serialized using pb
-   */
+  /** Returns The comparator serialized using pb */
   @Override
   public byte[] toByteArray() {
     ComparatorProtos.BinaryPrefixComparator.Builder builder =
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java
index 15ca8890aba..8abdcf4e8a6 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java
@@ -57,16 +57,12 @@ public class BitComparator extends ByteArrayComparable {
     this.bitOperator = bitOperator;
   }
 
-  /**
-   * @return the bitwise operator
-   */
+  /** Returns the bitwise operator */
   public BitwiseOp getOperator() {
     return bitOperator;
   }
 
-  /**
-   * @return The comparator serialized using pb
-   */
+  /** Returns The comparator serialized using pb */
   @Override
   public byte[] toByteArray() {
     ComparatorProtos.BitComparator.Builder builder = ComparatorProtos.BitComparator.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java
index f1e5644fd8a..a7cd0489ca5 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java
@@ -84,9 +84,7 @@ public class ColumnCountGetFilter extends FilterBase {
     return new ColumnCountGetFilter(limit);
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() {
     FilterProtos.ColumnCountGetFilter.Builder builder =
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java
index f11005e43d2..e9cf7dce76d 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java
@@ -157,9 +157,7 @@ public class ColumnPaginationFilter extends FilterBase {
     return new ColumnPaginationFilter(limit, offset);
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() {
     FilterProtos.ColumnPaginationFilter.Builder builder =
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java
index 86401564ed5..de76b8c5059 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java
@@ -108,9 +108,7 @@ public class ColumnPrefixFilter extends FilterBase {
     return new ColumnPrefixFilter(columnPrefix);
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() {
     FilterProtos.ColumnPrefixFilter.Builder builder = FilterProtos.ColumnPrefixFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java
index dca4609763c..1abd49fff3c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java
@@ -65,44 +65,32 @@ public class ColumnRangeFilter extends FilterBase {
     this.maxColumnInclusive = maxColumnInclusive;
   }
 
-  /**
-   * @return if min column range is inclusive.
-   */
+  /** Returns if min column range is inclusive. */
   public boolean isMinColumnInclusive() {
     return minColumnInclusive;
   }
 
-  /**
-   * @return if max column range is inclusive.
-   */
+  /** Returns if max column range is inclusive. */
   public boolean isMaxColumnInclusive() {
     return maxColumnInclusive;
   }
 
-  /**
-   * @return the min column range for the filter
-   */
+  /** Returns the min column range for the filter */
   public byte[] getMinColumn() {
     return this.minColumn;
   }
 
-  /**
-   * @return true if min column is inclusive, false otherwise
-   */
+  /** Returns true if min column is inclusive, false otherwise */
   public boolean getMinColumnInclusive() {
     return this.minColumnInclusive;
   }
 
-  /**
-   * @return the max column range for the filter
-   */
+  /** Returns the max column range for the filter */
   public byte[] getMaxColumn() {
     return this.maxColumn;
   }
 
-  /**
-   * @return true if max column is inclusive, false otherwise
-   */
+  /** Returns true if max column is inclusive, false otherwise */
   public boolean getMaxColumnInclusive() {
     return this.maxColumnInclusive;
   }
@@ -161,9 +149,7 @@ public class ColumnRangeFilter extends FilterBase {
     return new ColumnRangeFilter(minColumn, minColumnInclusive, maxColumn, maxColumnInclusive);
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() {
     FilterProtos.ColumnRangeFilter.Builder builder = FilterProtos.ColumnRangeFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnValueFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnValueFilter.java
index 0074fe40a3a..0b3b85dbb47 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnValueFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnValueFilter.java
@@ -75,23 +75,17 @@ public class ColumnValueFilter extends FilterBase {
     return op;
   }
 
-  /**
-   * @return the comparator
-   */
+  /** Returns the comparator */
   public ByteArrayComparable getComparator() {
     return comparator;
   }
 
-  /**
-   * @return the column family
-   */
+  /** Returns the column family */
   public byte[] getFamily() {
     return family;
   }
 
-  /**
-   * @return the qualifier
-   */
+  /** Returns the qualifier */
   public byte[] getQualifier() {
     return qualifier;
   }
@@ -161,9 +155,7 @@ public class ColumnValueFilter extends FilterBase {
     return new ColumnValueFilter(family, qualifier, operator, comparator);
   }
 
-  /**
-   * @return A pb instance to represent this instance.
-   */
+  /** Returns A pb instance to represent this instance. */
   FilterProtos.ColumnValueFilter convert() {
     FilterProtos.ColumnValueFilter.Builder builder = FilterProtos.ColumnValueFilter.newBuilder();
 
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/CompareFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/CompareFilter.java
index 7fc8be99396..5b25e3b2ae1 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/CompareFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/CompareFilter.java
@@ -113,9 +113,7 @@ public abstract class CompareFilter extends FilterBase {
     return op;
   }
 
-  /**
-   * @return the comparator
-   */
+  /** Returns the comparator */
   public ByteArrayComparable getComparator() {
     return comparator;
   }
@@ -279,9 +277,7 @@ public abstract class CompareFilter extends FilterBase {
     return arguments;
   }
 
-  /**
-   * @return A pb instance to represent this instance.
-   */
+  /** Returns A pb instance to represent this instance. */
   FilterProtos.CompareFilter convert() {
     FilterProtos.CompareFilter.Builder builder = FilterProtos.CompareFilter.newBuilder();
     HBaseProtos.CompareType compareOp = CompareType.valueOf(this.op.name());
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java
index c9a688028ca..0a39f4c94f2 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java
@@ -112,23 +112,17 @@ public class DependentColumnFilter extends CompareFilter {
     this(family, qualifier, dropDependentColumn, CompareOp.NO_OP, null);
   }
 
-  /**
-   * @return the column family
-   */
+  /** Returns the column family */
   public byte[] getFamily() {
     return this.columnFamily;
   }
 
-  /**
-   * @return the column qualifier
-   */
+  /** Returns the column qualifier */
   public byte[] getQualifier() {
     return this.columnQualifier;
   }
 
-  /**
-   * @return true if we should drop the dependent column, false otherwise
-   */
+  /** Returns true if we should drop the dependent column, false otherwise */
   public boolean dropDependentColumn() {
     return this.dropDependentColumn;
   }
@@ -219,9 +213,7 @@ public class DependentColumnFilter extends CompareFilter {
     }
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() {
     FilterProtos.DependentColumnFilter.Builder builder =
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java
index 489b1363fd7..917ac522de1 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java
@@ -92,9 +92,7 @@ public class FamilyFilter extends CompareFilter {
     return new FamilyFilter(compareOp, comparator);
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() {
     FilterProtos.FamilyFilter.Builder builder = FilterProtos.FamilyFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java
index 6daf97e931e..35313837843 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java
@@ -192,9 +192,7 @@ final public class FilterList extends FilterBase {
     return filterListBase.filterRow();
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() throws IOException {
     FilterProtos.FilterList.Builder builder = FilterProtos.FilterList.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.java
index fd6eddf005e..d779d36eef7 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.java
@@ -71,9 +71,7 @@ public class FirstKeyOnlyFilter extends FilterBase {
     return new FirstKeyOnlyFilter();
   }
 
-  /**
-   * @return true if first KV has been found.
-   */
+  /** Returns true if first KV has been found. */
   protected boolean hasFoundKV() {
     return this.foundKV;
   }
@@ -85,9 +83,7 @@ public class FirstKeyOnlyFilter extends FilterBase {
     this.foundKV = value;
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() {
     FilterProtos.FirstKeyOnlyFilter.Builder builder = FilterProtos.FirstKeyOnlyFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyValueMatchingQualifiersFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyValueMatchingQualifiersFilter.java
index d10a6ea8e90..b49e51f348c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyValueMatchingQualifiersFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyValueMatchingQualifiersFilter.java
@@ -83,9 +83,7 @@ public class FirstKeyValueMatchingQualifiersFilter extends FirstKeyOnlyFilter {
     return false;
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() {
     FilterProtos.FirstKeyValueMatchingQualifiersFilter.Builder builder =
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java
index 5f3e2b3da73..6036aebdd57 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java
@@ -284,9 +284,7 @@ public class FuzzyRowFilter extends FilterBase {
     return done;
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() {
     FilterProtos.FuzzyRowFilter.Builder builder = FilterProtos.FuzzyRowFilter.newBuilder()
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java
index b8d43cc5c39..67a214e5650 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java
@@ -81,9 +81,7 @@ public class InclusiveStopFilter extends FilterBase {
     return new InclusiveStopFilter(stopRowKey);
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() {
     FilterProtos.InclusiveStopFilter.Builder builder =
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/KeyOnlyFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/KeyOnlyFilter.java
index 6332d404ad1..60b3a5c0480 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/KeyOnlyFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/KeyOnlyFilter.java
@@ -99,9 +99,7 @@ public class KeyOnlyFilter extends FilterBase {
     return filter;
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() {
     FilterProtos.KeyOnlyFilter.Builder builder = FilterProtos.KeyOnlyFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/LongComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/LongComparator.java
index ead0ee10447..8f9a7f1dfc3 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/LongComparator.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/LongComparator.java
@@ -53,9 +53,7 @@ public class LongComparator extends ByteArrayComparable {
     return Long.compare(longValue, that);
   }
 
-  /**
-   * @return The comparator serialized using pb
-   */
+  /** Returns The comparator serialized using pb */
   @Override
   public byte[] toByteArray() {
     ComparatorProtos.LongComparator.Builder builder = ComparatorProtos.LongComparator.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java
index 55026e8f1bf..af07cc42b11 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java
@@ -182,9 +182,7 @@ public class MultiRowRangeFilter extends FilterBase {
     return PrivateCellUtil.createFirstOnRow(comparisonData, 0, (short) comparisonData.length);
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() {
     FilterProtos.MultiRowRangeFilter.Builder builder =
@@ -475,16 +473,12 @@ public class MultiRowRangeFilter extends FilterBase {
       return stopRow;
     }
 
-    /**
-     * @return if start row is inclusive.
-     */
+    /** Returns if start row is inclusive. */
     public boolean isStartRowInclusive() {
       return startRowInclusive;
     }
 
-    /**
-     * @return if stop row is inclusive.
-     */
+    /** Returns if stop row is inclusive. */
     public boolean isStopRowInclusive() {
       return stopRowInclusive;
     }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java
index 1744c20968e..e7c868e3bbb 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java
@@ -122,9 +122,7 @@ public class MultipleColumnPrefixFilter extends FilterBase {
     return new MultipleColumnPrefixFilter(prefixes);
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() {
     FilterProtos.MultipleColumnPrefixFilter.Builder builder =
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/NullComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/NullComparator.java
index fc0562ecb3e..6153e1e8080 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/NullComparator.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/NullComparator.java
@@ -63,9 +63,7 @@ public class NullComparator extends ByteArrayComparable {
     return value != null ? 1 : 0;
   }
 
-  /**
-   * @return The comparator serialized using pb
-   */
+  /** Returns The comparator serialized using pb */
   @Override
   public byte[] toByteArray() {
     ComparatorProtos.NullComparator.Builder builder = ComparatorProtos.NullComparator.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java
index 603778cac57..45c8a1f1a3f 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java
@@ -97,9 +97,7 @@ public class PageFilter extends FilterBase {
     return new PageFilter(pageSize);
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() {
     FilterProtos.PageFilter.Builder builder = FilterProtos.PageFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java
index c6fdcc88875..51752d2b72f 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java
@@ -107,9 +107,7 @@ public class PrefixFilter extends FilterBase {
     return new PrefixFilter(prefix);
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() {
     FilterProtos.PrefixFilter.Builder builder = FilterProtos.PrefixFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/QualifierFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/QualifierFilter.java
index 6ee70a79e9f..7f0b8642003 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/QualifierFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/QualifierFilter.java
@@ -86,9 +86,7 @@ public class QualifierFilter extends CompareFilter {
     return new QualifierFilter(compareOp, comparator);
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() {
     FilterProtos.QualifierFilter.Builder builder = FilterProtos.QualifierFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RandomRowFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RandomRowFilter.java
index ff73dadb9d8..656ceaf3e1a 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RandomRowFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RandomRowFilter.java
@@ -44,9 +44,7 @@ public class RandomRowFilter extends FilterBase {
     this.chance = chance;
   }
 
-  /**
-   * @return The chance that a row gets included.
-   */
+  /** Returns The chance that a row gets included. */
   public float getChance() {
     return chance;
   }
@@ -107,9 +105,7 @@ public class RandomRowFilter extends FilterBase {
     filterOutRow = false;
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() {
     FilterProtos.RandomRowFilter.Builder builder = FilterProtos.RandomRowFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RegexStringComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RegexStringComparator.java
index 75272c5f241..6841c80194a 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RegexStringComparator.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RegexStringComparator.java
@@ -145,9 +145,7 @@ public class RegexStringComparator extends ByteArrayComparable {
     return engine.compareTo(value, offset, length);
   }
 
-  /**
-   * @return The comparator serialized using pb
-   */
+  /** Returns The comparator serialized using pb */
   @Override
   public byte[] toByteArray() {
     return engine.toByteArray();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java
index 6c51e788091..fdefe382439 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java
@@ -106,9 +106,7 @@ public class RowFilter extends CompareFilter {
     return new RowFilter(compareOp, comparator);
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() {
     FilterProtos.RowFilter.Builder builder = FilterProtos.RowFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueExcludeFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueExcludeFilter.java
index a410884a8ab..bbd8ec67994 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueExcludeFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueExcludeFilter.java
@@ -163,9 +163,7 @@ public class SingleColumnValueExcludeFilter extends SingleColumnValueFilter {
     return filter;
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() {
     FilterProtos.SingleColumnValueExcludeFilter.Builder builder =
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java
index b049867d2ef..a68fdcd0577 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java
@@ -192,23 +192,17 @@ public class SingleColumnValueFilter extends FilterBase {
     return op;
   }
 
-  /**
-   * @return the comparator
-   */
+  /** Returns the comparator */
   public org.apache.hadoop.hbase.filter.ByteArrayComparable getComparator() {
     return comparator;
   }
 
-  /**
-   * @return the family
-   */
+  /** Returns the family */
   public byte[] getFamily() {
     return columnFamily;
   }
 
-  /**
-   * @return the qualifier
-   */
+  /** Returns the qualifier */
   public byte[] getQualifier() {
     return columnQualifier;
   }
@@ -356,9 +350,7 @@ public class SingleColumnValueFilter extends FilterBase {
     return builder.build();
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() {
     return convert().toByteArray();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java
index e3b334b8820..4b23a87537d 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java
@@ -102,9 +102,7 @@ public class SkipFilter extends FilterBase {
     return true;
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() throws IOException {
     FilterProtos.SkipFilter.Builder builder = FilterProtos.SkipFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SubstringComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SubstringComparator.java
index b8e33c438fe..b5fe22dca24 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SubstringComparator.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SubstringComparator.java
@@ -66,9 +66,7 @@ public class SubstringComparator extends ByteArrayComparable {
     return Bytes.toString(value, offset, length).toLowerCase(Locale.ROOT).contains(substr) ? 0 : 1;
   }
 
-  /**
-   * @return The comparator serialized using pb
-   */
+  /** Returns The comparator serialized using pb */
   @Override
   public byte[] toByteArray() {
     ComparatorProtos.SubstringComparator.Builder builder =
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/TimestampsFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/TimestampsFilter.java
index 2ed44ba77e4..eaeaaac3df1 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/TimestampsFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/TimestampsFilter.java
@@ -77,9 +77,7 @@ public class TimestampsFilter extends FilterBase {
     init();
   }
 
-  /**
-   * @return the list of timestamps
-   */
+  /** Returns the list of timestamps */
   public List<Long> getTimestamps() {
     List<Long> list = new ArrayList<>(timestamps.size());
     list.addAll(timestamps);
@@ -165,9 +163,7 @@ public class TimestampsFilter extends FilterBase {
     return new TimestampsFilter(timestamps);
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() {
     FilterProtos.TimestampsFilter.Builder builder = FilterProtos.TimestampsFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ValueFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ValueFilter.java
index 57a1c5c7c2a..e726d88f720 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ValueFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ValueFilter.java
@@ -87,9 +87,7 @@ public class ValueFilter extends CompareFilter {
     return new ValueFilter(compareOp, comparator);
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() {
     FilterProtos.ValueFilter.Builder builder = FilterProtos.ValueFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/WhileMatchFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/WhileMatchFilter.java
index c58e080b565..e31cbb97378 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/WhileMatchFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/WhileMatchFilter.java
@@ -107,9 +107,7 @@ public class WhileMatchFilter extends FilterBase {
     return true;
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() throws IOException {
     FilterProtos.WhileMatchFilter.Builder builder = FilterProtos.WhileMatchFilter.newBuilder();
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRpcController.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRpcController.java
index b33771e5b58..01412631ef7 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRpcController.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRpcController.java
@@ -60,9 +60,7 @@ public interface HBaseRpcController extends RpcController, CellScannable {
    */
   void setPriority(final TableName tn);
 
-  /**
-   * @return The priority of this request
-   */
+  /** Returns The priority of this request */
   int getPriority();
 
   int getCallTimeout();
@@ -110,16 +108,12 @@ public interface HBaseRpcController extends RpcController, CellScannable {
    */
   void notifyOnCancel(RpcCallback<Object> callback, CancellationCallback action) throws IOException;
 
-  /**
-   * @return True if this Controller is carrying the RPC target Region's RegionInfo.
-   */
+  /** Returns True if this Controller is carrying the RPC target Region's RegionInfo. */
   default boolean hasRegionInfo() {
     return false;
   }
 
-  /**
-   * @return Target Region's RegionInfo or null if not available or pertinent.
-   */
+  /** Returns Target Region's RegionInfo or null if not available or pertinent. */
   default RegionInfo getRegionInfo() {
     return null;
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRpcControllerImpl.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRpcControllerImpl.java
index a4c51ae62ff..f3e19aeadae 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRpcControllerImpl.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRpcControllerImpl.java
@@ -101,9 +101,7 @@ public class HBaseRpcControllerImpl implements HBaseRpcController {
     return this.regionInfo;
   }
 
-  /**
-   * @return One-shot cell scanner (you cannot back it up and restart)
-   */
+  /** Returns One-shot cell scanner (you cannot back it up and restart) */
   @Override
   public CellScanner cellScanner() {
     return cellScanner;
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/IPCUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/IPCUtil.java
index 57f8da98eff..b509dcbd27b 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/IPCUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/IPCUtil.java
@@ -96,9 +96,7 @@ class IPCUtil {
     return totalSize;
   }
 
-  /**
-   * @return Size on the wire when the two messages are written with writeDelimitedTo
-   */
+  /** Returns Size on the wire when the two messages are written with writeDelimitedTo */
   public static int getTotalSizeWhenWrittenDelimited(Message... messages) {
     int totalSize = 0;
     for (Message m : messages) {
@@ -149,9 +147,7 @@ class IPCUtil {
         serverOverloaded);
   }
 
-  /**
-   * @return True if the exception is a fatal connection exception.
-   */
+  /** Returns True if the exception is a fatal connection exception. */
   static boolean isFatalConnectionException(final ExceptionResponse e) {
     return e.getExceptionClassName().equals(FatalConnectionException.class.getName());
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RemoteWithExtrasException.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RemoteWithExtrasException.java
index 62d0bb1d455..a25708fe58e 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RemoteWithExtrasException.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RemoteWithExtrasException.java
@@ -123,30 +123,22 @@ public class RemoteWithExtrasException extends RemoteException {
     return ex;
   }
 
-  /**
-   * @return null if not set
-   */
+  /** Returns null if not set */
   public String getHostname() {
     return this.hostname;
   }
 
-  /**
-   * @return -1 if not set
-   */
+  /** Returns -1 if not set */
   public int getPort() {
     return this.port;
   }
 
-  /**
-   * @return True if origin exception was a do not retry type.
-   */
+  /** Returns True if origin exception was a do not retry type. */
   public boolean isDoNotRetry() {
     return this.doNotRetry;
   }
 
-  /**
-   * @return True if the server was considered overloaded when the exception was thrown.
-   */
+  /** Returns True if the server was considered overloaded when the exception was thrown. */
   public boolean isServerOverloaded() {
     return serverOverloaded;
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufMagic.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufMagic.java
index 7c6f780e069..15a196cbe5b 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufMagic.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufMagic.java
@@ -77,9 +77,7 @@ public class ProtobufMagic {
     return compareTo(PB_MAGIC, 0, PB_MAGIC.length, bytes, offset, PB_MAGIC.length) == 0;
   }
 
-  /**
-   * @return Length of {@link #PB_MAGIC}
-   */
+  /** Returns Length of {@link #PB_MAGIC} */
   public static int lengthOfPBMagic() {
     return PB_MAGIC.length;
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java
index a7bb4d0e01b..76ce56318ed 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java
@@ -223,9 +223,7 @@ public final class ProtobufUtil {
     }
   }
 
-  /**
-   * @return Length of {@link ProtobufMagic#lengthOfPBMagic()}
-   */
+  /** Returns Length of {@link ProtobufMagic#lengthOfPBMagic()} */
   public static int lengthOfPBMagic() {
     return ProtobufMagic.lengthOfPBMagic();
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/QuotaFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/QuotaFilter.java
index 9ddd408f845..9dd44c56d43 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/QuotaFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/QuotaFilter.java
@@ -92,32 +92,32 @@ public class QuotaFilter {
     return this;
   }
 
-  /** @return true if the filter is empty */
+  /** Returns true if the filter is empty */
   public boolean isNull() {
     return !hasFilters;
   }
 
-  /** @return the QuotaType types that we want to filter one */
+  /** Returns the QuotaType types that we want to filter one */
   public Set<QuotaType> getTypeFilters() {
     return types;
   }
 
-  /** @return the Namespace filter regex */
+  /** Returns the Namespace filter regex */
   public String getNamespaceFilter() {
     return namespaceRegex;
   }
 
-  /** @return the Table filter regex */
+  /** Returns the Table filter regex */
   public String getTableFilter() {
     return tableRegex;
   }
 
-  /** @return the User filter regex */
+  /** Returns the User filter regex */
   public String getUserFilter() {
     return userRegex;
   }
 
-  /** @return the RegionServer filter regex */
+  /** Returns the RegionServer filter regex */
   public String getRegionServerFilter() {
     return regionServerRegex;
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/SpaceQuotaSnapshot.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/SpaceQuotaSnapshot.java
index 4a64a170152..0dd4fd3b7d2 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/SpaceQuotaSnapshot.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/SpaceQuotaSnapshot.java
@@ -74,9 +74,7 @@ public class SpaceQuotaSnapshot implements SpaceQuotaSnapshotView {
       return policy;
     }
 
-    /**
-     * @return {@code true} if the quota is being violated, {@code false} otherwise.
-     */
+    /** Returns {@code true} if the quota is being violated, {@code false} otherwise. */
     @Override
     public boolean isInViolation() {
       return inViolation;
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/SpaceQuotaSnapshotView.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/SpaceQuotaSnapshotView.java
index 4bc47b9cddd..f7acff7a283 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/SpaceQuotaSnapshotView.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/SpaceQuotaSnapshotView.java
@@ -39,9 +39,7 @@ public interface SpaceQuotaSnapshotView {
      */
     Optional<SpaceViolationPolicy> getPolicy();
 
-    /**
-     * @return {@code true} if the quota is being violated, {@code false} otherwise.
-     */
+    /** Returns {@code true} if the quota is being violated, {@code false} otherwise. */
     boolean isInViolation();
   }
 
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/Permission.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/Permission.java
index b3ac386689e..ace6f302a19 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/Permission.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/Permission.java
@@ -203,7 +203,7 @@ public class Permission extends VersionedWritable {
     return raw.toString();
   }
 
-  /** @return the object version number */
+  /** Returns the object version number */
   @Override
   public byte getVersion() {
     return VERSION;
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/CellVisibility.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/CellVisibility.java
index 8abaee00509..05b1f6c596d 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/CellVisibility.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/CellVisibility.java
@@ -35,9 +35,7 @@ public class CellVisibility {
     this.expression = expression;
   }
 
-  /**
-   * @return The visibility expression
-   */
+  /** Returns The visibility expression */
   public String getExpression() {
     return this.expression;
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java
index cae868a01a9..cb063938848 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java
@@ -334,9 +334,7 @@ public final class ProtobufUtil {
     }
   }
 
-  /**
-   * @return Length of {@link ProtobufMagic#lengthOfPBMagic()}
-   */
+  /** Returns Length of {@link ProtobufMagic#lengthOfPBMagic()} */
   public static int lengthOfPBMagic() {
     return ProtobufMagic.lengthOfPBMagic();
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java
index 09f7895b977..0ca16009d6c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java
@@ -841,9 +841,7 @@ public final class RequestConverter {
       ProtobufUtil.toMutationNoData(MutationType.APPEND, append, mutationBuilder, nonce)));
   }
 
-  /**
-   * @return whether or not the rowMutations has a Increment or Append
-   */
+  /** Returns whether or not the rowMutations has a Increment or Append */
   private static boolean buildNoDataRegionAction(final RowMutations rowMutations,
     final List<CellScannable> cells, long nonce, final RegionAction.Builder regionActionBuilder,
     final ClientProtos.Action.Builder actionBuilder, final MutationProto.Builder mutationBuilder)
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/HBaseSnapshotException.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/HBaseSnapshotException.java
index 5fd62a86008..65481f7c9f6 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/HBaseSnapshotException.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/HBaseSnapshotException.java
@@ -69,9 +69,7 @@ public class HBaseSnapshotException extends DoNotRetryIOException {
     super(message, cause);
   }
 
-  /**
-   * @return the description of the snapshot that is being failed
-   */
+  /** Returns the description of the snapshot that is being failed */
   public SnapshotDescription getSnapshotDescription() {
     return this.description;
   }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java
index 003878e7b21..4208ea113e4 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java
@@ -137,9 +137,7 @@ public class ZNodePaths {
       .append("]").toString();
   }
 
-  /**
-   * @return the znode string corresponding to a replicaId
-   */
+  /** Returns the znode string corresponding to a replicaId */
   public String getZNodeForReplica(int replicaId) {
     if (RegionReplicaUtil.isDefaultReplica(replicaId)) {
       return joinZNode(baseZNode, metaZNodePrefix);
@@ -169,16 +167,12 @@ public class ZNodePaths {
       : Integer.parseInt(znode.substring(metaZNodePrefix.length() + 1));
   }
 
-  /**
-   * @return True if meta znode.
-   */
+  /** Returns True if meta znode. */
   public boolean isMetaZNodePrefix(String znode) {
     return znode != null && znode.startsWith(this.metaZNodePrefix);
   }
 
-  /**
-   * @return True is the fully qualified path is for meta location
-   */
+  /** Returns True is the fully qualified path is for meta location */
   public boolean isMetaZNodePath(String path) {
     int prefixLen = baseZNode.length() + 1;
     return path.length() > prefixLen && isMetaZNodePrefix(path.substring(prefixLen));
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/ArrayBackedTag.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/ArrayBackedTag.java
index 434f731883a..efa64f09e72 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/ArrayBackedTag.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/ArrayBackedTag.java
@@ -91,33 +91,25 @@ public class ArrayBackedTag implements Tag {
     this.type = bytes[offset + TAG_LENGTH_SIZE];
   }
 
-  /**
-   * @return The byte array backing this Tag.
-   */
+  /** Returns The byte array backing this Tag. */
   @Override
   public byte[] getValueArray() {
     return this.bytes;
   }
 
-  /**
-   * @return the tag type
-   */
+  /** Returns the tag type */
   @Override
   public byte getType() {
     return this.type;
   }
 
-  /**
-   * @return Length of actual tag bytes within the backed buffer
-   */
+  /** Returns Length of actual tag bytes within the backed buffer */
   @Override
   public int getValueLength() {
     return this.length - INFRASTRUCTURE_SIZE;
   }
 
-  /**
-   * @return Offset of actual tag bytes within the backed buffer
-   */
+  /** Returns Offset of actual tag bytes within the backed buffer */
   @Override
   public int getValueOffset() {
     return this.offset + INFRASTRUCTURE_SIZE;
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/ByteBufferExtendedCell.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/ByteBufferExtendedCell.java
index 824625074b5..cb3d379849e 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/ByteBufferExtendedCell.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/ByteBufferExtendedCell.java
@@ -60,53 +60,33 @@ import org.apache.yetus.audience.InterfaceAudience;
  */
 @InterfaceAudience.Private
 public abstract class ByteBufferExtendedCell implements ExtendedCell {
-  /**
-   * @return The {@link ByteBuffer} containing the row bytes.
-   */
+  /** Returns The {@link ByteBuffer} containing the row bytes. */
   public abstract ByteBuffer getRowByteBuffer();
 
-  /**
-   * @return Position in the {@link ByteBuffer} where row bytes start
-   */
+  /** Returns Position in the {@link ByteBuffer} where row bytes start */
   public abstract int getRowPosition();
 
-  /**
-   * @return The {@link ByteBuffer} containing the column family bytes.
-   */
+  /** Returns The {@link ByteBuffer} containing the column family bytes. */
   public abstract ByteBuffer getFamilyByteBuffer();
 
-  /**
-   * @return Position in the {@link ByteBuffer} where column family bytes start
-   */
+  /** Returns Position in the {@link ByteBuffer} where column family bytes start */
   public abstract int getFamilyPosition();
 
-  /**
-   * @return The {@link ByteBuffer} containing the column qualifier bytes.
-   */
+  /** Returns The {@link ByteBuffer} containing the column qualifier bytes. */
   public abstract ByteBuffer getQualifierByteBuffer();
 
-  /**
-   * @return Position in the {@link ByteBuffer} where column qualifier bytes start
-   */
+  /** Returns Position in the {@link ByteBuffer} where column qualifier bytes start */
   public abstract int getQualifierPosition();
 
-  /**
-   * @return The {@link ByteBuffer} containing the value bytes.
-   */
+  /** Returns The {@link ByteBuffer} containing the value bytes. */
   public abstract ByteBuffer getValueByteBuffer();
 
-  /**
-   * @return Position in the {@link ByteBuffer} where value bytes start
-   */
+  /** Returns Position in the {@link ByteBuffer} where value bytes start */
   public abstract int getValuePosition();
 
-  /**
-   * @return The {@link ByteBuffer} containing the tag bytes.
-   */
+  /** Returns The {@link ByteBuffer} containing the tag bytes. */
   public abstract ByteBuffer getTagsByteBuffer();
 
-  /**
-   * @return Position in the {@link ByteBuffer} where tag bytes start
-   */
+  /** Returns Position in the {@link ByteBuffer} where tag bytes start */
   public abstract int getTagsPosition();
 }
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/Cell.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/Cell.java
index 4b7ab303175..4fb254b35b4 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/Cell.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/Cell.java
@@ -68,14 +68,10 @@ public interface Cell extends HeapSize {
    */
   byte[] getRowArray();
 
-  /**
-   * @return Array index of first row byte
-   */
+  /** Returns Array index of first row byte */
   int getRowOffset();
 
-  /**
-   * @return Number of row bytes. Must be &lt; rowArray.length - offset.
-   */
+  /** Returns Number of row bytes. Must be &lt; rowArray.length - offset. */
   short getRowLength();
 
   // 2) Family
@@ -87,14 +83,10 @@ public interface Cell extends HeapSize {
    */
   byte[] getFamilyArray();
 
-  /**
-   * @return Array index of first family byte
-   */
+  /** Returns Array index of first family byte */
   int getFamilyOffset();
 
-  /**
-   * @return Number of family bytes. Must be &lt; familyArray.length - offset.
-   */
+  /** Returns Number of family bytes. Must be &lt; familyArray.length - offset. */
   byte getFamilyLength();
 
   // 3) Qualifier
@@ -105,14 +97,10 @@ public interface Cell extends HeapSize {
    */
   byte[] getQualifierArray();
 
-  /**
-   * @return Array index of first qualifier byte
-   */
+  /** Returns Array index of first qualifier byte */
   int getQualifierOffset();
 
-  /**
-   * @return Number of qualifier bytes. Must be &lt; qualifierArray.length - offset.
-   */
+  /** Returns Number of qualifier bytes. Must be &lt; qualifierArray.length - offset. */
   int getQualifierLength();
 
   // 4) Timestamp
@@ -154,19 +142,13 @@ public interface Cell extends HeapSize {
    */
   byte[] getValueArray();
 
-  /**
-   * @return Array index of first value byte
-   */
+  /** Returns Array index of first value byte */
   int getValueOffset();
 
-  /**
-   * @return Number of value bytes. Must be &lt; valueArray.length - offset.
-   */
+  /** Returns Number of value bytes. Must be &lt; valueArray.length - offset. */
   int getValueLength();
 
-  /**
-   * @return Serialized size (defaults to include tag length if has some tags).
-   */
+  /** Returns Serialized size (defaults to include tag length if has some tags). */
   int getSerializedSize();
 
   /**
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScannable.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScannable.java
index 454f5be3bee..b194f08d4c2 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScannable.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScannable.java
@@ -26,8 +26,6 @@ import org.apache.yetus.audience.InterfaceAudience;
  */
 @InterfaceAudience.Public
 public interface CellScannable {
-  /**
-   * @return A CellScanner over the contained {@link Cell}s
-   */
+  /** Returns A CellScanner over the contained {@link Cell}s */
   CellScanner cellScanner();
 }
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScanner.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScanner.java
index df2f24b5b1b..8d2573765be 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScanner.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScanner.java
@@ -47,9 +47,7 @@ import org.apache.yetus.audience.InterfaceAudience;
  */
 @InterfaceAudience.Public
 public interface CellScanner {
-  /**
-   * @return the current Cell which may be mutable
-   */
+  /** Returns the current Cell which may be mutable */
   Cell current();
 
   /**
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellUtil.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellUtil.java
index 687114e0921..604aa1dff0d 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellUtil.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellUtil.java
@@ -856,9 +856,7 @@ public final class CellUtil {
     return matchingFamily(left, fam) && matchingQualifier(left, qual);
   }
 
-  /**
-   * @return True if matching column family and the qualifier starts with <code>qual</code>
-   */
+  /** Returns True if matching column family and the qualifier starts with <code>qual</code> */
   public static boolean matchingColumnFamilyAndQualifierPrefix(final Cell left, final byte[] fam,
     final byte[] qual) {
     return matchingFamily(left, fam) && PrivateCellUtil.qualifierStartsWith(left, qual);
@@ -976,9 +974,7 @@ public final class CellUtil {
     return t == Type.DeleteColumn.getCode() || t == Type.DeleteFamily.getCode();
   }
 
-  /**
-   * @return True if this cell is a Put.
-   */
+  /** Returns True if this cell is a Put. */
   @SuppressWarnings("deprecation")
   public static boolean isPut(Cell cell) {
     return cell.getTypeByte() == Type.Put.getCode();
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/ChoreService.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/ChoreService.java
index 1bba8d49120..b93610f26d0 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/ChoreService.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/ChoreService.java
@@ -228,9 +228,7 @@ public class ChoreService {
     }
   }
 
-  /**
-   * @return true when the chore is scheduled with the implementer of this interface
-   */
+  /** Returns true when the chore is scheduled with the implementer of this interface */
   @InterfaceAudience.Private
   public synchronized boolean isChoreScheduled(ScheduledChore chore) {
     return chore != null && scheduledChores.containsKey(chore)
@@ -248,9 +246,7 @@ public class ChoreService {
     rescheduleChore(chore);
   }
 
-  /**
-   * @return number of chores that this service currently has scheduled
-   */
+  /** Returns number of chores that this service currently has scheduled */
   int getNumberOfScheduledChores() {
     return scheduledChores.size();
   }
@@ -263,9 +259,7 @@ public class ChoreService {
     return choresMissingStartTime.size();
   }
 
-  /**
-   * @return number of threads in the core pool of the underlying ScheduledThreadPoolExecutor
-   */
+  /** Returns number of threads in the core pool of the underlying ScheduledThreadPoolExecutor */
   int getCorePoolSize() {
     return scheduler.getCorePoolSize();
   }
@@ -375,16 +369,12 @@ public class ChoreService {
     choresMissingStartTime.clear();
   }
 
-  /**
-   * @return true when the service is shutdown and thus cannot be used anymore
-   */
+  /** Returns true when the service is shutdown and thus cannot be used anymore */
   public boolean isShutdown() {
     return scheduler.isShutdown();
   }
 
-  /**
-   * @return true when the service is shutdown and all threads have terminated
-   */
+  /** Returns true when the service is shutdown and all threads have terminated */
   public boolean isTerminated() {
     return scheduler.isTerminated();
   }
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/ExtendedCell.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/ExtendedCell.java
index b36e1c90aa5..b3b7a1c5e57 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/ExtendedCell.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/ExtendedCell.java
@@ -86,9 +86,7 @@ public interface ExtendedCell extends RawCell, HeapSize {
       getValueLength(), getTagsLength(), withTags);
   }
 
-  /**
-   * @return Serialized size (defaults to include tag length).
-   */
+  /** Returns Serialized size (defaults to include tag length). */
   @Override
   default int getSerializedSize() {
     return getSerializedSize(true);
@@ -155,9 +153,7 @@ public interface ExtendedCell extends RawCell, HeapSize {
    */
   byte[] getTagsArray();
 
-  /**
-   * @return the first offset where the tags start in the Cell
-   */
+  /** Returns the first offset where the tags start in the Cell */
   int getTagsOffset();
 
   /**
@@ -170,8 +166,6 @@ public interface ExtendedCell extends RawCell, HeapSize {
    */
   int getTagsLength();
 
-  /**
-   * @return The byte representation of the KeyValue.TYPE of this cell: one of Put, Delete, etc
-   */
+  /** Returns The byte representation of the KeyValue.TYPE of this cell: one of Put, Delete, etc */
   byte getTypeByte();
 }
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/HBaseConfiguration.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/HBaseConfiguration.java
index 91f232f2dc8..0573fe5e45c 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/HBaseConfiguration.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/HBaseConfiguration.java
@@ -185,9 +185,7 @@ public class HBaseConfiguration extends Configuration {
     }
   }
 
-  /**
-   * @return whether to show HBase Configuration in servlet
-   */
+  /** Returns whether to show HBase Configuration in servlet */
   public static boolean isShowConfInServlet() {
     boolean isShowConf = false;
     try {
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
index c8956bb02f4..85b8d9d6d31 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
@@ -1164,16 +1164,12 @@ public class KeyValue implements ExtendedCell, Cloneable {
     return this.bytes;
   }
 
-  /**
-   * @return Offset into {@link #getBuffer()} at which this KeyValue starts.
-   */
+  /** Returns Offset into {@link #getBuffer()} at which this KeyValue starts. */
   public int getOffset() {
     return this.offset;
   }
 
-  /**
-   * @return Length of bytes this KeyValue occupies in {@link #getBuffer()}.
-   */
+  /** Returns Length of bytes this KeyValue occupies in {@link #getBuffer()}. */
   public int getLength() {
     return length;
   }
@@ -1197,9 +1193,7 @@ public class KeyValue implements ExtendedCell, Cloneable {
     return klength + vlength;
   }
 
-  /**
-   * @return Key offset in backing buffer..
-   */
+  /** Returns Key offset in backing buffer.. */
   public int getKeyOffset() {
     return this.offset + ROW_OFFSET;
   }
@@ -1208,33 +1202,27 @@ public class KeyValue implements ExtendedCell, Cloneable {
     return Bytes.toStringBinary(getBuffer(), getKeyOffset(), getKeyLength());
   }
 
-  /**
-   * @return Length of key portion.
-   */
+  /** Returns Length of key portion. */
   public int getKeyLength() {
     return Bytes.toInt(this.bytes, this.offset);
   }
 
   /**
-   * @return the backing array of the entire KeyValue (all KeyValue fields are in a single array)
+   * Returns the backing array of the entire KeyValue (all KeyValue fields are in a single array)
    */
   @Override
   public byte[] getValueArray() {
     return bytes;
   }
 
-  /**
-   * @return the value offset
-   */
+  /** Returns the value offset */
   @Override
   public int getValueOffset() {
     int voffset = getKeyOffset() + getKeyLength();
     return voffset;
   }
 
-  /**
-   * @return Value length
-   */
+  /** Returns Value length */
   @Override
   public int getValueLength() {
     int vlength = Bytes.toInt(this.bytes, this.offset + Bytes.SIZEOF_INT);
@@ -1242,63 +1230,51 @@ public class KeyValue implements ExtendedCell, Cloneable {
   }
 
   /**
-   * @return the backing array of the entire KeyValue (all KeyValue fields are in a single array)
+   * Returns the backing array of the entire KeyValue (all KeyValue fields are in a single array)
    */
   @Override
   public byte[] getRowArray() {
     return bytes;
   }
 
-  /**
-   * @return Row offset
-   */
+  /** Returns Row offset */
   @Override
   public int getRowOffset() {
     return this.offset + ROW_KEY_OFFSET;
   }
 
-  /**
-   * @return Row length
-   */
+  /** Returns Row length */
   @Override
   public short getRowLength() {
     return Bytes.toShort(this.bytes, getKeyOffset());
   }
 
   /**
-   * @return the backing array of the entire KeyValue (all KeyValue fields are in a single array)
+   * Returns the backing array of the entire KeyValue (all KeyValue fields are in a single array)
    */
   @Override
   public byte[] getFamilyArray() {
     return bytes;
   }
 
-  /**
-   * @return Family offset
-   */
+  /** Returns Family offset */
   @Override
   public int getFamilyOffset() {
     return getFamilyOffset(getFamilyLengthPosition(getRowLength()));
   }
 
-  /**
-   * @return Family offset
-   */
+  /** Returns Family offset */
   int getFamilyOffset(int familyLenPosition) {
     return familyLenPosition + Bytes.SIZEOF_BYTE;
   }
 
-  /**
-   * @return Family length
-   */
+  /** Returns Family length */
   @Override
   public byte getFamilyLength() {
     return getFamilyLength(getFamilyLengthPosition(getRowLength()));
   }
 
-  /**
-   * @return Family length
-   */
+  /** Returns Family length */
   public byte getFamilyLength(int famLenPos) {
     return this.bytes[famLenPos];
   }
@@ -1308,60 +1284,46 @@ public class KeyValue implements ExtendedCell, Cloneable {
   }
 
   /**
-   * @return the backing array of the entire KeyValue (all KeyValue fields are in a single array)
+   * Returns the backing array of the entire KeyValue (all KeyValue fields are in a single array)
    */
   @Override
   public byte[] getQualifierArray() {
     return bytes;
   }
 
-  /**
-   * @return Qualifier offset
-   */
+  /** Returns Qualifier offset */
   @Override
   public int getQualifierOffset() {
     return getQualifierOffset(getFamilyOffset());
   }
 
-  /**
-   * @return Qualifier offset
-   */
+  /** Returns Qualifier offset */
   private int getQualifierOffset(int foffset) {
     return getQualifierOffset(foffset, getFamilyLength());
   }
 
-  /**
-   * @return Qualifier offset
-   */
+  /** Returns Qualifier offset */
   int getQualifierOffset(int foffset, int flength) {
     return foffset + flength;
   }
 
-  /**
-   * @return Qualifier length
-   */
+  /** Returns Qualifier length */
   @Override
   public int getQualifierLength() {
     return getQualifierLength(getRowLength(), getFamilyLength());
   }
 
-  /**
-   * @return Qualifier length
-   */
+  /** Returns Qualifier length */
   private int getQualifierLength(int rlength, int flength) {
     return getQualifierLength(getKeyLength(), rlength, flength);
   }
 
-  /**
-   * @return Qualifier length
-   */
+  /** Returns Qualifier length */
   int getQualifierLength(int keyLength, int rlength, int flength) {
     return keyLength - (int) getKeyDataStructureSize(rlength, flength, 0);
   }
 
-  /**
-   * @return Timestamp offset
-   */
+  /** Returns Timestamp offset */
   public int getTimestampOffset() {
     return getTimestampOffset(getKeyLength());
   }
@@ -1374,9 +1336,7 @@ public class KeyValue implements ExtendedCell, Cloneable {
     return getKeyOffset() + keylength - TIMESTAMP_TYPE_SIZE;
   }
 
-  /**
-   * @return True if this KeyValue has a LATEST_TIMESTAMP timestamp.
-   */
+  /** Returns True if this KeyValue has a LATEST_TIMESTAMP timestamp. */
   public boolean isLatestTimestamp() {
     return Bytes.equals(getBuffer(), getTimestampOffset(), Bytes.SIZEOF_LONG,
       HConstants.LATEST_TIMESTAMP_BYTES, 0, Bytes.SIZEOF_LONG);
@@ -1442,9 +1402,7 @@ public class KeyValue implements ExtendedCell, Cloneable {
     return Bytes.toLong(this.bytes, tsOffset);
   }
 
-  /**
-   * @return KeyValue.TYPE byte representation
-   */
+  /** Returns KeyValue.TYPE byte representation */
   @Override
   public byte getTypeByte() {
     return getTypeByte(getKeyLength());
@@ -1481,7 +1439,7 @@ public class KeyValue implements ExtendedCell, Cloneable {
   }
 
   /**
-   * @return the backing array of the entire KeyValue (all KeyValue fields are in a single array)
+   * Returns the backing array of the entire KeyValue (all KeyValue fields are in a single array)
    */
   @Override
   public byte[] getTagsArray() {
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/PrivateCellUtil.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/PrivateCellUtil.java
index 525433209a3..b3e70132dfc 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/PrivateCellUtil.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/PrivateCellUtil.java
@@ -100,16 +100,12 @@ public final class PrivateCellUtil {
     return buffer;
   }
 
-  /**
-   * @return A new cell which is having the extra tags also added to it.
-   */
+  /** Returns A new cell which is having the extra tags also added to it. */
   public static Cell createCell(Cell cell, List<Tag> tags) {
     return createCell(cell, TagUtil.fromList(tags));
   }
 
-  /**
-   * @return A new cell which is having the extra tags also added to it.
-   */
+  /** Returns A new cell which is having the extra tags also added to it. */
   public static Cell createCell(Cell cell, byte[] tags) {
     if (cell instanceof ByteBufferExtendedCell) {
       return new TagRewriteByteBufferExtendedCell((ByteBufferExtendedCell) cell, tags);
@@ -839,9 +835,7 @@ public final class PrivateCellUtil {
     return KeyValue.Type.Delete.getCode() <= type && type <= KeyValue.Type.DeleteFamily.getCode();
   }
 
-  /**
-   * @return True if this cell is a {@link KeyValue.Type#Delete} type.
-   */
+  /** Returns True if this cell is a {@link KeyValue.Type#Delete} type. */
   public static boolean isDeleteType(Cell cell) {
     return cell.getTypeByte() == KeyValue.Type.Delete.getCode();
   }
@@ -862,9 +856,7 @@ public final class PrivateCellUtil {
     return cell.getTypeByte() == KeyValue.Type.Delete.getCode();
   }
 
-  /**
-   * @return True if this cell is a delete family or column type.
-   */
+  /** Returns True if this cell is a delete family or column type. */
   public static boolean isDeleteColumnOrFamily(Cell cell) {
     int t = cell.getTypeByte();
     return t == KeyValue.Type.DeleteColumn.getCode() || t == KeyValue.Type.DeleteFamily.getCode();
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/RawCell.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/RawCell.java
index 3961e565721..b4d26dbfee9 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/RawCell.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/RawCell.java
@@ -69,9 +69,7 @@ public interface RawCell extends Cell {
     }
   }
 
-  /**
-   * @return A new cell which is having the extra tags also added to it.
-   */
+  /** Returns A new cell which is having the extra tags also added to it. */
   public static Cell createCell(Cell cell, List<Tag> tags) {
     return PrivateCellUtil.createCell(cell, tags);
   }
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/RawCellBuilderFactory.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/RawCellBuilderFactory.java
index 50a320b1033..dec7b083e74 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/RawCellBuilderFactory.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/RawCellBuilderFactory.java
@@ -28,9 +28,7 @@ import org.apache.yetus.audience.InterfaceAudience;
 @InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.UNITTEST)
 public final class RawCellBuilderFactory {
 
-  /**
-   * @return the cell that is created
-   */
+  /** Returns the cell that is created */
   public static RawCellBuilder create() {
     return new KeyValueBuilder();
   }
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/ScheduledChore.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/ScheduledChore.java
index 68c8d69410a..f6b6537f5d4 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/ScheduledChore.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/ScheduledChore.java
@@ -200,17 +200,13 @@ public abstract class ScheduledChore implements Runnable {
     return timeOfThisRun - timeOfLastRun;
   }
 
-  /**
-   * @return true when the time between runs exceeds the acceptable threshold
-   */
+  /** Returns true when the time between runs exceeds the acceptable threshold */
   private synchronized boolean missedStartTime() {
     return isValidTime(timeOfLastRun) && isValidTime(timeOfThisRun)
       && getTimeBetweenRuns() > getMaximumAllowedTimeBetweenRuns();
   }
 
-  /**
-   * @return max allowed time in millis between runs.
-   */
+  /** Returns max allowed time in millis between runs. */
   private double getMaximumAllowedTimeBetweenRuns() {
     // Threshold used to determine if the Chore's current run started too late
     return 1.5 * timeUnit.toMillis(period);
@@ -224,9 +220,7 @@ public abstract class ScheduledChore implements Runnable {
     return time > 0 && time <= EnvironmentEdgeManager.currentTime();
   }
 
-  /**
-   * @return false when the Chore is not currently scheduled with a ChoreService
-   */
+  /** Returns false when the Chore is not currently scheduled with a ChoreService */
   public synchronized boolean triggerNow() {
     if (choreService == null) {
       return false;
@@ -261,16 +255,12 @@ public abstract class ScheduledChore implements Runnable {
     return stopper;
   }
 
-  /**
-   * @return period to execute chore in getTimeUnit() units
-   */
+  /** Returns period to execute chore in getTimeUnit() units */
   public int getPeriod() {
     return period;
   }
 
-  /**
-   * @return initial delay before executing chore in getTimeUnit() units
-   */
+  /** Returns initial delay before executing chore in getTimeUnit() units */
   public long getInitialDelay() {
     return initialDelay;
   }
@@ -295,9 +285,7 @@ public abstract class ScheduledChore implements Runnable {
     return timeOfThisRun;
   }
 
-  /**
-   * @return true when this Chore is scheduled with a ChoreService
-   */
+  /** Returns true when this Chore is scheduled with a ChoreService */
   public synchronized boolean isScheduled() {
     return choreService != null && choreService.isChoreScheduled(this);
   }
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/ServerName.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/ServerName.java
index c65a821ad37..c6008561969 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/ServerName.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/ServerName.java
@@ -412,9 +412,7 @@ public class ServerName implements Comparable<ServerName>, Serializable {
     return SERVERNAME_PATTERN.matcher(str).matches() ? valueOf(str) : valueOf(str, NON_STARTCODE);
   }
 
-  /**
-   * @return true if the String follows the pattern of {@link #toString()}, false otherwise.
-   */
+  /** Returns true if the String follows the pattern of {@link #toString()}, false otherwise. */
   public static boolean isFullServerName(final String str) {
     if (str == null || str.isEmpty()) {
       return false;
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/Stoppable.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/Stoppable.java
index f2d9b6b4ead..5d6b4957298 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/Stoppable.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/Stoppable.java
@@ -30,8 +30,6 @@ public interface Stoppable {
    */
   void stop(String why);
 
-  /**
-   * @return True if {@link #stop(String)} has been closed.
-   */
+  /** Returns True if {@link #stop(String)} has been closed. */
   boolean isStopped();
 }
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/TableName.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/TableName.java
index 43732ad6c8a..18eba1ebb53 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/TableName.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/TableName.java
@@ -76,9 +76,7 @@ public final class TableName implements Comparable<TableName> {
   /** One globally disallowed name */
   public static final String DISALLOWED_TABLE_NAME = "zookeeper";
 
-  /**
-   * @return True if <code>tn</code> is the hbase:meta table name.
-   */
+  /** Returns True if <code>tn</code> is the hbase:meta table name. */
   public static boolean isMetaTableName(final TableName tn) {
     return tn.equals(TableName.META_TABLE_NAME);
   }
@@ -254,9 +252,7 @@ public final class TableName implements Comparable<TableName> {
     return qualifierAsString;
   }
 
-  /**
-   * @return A pointer to TableName as String bytes.
-   */
+  /** Returns A pointer to TableName as String bytes. */
   public byte[] toBytes() {
     return name;
   }
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/Tag.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/Tag.java
index cf8798a0e6b..fd0aa43190b 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/Tag.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/Tag.java
@@ -46,19 +46,13 @@ public interface Tag {
    */
   public static final byte CUSTOM_TAG_TYPE_RANGE = (byte) 64;
 
-  /**
-   * @return the tag type
-   */
+  /** Returns the tag type */
   byte getType();
 
-  /**
-   * @return Offset of tag value within the backed buffer
-   */
+  /** Returns Offset of tag value within the backed buffer */
   int getValueOffset();
 
-  /**
-   * @return Length of tag value within the backed buffer
-   */
+  /** Returns Length of tag value within the backed buffer */
   int getValueLength();
 
   /**
@@ -73,9 +67,7 @@ public interface Tag {
    */
   byte[] getValueArray();
 
-  /**
-   * @return The {@link java.nio.ByteBuffer} containing the value bytes.
-   */
+  /** Returns The {@link java.nio.ByteBuffer} containing the value bytes. */
   ByteBuffer getValueByteBuffer();
 
   /**
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/TagUtil.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/TagUtil.java
index 53158800cfd..c3c49410210 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/TagUtil.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/TagUtil.java
@@ -65,9 +65,7 @@ public final class TagUtil {
     return StreamUtils.readRawVarint32(tag.getValueByteBuffer(), offset);
   }
 
-  /**
-   * @return A List&lt;Tag&gt; of any Tags found in <code>cell</code> else null.
-   */
+  /** Returns A List&lt;Tag&gt; of any Tags found in <code>cell</code> else null. */
   public static List<Tag> carryForwardTags(final Cell cell) {
     return carryForwardTags(null, cell);
   }
@@ -108,9 +106,7 @@ public final class TagUtil {
     return b;
   }
 
-  /**
-   * @return Carry forward the TTL tag.
-   */
+  /** Returns Carry forward the TTL tag. */
   public static List<Tag> carryForwardTTLTag(final List<Tag> tagsOrNull, final long ttl) {
     if (ttl == Long.MAX_VALUE) {
       return tagsOrNull;
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/codec/CellCodec.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/codec/CellCodec.java
index 10a363a5f40..e7facdbfbf2 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/codec/CellCodec.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/codec/CellCodec.java
@@ -97,9 +97,7 @@ public class CellCodec implements Codec {
         .setTimestamp(timestamp).setType(type).setValue(value).setSequenceId(memstoreTS).build();
     }
 
-    /**
-     * @return Byte array read from the stream. n
-     */
+    /** Returns Byte array read from the stream. n */
     private byte[] readByteArray(final InputStream in) throws IOException {
       byte[] intArray = new byte[Bytes.SIZEOF_INT];
       IOUtils.readFully(in, intArray);
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/codec/CellCodecWithTags.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/codec/CellCodecWithTags.java
index fe9417b23c9..75e3d48d9fa 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/codec/CellCodecWithTags.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/codec/CellCodecWithTags.java
@@ -101,9 +101,7 @@ public class CellCodecWithTags implements Codec {
         .setTags(tags).build();
     }
 
-    /**
-     * @return Byte array read from the stream. n
-     */
+    /** Returns Byte array read from the stream. n */
     private byte[] readByteArray(final InputStream in) throws IOException {
       byte[] intArray = new byte[Bytes.SIZEOF_INT];
       IOUtils.readFully(in, intArray);
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/conf/ConfigurationManager.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/conf/ConfigurationManager.java
index 91d2a7be266..0605b3e5190 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/conf/ConfigurationManager.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/conf/ConfigurationManager.java
@@ -117,9 +117,7 @@ public class ConfigurationManager {
     }
   }
 
-  /**
-   * @return the number of observers.
-   */
+  /** Returns the number of observers. */
   public int getNumObservers() {
     synchronized (configurationObservers) {
       return configurationObservers.size();
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/filter/ByteArrayComparable.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/filter/ByteArrayComparable.java
index 332d44a84e5..480680b00f3 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/filter/ByteArrayComparable.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/filter/ByteArrayComparable.java
@@ -46,9 +46,7 @@ public abstract class ByteArrayComparable implements Comparable<byte[]> {
     return value;
   }
 
-  /**
-   * @return The comparator serialized using pb
-   */
+  /** Returns The comparator serialized using pb */
   public abstract byte[] toByteArray();
 
   /**
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ByteArrayOutputStream.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ByteArrayOutputStream.java
index 64bef1b83dc..d6b4c34aebb 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ByteArrayOutputStream.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ByteArrayOutputStream.java
@@ -116,16 +116,12 @@ public class ByteArrayOutputStream extends OutputStream implements ByteBufferWri
     buff.put(buf, 0, pos);
   }
 
-  /**
-   * @return the underlying array where the data gets accumulated
-   */
+  /** Returns the underlying array where the data gets accumulated */
   public byte[] getBuffer() {
     return this.buf;
   }
 
-  /**
-   * @return The current size of the buffer.
-   */
+  /** Returns The current size of the buffer. */
   public int size() {
     return this.pos;
   }
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ImmutableBytesWritable.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ImmutableBytesWritable.java
index 4074eed46e5..da897b04ae1 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ImmutableBytesWritable.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ImmutableBytesWritable.java
@@ -122,9 +122,7 @@ public class ImmutableBytesWritable implements WritableComparable<ImmutableBytes
     return this.length;
   }
 
-  /**
-   * @return the number of valid bytes in the buffer
-   */
+  /** Returns the number of valid bytes in the buffer */
   public int getLength() {
     if (this.bytes == null) {
       throw new IllegalStateException(
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/TimeRange.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/TimeRange.java
index b848887bdb6..0c96039c4f8 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/TimeRange.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/TimeRange.java
@@ -162,16 +162,12 @@ public class TimeRange {
     }
   }
 
-  /**
-   * @return the smallest timestamp that should be considered
-   */
+  /** Returns the smallest timestamp that should be considered */
   public long getMin() {
     return minStamp;
   }
 
-  /**
-   * @return the biggest timestamp that should be considered
-   */
+  /** Returns the biggest timestamp that should be considered */
   public long getMax() {
     return maxStamp;
   }
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.java
index b155d9b854e..c2bd9d39da7 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.java
@@ -1001,9 +1001,7 @@ abstract class BufferedDataBlockEncoder extends AbstractDataBlockEncoder {
     abstract protected void decodeNext();
   }
 
-  /**
-   * @return unencoded size added
-   */
+  /** Returns unencoded size added */
   protected final int afterEncodingKeyValue(Cell cell, DataOutputStream out,
     HFileBlockDefaultEncodingContext encodingCtx) throws IOException {
     int size = 0;
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoder.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoder.java
index 266dc26d594..7f13b2c6f66 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoder.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoder.java
@@ -129,9 +129,7 @@ public interface DataBlockEncoder {
      */
     ByteBuffer getValueShallowCopy();
 
-    /**
-     * @return the Cell at the current position. Includes memstore timestamp.
-     */
+    /** Returns the Cell at the current position. Includes memstore timestamp. */
     Cell getCell();
 
     /** Set position to beginning of given block */
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoding.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoding.java
index 54d045ba95d..71808adf75d 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoding.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoding.java
@@ -77,16 +77,12 @@ public enum DataBlockEncoding {
     this.encoderCls = encoderClsName;
   }
 
-  /**
-   * @return name converted to bytes.
-   */
+  /** Returns name converted to bytes. */
   public byte[] getNameInBytes() {
     return Bytes.toBytes(toString());
   }
 
-  /**
-   * @return The id of a data block encoder.
-   */
+  /** Returns The id of a data block encoder. */
   public short getId() {
     return id;
   }
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/EncodedDataBlock.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/EncodedDataBlock.java
index 4092a1ed2d6..d09ddc9126f 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/EncodedDataBlock.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/EncodedDataBlock.java
@@ -223,7 +223,7 @@ public class EncodedDataBlock {
     return getCompressedSize(comprAlgo, compressor, compressedBytes, 0, compressedBytes.length);
   }
 
-  /** @return encoded data with header */
+  /** Returns encoded data with header */
   private byte[] getEncodedData() {
     if (cachedEncodedData != null) {
       return cachedEncodedData;
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/HFileBlockDecodingContext.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/HFileBlockDecodingContext.java
index 16de5fb2669..6835a8bac3c 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/HFileBlockDecodingContext.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/HFileBlockDecodingContext.java
@@ -39,8 +39,6 @@ public interface HFileBlockDecodingContext {
   void prepareDecoding(int onDiskSizeWithoutHeader, int uncompressedSizeWithoutHeader,
     ByteBuff blockBufferWithoutHeader, ByteBuff onDiskBlock) throws IOException;
 
-  /**
-   * @return HFile meta information
-   */
+  /** Returns HFile meta information */
   HFileContext getHFileContext();
 }
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/HFileBlockEncodingContext.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/HFileBlockEncodingContext.java
index 15fe5eee5ea..fcac1549b9f 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/HFileBlockEncodingContext.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/HFileBlockEncodingContext.java
@@ -31,14 +31,10 @@ import org.apache.yetus.audience.InterfaceAudience;
 @InterfaceAudience.Private
 public interface HFileBlockEncodingContext {
 
-  /**
-   * @return the block type after encoding
-   */
+  /** Returns the block type after encoding */
   BlockType getBlockType();
 
-  /**
-   * @return the {@link DataBlockEncoding} encoding used
-   */
+  /** Returns the {@link DataBlockEncoding} encoding used */
   DataBlockEncoding getDataBlockEncoding();
 
   /**
@@ -52,9 +48,7 @@ public interface HFileBlockEncodingContext {
    */
   void close();
 
-  /**
-   * @return HFile context information
-   */
+  /** Returns HFile context information */
   HFileContext getHFileContext();
 
   /**
@@ -62,9 +56,7 @@ public interface HFileBlockEncodingContext {
    */
   void setEncodingState(EncodingState state);
 
-  /**
-   * @return the encoding state
-   */
+  /** Returns the encoding state */
   EncodingState getEncodingState();
 
   /**
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockType.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockType.java
index 5c8c366a228..3bd98e6388c 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockType.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockType.java
@@ -210,9 +210,7 @@ public enum BlockType {
     }
   }
 
-  /**
-   * @return whether this block type is encoded or unencoded data block
-   */
+  /** Returns whether this block type is encoded or unencoded data block */
   public final boolean isData() {
     return this == DATA || this == ENCODED_DATA;
   }
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileContext.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileContext.java
index 843c65a5f41..5064c4a4b8f 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileContext.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileContext.java
@@ -119,9 +119,7 @@ public class HFileContext implements HeapSize, Cloneable {
       : CellComparator.getInstance();
   }
 
-  /**
-   * @return true when on-disk blocks are compressed, and/or encrypted; false otherwise.
-   */
+  /** Returns true when on-disk blocks are compressed, and/or encrypted; false otherwise. */
   public boolean isCompressedOrEncrypted() {
     Compression.Algorithm compressAlgo = getCompression();
     boolean compressed = compressAlgo != null && compressAlgo != Compression.Algorithm.NONE;
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/ByteBuff.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/ByteBuff.java
index 2925fab161b..2285917364f 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/ByteBuff.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/ByteBuff.java
@@ -80,9 +80,7 @@ public abstract class ByteBuff implements HBaseReferenceCounted {
 
   /******************************* Methods for ByteBuff **************************************/
 
-  /**
-   * @return this ByteBuff's current position
-   */
+  /** Returns this ByteBuff's current position */
   public abstract int position();
 
   /**
@@ -102,9 +100,7 @@ public abstract class ByteBuff implements HBaseReferenceCounted {
    */
   public abstract ByteBuff moveBack(int len);
 
-  /**
-   * @return the total capacity of this ByteBuff.
-   */
+  /** Returns the total capacity of this ByteBuff. */
   public abstract int capacity();
 
   /**
@@ -263,19 +259,13 @@ public abstract class ByteBuff implements HBaseReferenceCounted {
    */
   public abstract ByteBuff put(byte[] src);
 
-  /**
-   * @return true or false if the underlying BB support hasArray
-   */
+  /** Returns true or false if the underlying BB support hasArray */
   public abstract boolean hasArray();
 
-  /**
-   * @return the byte[] if the underlying BB has single BB and hasArray true
-   */
+  /** Returns the byte[] if the underlying BB has single BB and hasArray true */
   public abstract byte[] array();
 
-  /**
-   * @return the arrayOffset of the byte[] incase of a single BB backed ByteBuff
-   */
+  /** Returns the arrayOffset of the byte[] incase of a single BB backed ByteBuff */
   public abstract int arrayOffset();
 
   /**
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/MultiByteBuff.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/MultiByteBuff.java
index a3c0b5c528c..042b21f6c00 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/MultiByteBuff.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/MultiByteBuff.java
@@ -131,17 +131,13 @@ public class MultiByteBuff extends ByteBuff {
     throw new UnsupportedOperationException();
   }
 
-  /**
-   * @return false. MBB does not support array based operations
-   */
+  /** Returns false. MBB does not support array based operations */
   @Override
   public boolean hasArray() {
     return false;
   }
 
-  /**
-   * @return the total capacity of this MultiByteBuffer.
-   */
+  /** Returns the total capacity of this MultiByteBuffer. */
   @Override
   public int capacity() {
     checkRefCount();
@@ -384,9 +380,7 @@ public class MultiByteBuff extends ByteBuff {
     return getLong(index, itemIndex);
   }
 
-  /**
-   * @return this MBB's current position
-   */
+  /** Returns this MBB's current position */
   @Override
   public int position() {
     checkRefCount();
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/security/User.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/security/User.java
index a353f11c488..e7d37069ac5 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/security/User.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/security/User.java
@@ -129,9 +129,7 @@ public abstract class User {
     ugi.addToken(token);
   }
 
-  /**
-   * @return true if user credentials are obtained from keytab.
-   */
+  /** Returns true if user credentials are obtained from keytab. */
   public boolean isLoginFromKeytab() {
     return ugi.isFromKeytab();
   }
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/security/UserProvider.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/security/UserProvider.java
index 5f1f65f40ee..f6766bd1f3c 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/security/UserProvider.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/security/UserProvider.java
@@ -149,9 +149,7 @@ public class UserProvider extends BaseConfigurable {
     return user == null ? null : user.getName();
   }
 
-  /**
-   * @return <tt>true</tt> if security is enabled, <tt>false</tt> otherwise
-   */
+  /** Returns <tt>true</tt> if security is enabled, <tt>false</tt> otherwise */
   public boolean isHBaseSecurityEnabled() {
     return User.isHBaseSecurityEnabled(this.getConf());
   }
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/AbstractByteRange.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/AbstractByteRange.java
index f2bf5cfa36d..c5fbf52fd46 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/AbstractByteRange.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/AbstractByteRange.java
@@ -152,9 +152,7 @@ public abstract class AbstractByteRange implements ByteRange {
     return isEmpty(this);
   }
 
-  /**
-   * @return true when {@code range} is of zero length, false otherwise.
-   */
+  /** Returns true when {@code range} is of zero length, false otherwise. */
   public static boolean isEmpty(ByteRange range) {
     return range == null || range.getLength() == 0;
   }
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteRange.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteRange.java
index 129f17d54df..64bd5cb3b6c 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteRange.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteRange.java
@@ -120,9 +120,7 @@ public interface ByteRange extends Comparable<ByteRange> {
    */
   public ByteRange setLength(int length);
 
-  /**
-   * @return true when this range is of zero length, false otherwise.
-   */
+  /** Returns true when this range is of zero length, false otherwise. */
   public boolean isEmpty();
 
   /**
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Bytes.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Bytes.java
index 5505f5f3bdf..1955aecf433 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Bytes.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Bytes.java
@@ -230,9 +230,7 @@ public class Bytes implements Comparable<Bytes> {
     return this.length;
   }
 
-  /**
-   * @return the number of valid bytes in the buffer
-   */
+  /** Returns the number of valid bytes in the buffer */
   public int getLength() {
     if (this.bytes == null) {
       throw new IllegalStateException(
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ExceptionUtil.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ExceptionUtil.java
index 26625a2862e..d3b0e82b1dc 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ExceptionUtil.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ExceptionUtil.java
@@ -33,9 +33,7 @@ public final class ExceptionUtil {
   private ExceptionUtil() {
   }
 
-  /**
-   * @return true if the throwable comes an interruption, false otherwise.
-   */
+  /** Returns true if the throwable comes an interruption, false otherwise. */
   public static boolean isInterrupt(Throwable t) {
     if (t instanceof InterruptedException) {
       return true;
@@ -59,9 +57,7 @@ public final class ExceptionUtil {
     }
   }
 
-  /**
-   * @return an InterruptedIOException if t was an interruption, null otherwise
-   */
+  /** Returns an InterruptedIOException if t was an interruption, null otherwise */
   public static InterruptedIOException asInterrupt(Throwable t) {
     if (t instanceof SocketTimeoutException) {
       return null;
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/HashKey.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/HashKey.java
index cf573aa394e..e4e89f7711f 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/HashKey.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/HashKey.java
@@ -36,8 +36,6 @@ public abstract class HashKey<T> {
    */
   public abstract byte get(int pos);
 
-  /**
-   * @return The number of bytes in this HashKey
-   */
+  /** Returns The number of bytes in this HashKey */
   public abstract int length();
 }
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Sleeper.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Sleeper.java
index 23ca92251c1..8fdcc189ea8 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Sleeper.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Sleeper.java
@@ -105,9 +105,7 @@ public class Sleeper {
     }
   }
 
-  /**
-   * @return the sleep period in milliseconds
-   */
+  /** Returns the sleep period in milliseconds */
   public final int getPeriod() {
     return period;
   }
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/WindowMovingAverage.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/WindowMovingAverage.java
index b79109cc991..b79924a352c 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/WindowMovingAverage.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/WindowMovingAverage.java
@@ -66,9 +66,7 @@ public class WindowMovingAverage extends MovingAverage {
     return oneRound;
   }
 
-  /**
-   * @return number of statistics
-   */
+  /** Returns number of statistics */
   protected int getNumberOfStatistics() {
     return lastN.length;
   }
@@ -85,9 +83,7 @@ public class WindowMovingAverage extends MovingAverage {
     return lastN[index];
   }
 
-  /**
-   * @return index of most recent
-   */
+  /** Returns index of most recent */
   protected int getMostRecentPosistion() {
     return mostRecent;
   }
diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/HBaseCommonTestingUtility.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/HBaseCommonTestingUtility.java
index cbc7717ca97..76086a3e732 100644
--- a/hbase-common/src/test/java/org/apache/hadoop/hbase/HBaseCommonTestingUtility.java
+++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/HBaseCommonTestingUtility.java
@@ -161,18 +161,14 @@ public class HBaseCommonTestingUtility {
     conf.set(propertyName, newDir.getAbsolutePath());
   }
 
-  /**
-   * @return True if we should delete testing dirs on exit.
-   */
+  /** Returns True if we should delete testing dirs on exit. */
   boolean deleteOnExit() {
     String v = System.getProperty("hbase.testing.preserve.testdir");
     // Let default be true, to delete on exit.
     return v == null ? true : !Boolean.parseBoolean(v);
   }
 
-  /**
-   * @return True if we removed the test dirs
-   */
+  /** Returns True if we removed the test dirs */
   public boolean cleanupTestDir() {
     if (deleteDir(this.dataTestDir)) {
       this.dataTestDir = null;
diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/net/BoundSocketMaker.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/net/BoundSocketMaker.java
index 69f072a0c6d..c8eb875ec0e 100644
--- a/hbase-common/src/test/java/org/apache/hadoop/hbase/net/BoundSocketMaker.java
+++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/net/BoundSocketMaker.java
@@ -53,9 +53,7 @@ public final class BoundSocketMaker implements Closeable {
     return this.socket.getLocalPort();
   }
 
-  /**
-   * @return Returns a bound socket; be sure to close when done.
-   */
+  /** Returns Returns a bound socket; be sure to close when done. */
   private ServerSocket get(String hostname, Supplier<Integer> randomPortMaker) {
     ServerSocket ss = null;
     int port = -1;
diff --git a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/master/MetricsAssignmentManagerSource.java b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/master/MetricsAssignmentManagerSource.java
index 76a70edbddf..6870c60838d 100644
--- a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/master/MetricsAssignmentManagerSource.java
+++ b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/master/MetricsAssignmentManagerSource.java
@@ -166,43 +166,27 @@ public interface MetricsAssignmentManagerSource extends BaseSource {
    */
   void incrementOperationCounter();
 
-  /**
-   * @return {@link OperationMetrics} containing common metrics for assign region operation
-   */
+  /** Returns {@link OperationMetrics} containing common metrics for assign region operation */
   OperationMetrics getAssignMetrics();
 
-  /**
-   * @return {@link OperationMetrics} containing common metrics for unassign region operation
-   */
+  /** Returns {@link OperationMetrics} containing common metrics for unassign region operation */
   OperationMetrics getUnassignMetrics();
 
-  /**
-   * @return {@link OperationMetrics} containing common metrics for move region operation
-   */
+  /** Returns {@link OperationMetrics} containing common metrics for move region operation */
   OperationMetrics getMoveMetrics();
 
-  /**
-   * @return {@link OperationMetrics} containing common metrics for reopen region operation
-   */
+  /** Returns {@link OperationMetrics} containing common metrics for reopen region operation */
   OperationMetrics getReopenMetrics();
 
-  /**
-   * @return {@link OperationMetrics} containing common metrics for open region request
-   */
+  /** Returns {@link OperationMetrics} containing common metrics for open region request */
   OperationMetrics getOpenMetrics();
 
-  /**
-   * @return {@link OperationMetrics} containing common metrics for close region request
-   */
+  /** Returns {@link OperationMetrics} containing common metrics for close region request */
   OperationMetrics getCloseMetrics();
 
-  /**
-   * @return {@link OperationMetrics} containing common metrics for split operation
-   */
+  /** Returns {@link OperationMetrics} containing common metrics for split operation */
   OperationMetrics getSplitMetrics();
 
-  /**
-   * @return {@link OperationMetrics} containing common metrics for merge operation
-   */
+  /** Returns {@link OperationMetrics} containing common metrics for merge operation */
   OperationMetrics getMergeMetrics();
 }
diff --git a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/master/MetricsMasterSource.java b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/master/MetricsMasterSource.java
index 06b2167a243..5952bdc4d8e 100644
--- a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/master/MetricsMasterSource.java
+++ b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/master/MetricsMasterSource.java
@@ -92,8 +92,6 @@ public interface MetricsMasterSource extends BaseSource {
    */
   void incRequests(final long inc);
 
-  /**
-   * @return {@link OperationMetrics} containing common metrics for server crash operation
-   */
+  /** Returns {@link OperationMetrics} containing common metrics for server crash operation */
   OperationMetrics getServerCrashMetrics();
 }
diff --git a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionServerWrapper.java b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionServerWrapper.java
index 1d5dcb23d83..46de6742db6 100644
--- a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionServerWrapper.java
+++ b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionServerWrapper.java
@@ -116,24 +116,16 @@ public interface MetricsRegionServerWrapper {
    */
   long getStoreFileSize();
 
-  /**
-   * @return Max age of store files hosted on this region server
-   */
+  /** Returns Max age of store files hosted on this region server */
   long getMaxStoreFileAge();
 
-  /**
-   * @return Min age of store files hosted on this region server
-   */
+  /** Returns Min age of store files hosted on this region server */
   long getMinStoreFileAge();
 
-  /**
-   * @return Average age of store files hosted on this region server
-   */
+  /** Returns Average age of store files hosted on this region server */
   long getAvgStoreFileAge();
 
-  /**
-   * @return Number of reference files on this region server
-   */
+  /** Returns Number of reference files on this region server */
   long getNumReferenceFiles();
 
   /**
@@ -507,43 +499,29 @@ public interface MetricsRegionServerWrapper {
    */
   double getMobFileCacheHitPercent();
 
-  /**
-   * @return Count of hedged read operations
-   */
+  /** Returns Count of hedged read operations */
   long getHedgedReadOps();
 
-  /**
-   * @return Count of times a hedged read beat out the primary read.
-   */
+  /** Returns Count of times a hedged read beat out the primary read. */
   long getHedgedReadWins();
 
-  /**
-   * @return Count of times a hedged read executes in current thread
-   */
+  /** Returns Count of times a hedged read executes in current thread */
   long getHedgedReadOpsInCurThread();
 
-  /**
-   * @return Number of total bytes read from HDFS.
-   */
+  /** Returns Number of total bytes read from HDFS. */
   long getTotalBytesRead();
 
-  /**
-   * @return Number of bytes read from the local HDFS DataNode.
-   */
+  /** Returns Number of bytes read from the local HDFS DataNode. */
   long getLocalBytesRead();
 
-  /**
-   * @return Number of bytes read locally through HDFS short circuit.
-   */
+  /** Returns Number of bytes read locally through HDFS short circuit. */
   long getShortCircuitBytesRead();
 
-  /**
-   * @return Number of bytes read locally through HDFS zero copy.
-   */
+  /** Returns Number of bytes read locally through HDFS zero copy. */
   long getZeroCopyBytesRead();
 
   /**
-   * @return Count of requests blocked because the memstore size is larger than blockingMemStoreSize
+   * Returns Count of requests blocked because the memstore size is larger than blockingMemStoreSize
    */
   long getBlockedRequestsCount();
 
diff --git a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionWrapper.java b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionWrapper.java
index b20d1c1ec80..2782af18b83 100644
--- a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionWrapper.java
+++ b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionWrapper.java
@@ -75,24 +75,16 @@ public interface MetricsRegionWrapper {
    */
   long getFilteredReadRequestCount();
 
-  /**
-   * @return Max age of store files under this region
-   */
+  /** Returns Max age of store files under this region */
   long getMaxStoreFileAge();
 
-  /**
-   * @return Min age of store files under this region
-   */
+  /** Returns Min age of store files under this region */
   long getMinStoreFileAge();
 
-  /**
-   * @return Average age of store files under this region
-   */
+  /** Returns Average age of store files under this region */
   long getAvgStoreFileAge();
 
-  /**
-   * @return Number of reference files under this region
-   */
+  /** Returns Number of reference files under this region */
   long getNumReferenceFiles();
 
   /**
@@ -108,9 +100,7 @@ public interface MetricsRegionWrapper {
 
   long getNumCompactionsCompleted();
 
-  /**
-   * @return Age of the last major compaction
-   */
+  /** Returns Age of the last major compaction */
   long getLastMajorCompactionAge();
 
   /**
@@ -151,9 +141,7 @@ public interface MetricsRegionWrapper {
    */
   int getReplicaId();
 
-  /**
-   * @return the number of references active on the store
-   */
+  /** Returns the number of references active on the store */
   long getStoreRefCount();
 
   /**
@@ -162,14 +150,10 @@ public interface MetricsRegionWrapper {
    */
   long getMaxCompactedStoreFileRefCount();
 
-  /**
-   * @return the number of row reads completely on memstore per store
-   */
+  /** Returns the number of row reads completely on memstore per store */
   Map<String, Long> getMemstoreOnlyRowReadsCount();
 
-  /**
-   * @return the number of row reads on memstore and file per store
-   */
+  /** Returns the number of row reads on memstore and file per store */
   Map<String, Long> getMixedRowReadsCount();
 
 }
diff --git a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsTableWrapperAggregate.java b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsTableWrapperAggregate.java
index 95489cb10b0..e18ba51b3c2 100644
--- a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsTableWrapperAggregate.java
+++ b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsTableWrapperAggregate.java
@@ -88,33 +88,21 @@ public interface MetricsTableWrapperAggregate {
    */
   long getMaxStoreFiles(String table);
 
-  /**
-   * @return Max age of store files for this table
-   */
+  /** Returns Max age of store files for this table */
   long getMaxStoreFileAge(String table);
 
-  /**
-   * @return Min age of store files for this table
-   */
+  /** Returns Min age of store files for this table */
   long getMinStoreFileAge(String table);
 
-  /**
-   * @return Average age of store files for this table
-   */
+  /** Returns Average age of store files for this table */
   long getAvgStoreFileAge(String table);
 
-  /**
-   * @return Number of reference files for this table
-   */
+  /** Returns Number of reference files for this table */
   long getNumReferenceFiles(String table);
 
-  /**
-   * @return number of row reads completely from memstore per store for this table
-   */
+  /** Returns number of row reads completely from memstore per store for this table */
   Map<String, Long> getMemstoreOnlyRowReadsCount(String table);
 
-  /**
-   * @return number of row reads from file and memstore per store for this table
-   */
+  /** Returns number of row reads from file and memstore per store for this table */
   Map<String, Long> getMixedRowReadsCount(String table);
 }
diff --git a/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/metrics2/lib/DynamicMetricsRegistry.java b/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/metrics2/lib/DynamicMetricsRegistry.java
index 7177f322210..f1b1196cea4 100644
--- a/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/metrics2/lib/DynamicMetricsRegistry.java
+++ b/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/metrics2/lib/DynamicMetricsRegistry.java
@@ -68,9 +68,7 @@ public class DynamicMetricsRegistry {
     metricsInfo = info;
   }
 
-  /**
-   * @return the info object of the metrics registry
-   */
+  /** Returns the info object of the metrics registry */
   public MetricsInfo info() {
     return metricsInfo;
   }
diff --git a/hbase-http/src/main/java/org/apache/hadoop/hbase/util/JSONBean.java b/hbase-http/src/main/java/org/apache/hadoop/hbase/util/JSONBean.java
index 2360dcb32a2..687b1e34041 100644
--- a/hbase-http/src/main/java/org/apache/hadoop/hbase/util/JSONBean.java
+++ b/hbase-http/src/main/java/org/apache/hadoop/hbase/util/JSONBean.java
@@ -129,9 +129,7 @@ public class JSONBean {
     };
   }
 
-  /**
-   * @return Return non-zero if failed to find bean. 0
-   */
+  /** Returns Return non-zero if failed to find bean. 0 */
   private static int write(JsonWriter writer, MBeanServer mBeanServer, ObjectName qry,
     String attribute, boolean description, ObjectName excluded) throws IOException {
     LOG.debug("Listing beans for {}", qry);
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestMobCompaction.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestMobCompaction.java
index 51f9bca5580..a5172997eba 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestMobCompaction.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestMobCompaction.java
@@ -162,9 +162,7 @@ public class IntegrationTestMobCompaction extends IntegrationTestBase {
     LOG.debug("Done initializing/checking cluster");
   }
 
-  /**
-   * @return status of CLI execution
-   */
+  /** Returns status of CLI execution */
   @Override
   public int runTestFromCommandLine() throws Exception {
     testMobCompaction();
diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapred/TableSplit.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapred/TableSplit.java
index 8cc03b22254..2cb63ba7a6a 100644
--- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapred/TableSplit.java
+++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapred/TableSplit.java
@@ -56,27 +56,27 @@ public class TableSplit implements InputSplit, Comparable<TableSplit> {
     this(TableName.valueOf(tableName), startRow, endRow, location);
   }
 
-  /** @return table name */
+  /** Returns table name */
   public TableName getTable() {
     return this.m_tableName;
   }
 
-  /** @return table name */
+  /** Returns table name */
   public byte[] getTableName() {
     return this.m_tableName.getName();
   }
 
-  /** @return starting row key */
+  /** Returns starting row key */
   public byte[] getStartRow() {
     return this.m_startRow;
   }
 
-  /** @return end row key */
+  /** Returns end row key */
   public byte[] getEndRow() {
     return this.m_endRow;
   }
 
-  /** @return the region's hostname */
+  /** Returns the region's hostname */
   public String getRegionLocation() {
     return this.m_regionLocation;
   }
diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/CellCreator.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/CellCreator.java
index bc2f8040db1..a48ba49058a 100644
--- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/CellCreator.java
+++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/CellCreator.java
@@ -120,9 +120,7 @@ public class CellCreator {
       qlength, timestamp, KeyValue.Type.Put, value, voffset, vlength, tags);
   }
 
-  /**
-   * @return Visibility expression resolver
-   */
+  /** Returns Visibility expression resolver */
   public VisibilityExpressionResolver getVisibilityExpressionResolver() {
     return this.visExpResolver;
   }
diff --git a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java
index beef9308ca6..ee36048cb3e 100644
--- a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java
+++ b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java
@@ -1406,16 +1406,12 @@ public class PerformanceEvaluation extends Configured implements Tool {
       }
     }
 
-    /**
-     * @return Subset of the histograms' calculation.
-     */
+    /** Returns Subset of the histograms' calculation. */
     public String getShortLatencyReport() {
       return YammerHistogramUtils.getShortHistogramReport(this.latencyHistogram);
     }
 
-    /**
-     * @return Subset of the histograms' calculation.
-     */
+    /** Returns Subset of the histograms' calculation. */
     public String getShortValueSizeReport() {
       return YammerHistogramUtils.getShortHistogramReport(this.valueSizeHistogram);
     }
diff --git a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestRoundRobinTableInputFormat.java b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestRoundRobinTableInputFormat.java
index df9d1597857..34bd76937d3 100644
--- a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestRoundRobinTableInputFormat.java
+++ b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestRoundRobinTableInputFormat.java
@@ -77,9 +77,7 @@ public class TestRoundRobinTableInputFormat {
     }
   }
 
-  /**
-   * @return Splits made out of {@link #KEYS}. Splits are for five Servers. Length is ZERO!
-   */
+  /** Returns Splits made out of {@link #KEYS}. Splits are for five Servers. Length is ZERO! */
   private List<InputSplit> createSplits() {
     List<InputSplit> splits = new ArrayList<>(KEYS.length - 1);
     for (int i = 0; i < KEYS.length - 1; i++) {
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/LockAndQueue.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/LockAndQueue.java
index e6b7d7d0397..fa6c72c490f 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/LockAndQueue.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/LockAndQueue.java
@@ -105,9 +105,7 @@ public class LockAndQueue implements LockStatus {
   // try/release Shared/Exclusive lock
   // ======================================================================
 
-  /**
-   * @return whether we have succesfully acquired the shared lock.
-   */
+  /** Returns whether we have succesfully acquired the shared lock. */
   public boolean trySharedLock(Procedure<?> proc) {
     if (hasExclusiveLock() && !hasLockAccess(proc)) {
       return false;
@@ -119,9 +117,7 @@ public class LockAndQueue implements LockStatus {
     return true;
   }
 
-  /**
-   * @return whether we should wake the procedures waiting on the lock here.
-   */
+  /** Returns whether we should wake the procedures waiting on the lock here. */
   public boolean releaseSharedLock() {
     // hasExclusiveLock could be true, it usually means we acquire shared lock while we or our
     // parent have held the xlock. And since there is still an exclusive lock, we do not need to
@@ -137,9 +133,7 @@ public class LockAndQueue implements LockStatus {
     return true;
   }
 
-  /**
-   * @return whether we should wake the procedures waiting on the lock here.
-   */
+  /** Returns whether we should wake the procedures waiting on the lock here. */
   public boolean releaseExclusiveLock(Procedure<?> proc) {
     if (
       exclusiveLockOwnerProcedure == null
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/Procedure.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/Procedure.java
index ec46e523fc5..a71026e6007 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/Procedure.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/Procedure.java
@@ -628,9 +628,7 @@ public abstract class Procedure<TEnvironment> implements Comparable<Procedure<TE
     return timeout != NO_TIMEOUT;
   }
 
-  /**
-   * @return the timeout in msec
-   */
+  /** Returns the timeout in msec */
   public int getTimeout() {
     return timeout;
   }
@@ -665,16 +663,12 @@ public abstract class Procedure<TEnvironment> implements Comparable<Procedure<TE
   // ==========================================================================
   // runtime state
   // ==========================================================================
-  /**
-   * @return the time elapsed between the last update and the start time of the procedure.
-   */
+  /** Returns the time elapsed between the last update and the start time of the procedure. */
   public long elapsedTime() {
     return getLastUpdate() - getSubmittedTime();
   }
 
-  /**
-   * @return the serialized result if any, otherwise null
-   */
+  /** Returns the serialized result if any, otherwise null */
   public byte[] getResult() {
     return result;
   }
@@ -717,9 +711,7 @@ public abstract class Procedure<TEnvironment> implements Comparable<Procedure<TE
   // just because the procedure can get scheduled on different executor threads on each step.
   // ==============================================================================================
 
-  /**
-   * @return true if the procedure is in a RUNNABLE state.
-   */
+  /** Returns true if the procedure is in a RUNNABLE state. */
   public synchronized boolean isRunnable() {
     return state == ProcedureState.RUNNABLE;
   }
@@ -728,16 +720,12 @@ public abstract class Procedure<TEnvironment> implements Comparable<Procedure<TE
     return state == ProcedureState.INITIALIZING;
   }
 
-  /**
-   * @return true if the procedure has failed. It may or may not have rolled back.
-   */
+  /** Returns true if the procedure has failed. It may or may not have rolled back. */
   public synchronized boolean isFailed() {
     return state == ProcedureState.FAILED || state == ProcedureState.ROLLEDBACK;
   }
 
-  /**
-   * @return true if the procedure is finished successfully.
-   */
+  /** Returns true if the procedure is finished successfully. */
   public synchronized boolean isSuccess() {
     return state == ProcedureState.SUCCESS && !hasException();
   }
@@ -750,9 +738,7 @@ public abstract class Procedure<TEnvironment> implements Comparable<Procedure<TE
     return isSuccess() || state == ProcedureState.ROLLEDBACK;
   }
 
-  /**
-   * @return true if the procedure is waiting for a child to finish or for an external event.
-   */
+  /** Returns true if the procedure is waiting for a child to finish or for an external event. */
   public synchronized boolean isWaiting() {
     switch (state) {
       case WAITING:
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureEvent.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureEvent.java
index e55f73211d7..ab1929fb8ce 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureEvent.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureEvent.java
@@ -42,7 +42,7 @@ public class ProcedureEvent<T> {
   }
 
   /**
-   * @return true if event is not ready and adds procedure to suspended queue, else returns false.
+   * Returns true if event is not ready and adds procedure to suspended queue, else returns false.
    */
   public synchronized boolean suspendIfNotReady(Procedure proc) {
     if (!ready) {
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.java
index 530e5305ba2..4e7c1ca04a7 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.java
@@ -680,16 +680,12 @@ public class ProcedureExecutor<TEnvironment> {
     return running.get();
   }
 
-  /**
-   * @return the current number of worker threads.
-   */
+  /** Returns the current number of worker threads. */
   public int getWorkerThreadCount() {
     return workerThreads.size();
   }
 
-  /**
-   * @return the core pool size settings.
-   */
+  /** Returns the core pool size settings. */
   public int getCorePoolSize() {
     return corePoolSize;
   }
@@ -2004,9 +2000,7 @@ public class ProcedureExecutor<TEnvironment> {
       return getName() + "(pid=" + (p == null ? Procedure.NO_PROC_ID : p.getProcId() + ")");
     }
 
-    /**
-     * @return the time since the current procedure is running
-     */
+    /** Returns the time since the current procedure is running */
     public long getCurrentRunTime() {
       return EnvironmentEdgeManager.currentTime() - executionStartTime.get();
     }
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureMetrics.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureMetrics.java
index f86a2b2d00a..bc82145cfe2 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureMetrics.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureMetrics.java
@@ -33,18 +33,14 @@ import org.apache.yetus.audience.InterfaceAudience;
  */
 @InterfaceAudience.Private
 public interface ProcedureMetrics {
-  /**
-   * @return Total number of instances submitted for a type of a procedure
-   */
+  /** Returns Total number of instances submitted for a type of a procedure */
   Counter getSubmittedCounter();
 
   /**
-   * @return Histogram of runtimes for all successfully completed instances of a type of a procedure
+   * Returns Histogram of runtimes for all successfully completed instances of a type of a procedure
    */
   Histogram getTimeHisto();
 
-  /**
-   * @return Total number of instances failed for a type of a procedure
-   */
+  /** Returns Total number of instances failed for a type of a procedure */
   Counter getFailedCounter();
 }
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureScheduler.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureScheduler.java
index f89cac5137c..c0ef8501f4c 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureScheduler.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureScheduler.java
@@ -87,9 +87,7 @@ public interface ProcedureScheduler {
    */
   void completionCleanup(Procedure proc);
 
-  /**
-   * @return true if there are procedures available to process, otherwise false.
-   */
+  /** Returns true if there are procedures available to process, otherwise false. */
   boolean hasRunnables();
 
   /**
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/ProcedureStore.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/ProcedureStore.java
index 7e54cfa9a12..386886d0c8c 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/ProcedureStore.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/ProcedureStore.java
@@ -165,14 +165,10 @@ public interface ProcedureStore {
    */
   void stop(boolean abort);
 
-  /**
-   * @return true if the store is running, otherwise false.
-   */
+  /** Returns true if the store is running, otherwise false. */
   boolean isRunning();
 
-  /**
-   * @return the number of threads/slots passed to start()
-   */
+  /** Returns the number of threads/slots passed to start() */
   int getNumThreads();
 
   /**
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/BitSetNode.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/BitSetNode.java
index fdbf183b19f..e97fe9bfc84 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/BitSetNode.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/BitSetNode.java
@@ -202,9 +202,7 @@ class BitSetNode {
     return (modified[wordIndex] & (1L << bitmapIndex)) != 0;
   }
 
-  /**
-   * @return true, if all the procedures has been modified.
-   */
+  /** Returns true, if all the procedures has been modified. */
   public boolean isAllModified() {
     // TODO: cache the value
     for (int i = 0; i < modified.length; ++i) {
@@ -215,9 +213,7 @@ class BitSetNode {
     return true;
   }
 
-  /**
-   * @return all the active procedure ids in this bit set.
-   */
+  /** Returns all the active procedure ids in this bit set. */
   public long[] getActiveProcIds() {
     List<Long> procIds = new ArrayList<>();
     for (int wordIndex = 0; wordIndex < modified.length; wordIndex++) {
@@ -236,9 +232,7 @@ class BitSetNode {
     return procIds.stream().mapToLong(Long::longValue).toArray();
   }
 
-  /**
-   * @return true, if there are no active procedures in this BitSetNode, else false.
-   */
+  /** Returns true, if there are no active procedures in this BitSetNode, else false. */
   public boolean isEmpty() {
     // TODO: cache the value
     for (int i = 0; i < deleted.length; ++i) {
@@ -429,16 +423,12 @@ class BitSetNode {
   // ========================================================================
   // Helpers
   // ========================================================================
-  /**
-   * @return upper boundary (aligned to multiple of BITS_PER_WORD) of bitmap range x belongs to.
-   */
+  /** Returns upper boundary (aligned to multiple of BITS_PER_WORD) of bitmap range x belongs to. */
   private static long alignUp(final long x) {
     return (x + (BITS_PER_WORD - 1)) & -BITS_PER_WORD;
   }
 
-  /**
-   * @return lower boundary (aligned to multiple of BITS_PER_WORD) of bitmap range x belongs to.
-   */
+  /** Returns lower boundary (aligned to multiple of BITS_PER_WORD) of bitmap range x belongs to. */
   private static long alignDown(final long x) {
     return x & -BITS_PER_WORD;
   }
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureStoreTracker.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureStoreTracker.java
index eeb5a3a827b..692333ff733 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureStoreTracker.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureStoreTracker.java
@@ -343,9 +343,7 @@ class ProcedureStoreTracker {
     this.partial = isPartial;
   }
 
-  /**
-   * @return true, if no procedure is active, else false.
-   */
+  /** Returns true, if no procedure is active, else false. */
   public boolean isEmpty() {
     for (Map.Entry<Long, BitSetNode> entry : map.entrySet()) {
       if (!entry.getValue().isEmpty()) {
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/util/DelayedUtil.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/util/DelayedUtil.java
index 19c4f8bcb58..32232ac18e6 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/util/DelayedUtil.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/util/DelayedUtil.java
@@ -74,9 +74,7 @@ public final class DelayedUtil {
     }
   };
 
-  /**
-   * @return null (if an interrupt) or an instance of E; resets interrupt on calling thread.
-   */
+  /** Returns null (if an interrupt) or an instance of E; resets interrupt on calling thread. */
   public static <E extends Delayed> E takeWithoutInterrupt(final DelayQueue<E> queue,
     final long timeout, final TimeUnit timeUnit) {
     try {
@@ -87,9 +85,7 @@ public final class DelayedUtil {
     }
   }
 
-  /**
-   * @return Time remaining as milliseconds.
-   */
+  /** Returns Time remaining as milliseconds. */
   public static long getRemainingTime(final TimeUnit resultUnit, final long timeout) {
     final long currentTime = EnvironmentEdgeManager.currentTime();
     if (currentTime >= timeout) {
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/ProtobufMessageHandler.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/ProtobufMessageHandler.java
index f90354b7ab7..39a7ba71dd6 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/ProtobufMessageHandler.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/ProtobufMessageHandler.java
@@ -26,9 +26,7 @@ import org.apache.yetus.audience.InterfaceAudience;
  */
 @InterfaceAudience.Private
 public interface ProtobufMessageHandler {
-  /**
-   * @return the protobuf represention of the model
-   */
+  /** Returns the protobuf represention of the model */
   byte[] createProtobufOutput();
 
   /**
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/RESTServlet.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/RESTServlet.java
index 10b96ec9284..66834f36176 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/RESTServlet.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/RESTServlet.java
@@ -52,17 +52,13 @@ public class RESTServlet implements Constants {
     return realUser;
   }
 
-  /**
-   * @return the RESTServlet singleton instance
-   */
+  /** Returns the RESTServlet singleton instance */
   public synchronized static RESTServlet getInstance() {
     assert (INSTANCE != null);
     return INSTANCE;
   }
 
-  /**
-   * @return the ConnectionCache instance
-   */
+  /** Returns the ConnectionCache instance */
   public ConnectionCache getConnectionCache() {
     return connectionCache;
   }
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java
index c10ad88b4b7..c40165c8254 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java
@@ -53,14 +53,12 @@ public class TableResource extends ResourceBase {
     this.table = table;
   }
 
-  /** @return the table name */
+  /** Returns the table name */
   String getName() {
     return table;
   }
 
-  /**
-   * @return true if the table exists n
-   */
+  /** Returns true if the table exists n */
   boolean exists() throws IOException {
     return servlet.getAdmin().tableExists(TableName.valueOf(table));
   }
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/Client.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/Client.java
index ff1fe141bb8..85cb2af86a8 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/Client.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/Client.java
@@ -212,9 +212,7 @@ public class Client {
   public void shutdown() {
   }
 
-  /**
-   * @return the wrapped HttpClient
-   */
+  /** Returns the wrapped HttpClient */
   public HttpClient getHttpClient() {
     return httpClient;
   }
@@ -394,9 +392,7 @@ public class Client {
     }
   }
 
-  /**
-   * @return the cluster definition
-   */
+  /** Returns the cluster definition */
   public Cluster getCluster() {
     return cluster;
   }
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/Cluster.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/Cluster.java
index dbb30adbc74..9070dfd2800 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/Cluster.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/Cluster.java
@@ -45,9 +45,7 @@ public class Cluster {
     this.nodes.addAll(nodes);
   }
 
-  /**
-   * @return true if no locations have been added, false otherwise
-   */
+  /** Returns true if no locations have been added, false otherwise */
   public boolean isEmpty() {
     return nodes.isEmpty();
   }
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/Response.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/Response.java
index cd0ac33dd79..7041d88ce30 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/Response.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/Response.java
@@ -82,9 +82,7 @@ public class Response {
     this.stream = in;
   }
 
-  /**
-   * @return the HTTP response code
-   */
+  /** Returns the HTTP response code */
   public int getCode() {
     return code;
   }
@@ -97,9 +95,7 @@ public class Response {
     return this.stream;
   }
 
-  /**
-   * @return the HTTP response headers
-   */
+  /** Returns the HTTP response headers */
   public Header[] getHeaders() {
     return headers;
   }
@@ -113,23 +109,17 @@ public class Response {
     return null;
   }
 
-  /**
-   * @return the value of the Location header
-   */
+  /** Returns the value of the Location header */
   public String getLocation() {
     return getHeader("Location");
   }
 
-  /**
-   * @return true if a response body was sent
-   */
+  /** Returns true if a response body was sent */
   public boolean hasBody() {
     return body != null;
   }
 
-  /**
-   * @return the HTTP response body
-   */
+  /** Returns the HTTP response body */
   public byte[] getBody() {
     if (body == null) {
       try {
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/CellModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/CellModel.java
index ae64a009289..6e2391546bb 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/CellModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/CellModel.java
@@ -119,9 +119,7 @@ public class CellModel implements ProtobufMessageHandler, Serializable {
     this.value = value;
   }
 
-  /**
-   * @return the column
-   */
+  /** Returns the column */
   public byte[] getColumn() {
     return column;
   }
@@ -133,16 +131,12 @@ public class CellModel implements ProtobufMessageHandler, Serializable {
     this.column = column;
   }
 
-  /**
-   * @return true if the timestamp property has been specified by the user
-   */
+  /** Returns true if the timestamp property has been specified by the user */
   public boolean hasUserTimestamp() {
     return timestamp != HConstants.LATEST_TIMESTAMP;
   }
 
-  /**
-   * @return the timestamp
-   */
+  /** Returns the timestamp */
   public long getTimestamp() {
     return timestamp;
   }
@@ -154,9 +148,7 @@ public class CellModel implements ProtobufMessageHandler, Serializable {
     this.timestamp = timestamp;
   }
 
-  /**
-   * @return the value
-   */
+  /** Returns the value */
   public byte[] getValue() {
     return value;
   }
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/CellSetModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/CellSetModel.java
index d962c71d314..7d99ea6219a 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/CellSetModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/CellSetModel.java
@@ -98,9 +98,7 @@ public class CellSetModel implements Serializable, ProtobufMessageHandler {
     rows.add(row);
   }
 
-  /**
-   * @return the rows
-   */
+  /** Returns the rows */
   public List<RowModel> getRows() {
     return rows;
   }
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/ColumnSchemaModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/ColumnSchemaModel.java
index c15a96d2d67..7712641e220 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/ColumnSchemaModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/ColumnSchemaModel.java
@@ -80,17 +80,13 @@ public class ColumnSchemaModel implements Serializable {
     return o != null ? o.toString() : null;
   }
 
-  /**
-   * @return the column name
-   */
+  /** Returns the column name */
   @XmlAttribute
   public String getName() {
     return name;
   }
 
-  /**
-   * @return the map for holding unspecified (user) attributes
-   */
+  /** Returns the map for holding unspecified (user) attributes */
   @XmlAnyAttribute
   @JsonAnyGetter
   public Map<QName, Object> getAny() {
@@ -130,57 +126,43 @@ public class ColumnSchemaModel implements Serializable {
   // cannot be standard bean type getters and setters, otherwise this would
   // confuse JAXB
 
-  /**
-   * @return true if the BLOCKCACHE attribute is present and true
-   */
+  /** Returns true if the BLOCKCACHE attribute is present and true */
   public boolean __getBlockcache() {
     Object o = attrs.get(BLOCKCACHE);
     return o != null ? Boolean.parseBoolean(o.toString()) : HColumnDescriptor.DEFAULT_BLOCKCACHE;
   }
 
-  /**
-   * @return the value of the BLOCKSIZE attribute or its default if it is unset
-   */
+  /** Returns the value of the BLOCKSIZE attribute or its default if it is unset */
   public int __getBlocksize() {
     Object o = attrs.get(BLOCKSIZE);
     return o != null ? Integer.parseInt(o.toString()) : HColumnDescriptor.DEFAULT_BLOCKSIZE;
   }
 
-  /**
-   * @return the value of the BLOOMFILTER attribute or its default if unset
-   */
+  /** Returns the value of the BLOOMFILTER attribute or its default if unset */
   public String __getBloomfilter() {
     Object o = attrs.get(BLOOMFILTER);
     return o != null ? o.toString() : HColumnDescriptor.DEFAULT_BLOOMFILTER;
   }
 
-  /**
-   * @return the value of the COMPRESSION attribute or its default if unset
-   */
+  /** Returns the value of the COMPRESSION attribute or its default if unset */
   public String __getCompression() {
     Object o = attrs.get(COMPRESSION);
     return o != null ? o.toString() : HColumnDescriptor.DEFAULT_COMPRESSION;
   }
 
-  /**
-   * @return true if the IN_MEMORY attribute is present and true
-   */
+  /** Returns true if the IN_MEMORY attribute is present and true */
   public boolean __getInMemory() {
     Object o = attrs.get(IN_MEMORY);
     return o != null ? Boolean.parseBoolean(o.toString()) : HColumnDescriptor.DEFAULT_IN_MEMORY;
   }
 
-  /**
-   * @return the value of the TTL attribute or its default if it is unset
-   */
+  /** Returns the value of the TTL attribute or its default if it is unset */
   public int __getTTL() {
     Object o = attrs.get(TTL);
     return o != null ? Integer.parseInt(o.toString()) : HColumnDescriptor.DEFAULT_TTL;
   }
 
-  /**
-   * @return the value of the VERSIONS attribute or its default if it is unset
-   */
+  /** Returns the value of the VERSIONS attribute or its default if it is unset */
   public int __getVersions() {
     Object o = attrs.get(VERSIONS);
     return o != null ? Integer.parseInt(o.toString()) : HColumnDescriptor.DEFAULT_VERSIONS;
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/NamespacesInstanceModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/NamespacesInstanceModel.java
index aeeff20b041..475f9185976 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/NamespacesInstanceModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/NamespacesInstanceModel.java
@@ -103,9 +103,7 @@ public class NamespacesInstanceModel implements Serializable, ProtobufMessageHan
     properties.put(key, value);
   }
 
-  /**
-   * @return The map of uncategorized namespace properties.
-   */
+  /** Returns The map of uncategorized namespace properties. */
   public Map<String, String> getProperties() {
     if (properties == null) {
       properties = new HashMap<>();
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/NamespacesModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/NamespacesModel.java
index cfd94d1ebc4..3b0e2d1e58b 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/NamespacesModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/NamespacesModel.java
@@ -67,9 +67,7 @@ public class NamespacesModel implements Serializable, ProtobufMessageHandler {
     }
   }
 
-  /**
-   * @return all namespaces
-   */
+  /** Returns all namespaces */
   public List<String> getNamespaces() {
     return namespaces;
   }
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/RowModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/RowModel.java
index 34f065105d6..f3e892aba0c 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/RowModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/RowModel.java
@@ -112,9 +112,7 @@ public class RowModel implements ProtobufMessageHandler, Serializable {
     cells.add(cell);
   }
 
-  /**
-   * @return the row key
-   */
+  /** Returns the row key */
   public byte[] getKey() {
     return key;
   }
@@ -126,9 +124,7 @@ public class RowModel implements ProtobufMessageHandler, Serializable {
     this.key = key;
   }
 
-  /**
-   * @return the cells
-   */
+  /** Returns the cells */
   public List<CellModel> getCells() {
     return cells;
   }
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/ScannerModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/ScannerModel.java
index fc59d14d56c..2eba8d4cad8 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/ScannerModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/ScannerModel.java
@@ -644,39 +644,29 @@ public class ScannerModel implements ProtobufMessageHandler, Serializable {
     labels.add(label);
   }
 
-  /**
-   * @return true if a start row was specified
-   */
+  /** Returns true if a start row was specified */
   public boolean hasStartRow() {
     return !Bytes.equals(startRow, HConstants.EMPTY_START_ROW);
   }
 
-  /**
-   * @return start row
-   */
+  /** Returns start row */
   @XmlAttribute
   public byte[] getStartRow() {
     return startRow;
   }
 
-  /**
-   * @return true if an end row was specified
-   */
+  /** Returns true if an end row was specified */
   public boolean hasEndRow() {
     return !Bytes.equals(endRow, HConstants.EMPTY_END_ROW);
   }
 
-  /**
-   * @return end row
-   */
+  /** Returns end row */
   @XmlAttribute
   public byte[] getEndRow() {
     return endRow;
   }
 
-  /**
-   * @return list of columns of interest in column:qualifier format, or empty for all
-   */
+  /** Returns list of columns of interest in column:qualifier format, or empty for all */
   @XmlElement(name = "column")
   public List<byte[]> getColumns() {
     return columns;
@@ -687,57 +677,43 @@ public class ScannerModel implements ProtobufMessageHandler, Serializable {
     return labels;
   }
 
-  /**
-   * @return the number of cells to return in batch
-   */
+  /** Returns the number of cells to return in batch */
   @XmlAttribute
   public int getBatch() {
     return batch;
   }
 
-  /**
-   * @return the number of rows that the scanner to fetch at once
-   */
+  /** Returns the number of rows that the scanner to fetch at once */
   @XmlAttribute
   public int getCaching() {
     return caching;
   }
 
-  /**
-   * @return true if HFile blocks should be cached on the servers for this scan, false otherwise
-   */
+  /** Returns true if HFile blocks should be cached on the servers for this scan, false otherwise */
   @XmlAttribute
   public boolean getCacheBlocks() {
     return cacheBlocks;
   }
 
-  /**
-   * @return the lower bound on timestamps of items of interest
-   */
+  /** Returns the lower bound on timestamps of items of interest */
   @XmlAttribute
   public long getStartTime() {
     return startTime;
   }
 
-  /**
-   * @return the upper bound on timestamps of items of interest
-   */
+  /** Returns the upper bound on timestamps of items of interest */
   @XmlAttribute
   public long getEndTime() {
     return endTime;
   }
 
-  /**
-   * @return maximum number of versions to return
-   */
+  /** Returns maximum number of versions to return */
   @XmlAttribute
   public int getMaxVersions() {
     return maxVersions;
   }
 
-  /**
-   * @return the filter specification
-   */
+  /** Returns the filter specification */
   @XmlElement
   public String getFilter() {
     return filter;
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterStatusModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterStatusModel.java
index 46bfb9091c3..027534c2888 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterStatusModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterStatusModel.java
@@ -162,105 +162,79 @@ public class StorageClusterStatusModel implements Serializable, ProtobufMessageH
         this.currentCompactedKVs = currentCompactedKVs;
       }
 
-      /**
-       * @return the region name
-       */
+      /** Returns the region name */
       @XmlAttribute
       public byte[] getName() {
         return name;
       }
 
-      /**
-       * @return the number of stores
-       */
+      /** Returns the number of stores */
       @XmlAttribute
       public int getStores() {
         return stores;
       }
 
-      /**
-       * @return the number of store files
-       */
+      /** Returns the number of store files */
       @XmlAttribute
       public int getStorefiles() {
         return storefiles;
       }
 
-      /**
-       * @return the total size of store files, in MB
-       */
+      /** Returns the total size of store files, in MB */
       @XmlAttribute
       public int getStorefileSizeMB() {
         return storefileSizeMB;
       }
 
-      /**
-       * @return memstore size, in MB
-       */
+      /** Returns memstore size, in MB */
       @XmlAttribute
       public int getMemStoreSizeMB() {
         return memstoreSizeMB;
       }
 
-      /**
-       * @return the total size of store file indexes, in KB
-       */
+      /** Returns the total size of store file indexes, in KB */
       @XmlAttribute
       public long getStorefileIndexSizeKB() {
         return storefileIndexSizeKB;
       }
 
-      /**
-       * @return the current total read requests made to region
-       */
+      /** Returns the current total read requests made to region */
       @XmlAttribute
       public long getReadRequestsCount() {
         return readRequestsCount;
       }
 
-      /**
-       * @return the current total write requests made to region
-       */
+      /** Returns the current total write requests made to region */
       @XmlAttribute
       public long getWriteRequestsCount() {
         return writeRequestsCount;
       }
 
-      /**
-       * @return The current total size of root-level indexes for the region, in KB.
-       */
+      /** Returns The current total size of root-level indexes for the region, in KB. */
       @XmlAttribute
       public int getRootIndexSizeKB() {
         return rootIndexSizeKB;
       }
 
-      /**
-       * @return The total size of static index, in KB
-       */
+      /** Returns The total size of static index, in KB */
       @XmlAttribute
       public int getTotalStaticIndexSizeKB() {
         return totalStaticIndexSizeKB;
       }
 
-      /**
-       * @return The total size of static bloom, in KB
-       */
+      /** Returns The total size of static bloom, in KB */
       @XmlAttribute
       public int getTotalStaticBloomSizeKB() {
         return totalStaticBloomSizeKB;
       }
 
-      /**
-       * @return The total number of compacting key-values
-       */
+      /** Returns The total number of compacting key-values */
       @XmlAttribute
       public long getTotalCompactingKVs() {
         return totalCompactingKVs;
       }
 
-      /**
-       * @return The number of current compacted key-values
-       */
+      /** Returns The number of current compacted key-values */
       @XmlAttribute
       public long getCurrentCompactedKVs() {
         return currentCompactedKVs;
@@ -405,49 +379,37 @@ public class StorageClusterStatusModel implements Serializable, ProtobufMessageH
       this.startCode = startCode;
     }
 
-    /**
-     * @return the region server's name
-     */
+    /** Returns the region server's name */
     @XmlAttribute
     public String getName() {
       return name;
     }
 
-    /**
-     * @return the region server's start code
-     */
+    /** Returns the region server's start code */
     @XmlAttribute
     public long getStartCode() {
       return startCode;
     }
 
-    /**
-     * @return the current heap size, in MB
-     */
+    /** Returns the current heap size, in MB */
     @XmlAttribute
     public int getHeapSizeMB() {
       return heapSizeMB;
     }
 
-    /**
-     * @return the maximum heap size, in MB
-     */
+    /** Returns the maximum heap size, in MB */
     @XmlAttribute
     public int getMaxHeapSizeMB() {
       return maxHeapSizeMB;
     }
 
-    /**
-     * @return the list of regions served by the region server
-     */
+    /** Returns the list of regions served by the region server */
     @XmlElement(name = "Region")
     public List<Region> getRegions() {
       return regions;
     }
 
-    /**
-     * @return the number of requests per second processed by the region server
-     */
+    /** Returns the number of requests per second processed by the region server */
     @XmlAttribute
     public long getRequests() {
       return requests;
@@ -547,9 +509,7 @@ public class StorageClusterStatusModel implements Serializable, ProtobufMessageH
   public StorageClusterStatusModel() {
   }
 
-  /**
-   * @return the list of live nodes
-   */
+  /** Returns the list of live nodes */
   @XmlElement(name = "Node")
   @XmlElementWrapper(name = "LiveNodes")
   // workaround https://github.com/FasterXML/jackson-dataformat-xml/issues/192
@@ -558,9 +518,7 @@ public class StorageClusterStatusModel implements Serializable, ProtobufMessageH
     return liveNodes;
   }
 
-  /**
-   * @return the list of dead nodes
-   */
+  /** Returns the list of dead nodes */
   @XmlElement(name = "Node")
   @XmlElementWrapper(name = "DeadNodes")
   // workaround https://github.com/FasterXML/jackson-dataformat-xml/issues/192
@@ -569,9 +527,7 @@ public class StorageClusterStatusModel implements Serializable, ProtobufMessageH
     return deadNodes;
   }
 
-  /**
-   * @return the total number of regions served by the cluster
-   */
+  /** Returns the total number of regions served by the cluster */
   @XmlAttribute
   public int getRegions() {
     return regions;
@@ -586,9 +542,7 @@ public class StorageClusterStatusModel implements Serializable, ProtobufMessageH
     return requests;
   }
 
-  /**
-   * @return the average load of the region servers in the cluster
-   */
+  /** Returns the average load of the region servers in the cluster */
   @XmlAttribute
   public double getAverageLoad() {
     return averageLoad;
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterVersionModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterVersionModel.java
index e23c6af8177..0125e321f75 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterVersionModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterVersionModel.java
@@ -38,9 +38,7 @@ public class StorageClusterVersionModel implements Serializable {
 
   private String version;
 
-  /**
-   * @return the storage cluster version
-   */
+  /** Returns the storage cluster version */
   @XmlAttribute(name = "Version")
   public String getVersion() {
     return version;
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableInfoModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableInfoModel.java
index 6a73c9227a5..37f7f79f89f 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableInfoModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableInfoModel.java
@@ -80,17 +80,13 @@ public class TableInfoModel implements Serializable, ProtobufMessageHandler {
     return regions.get(index);
   }
 
-  /**
-   * @return the table name
-   */
+  /** Returns the table name */
   @XmlAttribute
   public String getName() {
     return name;
   }
 
-  /**
-   * @return the regions
-   */
+  /** Returns the regions */
   @XmlElement(name = "Region")
   public List<TableRegionModel> getRegions() {
     return regions;
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableListModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableListModel.java
index 7118d159c2e..0b7ea10ab40 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableListModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableListModel.java
@@ -61,9 +61,7 @@ public class TableListModel implements Serializable, ProtobufMessageHandler {
     return tables.get(index);
   }
 
-  /**
-   * @return the tables
-   */
+  /** Returns the tables */
   @XmlElementRef(name = "table")
   public List<TableModel> getTables() {
     return tables;
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableModel.java
index 75513eadbaa..51a2bc567cd 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableModel.java
@@ -55,9 +55,7 @@ public class TableModel implements Serializable {
     this.name = name;
   }
 
-  /**
-   * @return the name
-   */
+  /** Returns the name */
   @XmlAttribute
   public String getName() {
     return name;
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableRegionModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableRegionModel.java
index 63d955a9929..b0d4b2f47a6 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableRegionModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableRegionModel.java
@@ -83,9 +83,7 @@ public class TableRegionModel implements Serializable {
     this.location = location;
   }
 
-  /**
-   * @return the region name
-   */
+  /** Returns the region name */
   @XmlAttribute
   public String getName() {
     byte[] tableNameAsBytes = Bytes.toBytes(this.table);
@@ -95,33 +93,25 @@ public class TableRegionModel implements Serializable {
     return Bytes.toString(nameAsBytes);
   }
 
-  /**
-   * @return the encoded region id
-   */
+  /** Returns the encoded region id */
   @XmlAttribute
   public long getId() {
     return id;
   }
 
-  /**
-   * @return the start key
-   */
+  /** Returns the start key */
   @XmlAttribute
   public byte[] getStartKey() {
     return startKey;
   }
 
-  /**
-   * @return the end key
-   */
+  /** Returns the end key */
   @XmlAttribute
   public byte[] getEndKey() {
     return endKey;
   }
 
-  /**
-   * @return the name and port of the region server hosting the region
-   */
+  /** Returns the name and port of the region server hosting the region */
   @XmlAttribute
   public String getLocation() {
     return location;
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableSchemaModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableSchemaModel.java
index 80896afa18f..1756db20d02 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableSchemaModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableSchemaModel.java
@@ -136,26 +136,20 @@ public class TableSchemaModel implements Serializable, ProtobufMessageHandler {
     return columns.get(index);
   }
 
-  /**
-   * @return the table name
-   */
+  /** Returns the table name */
   @XmlAttribute
   public String getName() {
     return name;
   }
 
-  /**
-   * @return the map for holding unspecified (user) attributes
-   */
+  /** Returns the map for holding unspecified (user) attributes */
   @XmlAnyAttribute
   @JsonAnyGetter
   public Map<QName, Object> getAny() {
     return attrs;
   }
 
-  /**
-   * @return the columns
-   */
+  /** Returns the columns */
   @XmlElement(name = "ColumnSchema")
   public List<ColumnSchemaModel> getColumns() {
     return columns;
@@ -211,25 +205,19 @@ public class TableSchemaModel implements Serializable, ProtobufMessageHandler {
   // cannot be standard bean type getters and setters, otherwise this would
   // confuse JAXB
 
-  /**
-   * @return true if IS_META attribute exists and is truel
-   */
+  /** Returns true if IS_META attribute exists and is truel */
   public boolean __getIsMeta() {
     Object o = attrs.get(IS_META);
     return o != null && Boolean.parseBoolean(o.toString());
   }
 
-  /**
-   * @return true if IS_ROOT attribute exists and is truel
-   */
+  /** Returns true if IS_ROOT attribute exists and is truel */
   public boolean __getIsRoot() {
     Object o = attrs.get(IS_ROOT);
     return o != null && Boolean.parseBoolean(o.toString());
   }
 
-  /**
-   * @return true if READONLY attribute exists and is truel
-   */
+  /** Returns true if READONLY attribute exists and is truel */
   public boolean __getReadOnly() {
     Object o = attrs.get(READONLY);
     return o != null ? Boolean.parseBoolean(o.toString()) : HTableDescriptor.DEFAULT_READONLY;
@@ -324,9 +312,7 @@ public class TableSchemaModel implements Serializable, ProtobufMessageHandler {
     return this;
   }
 
-  /**
-   * @return a table descriptor
-   */
+  /** Returns a table descriptor */
   @JsonIgnore
   public HTableDescriptor getTableDescriptor() {
     HTableDescriptor htd = new HTableDescriptor(TableName.valueOf(getName()));
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/VersionModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/VersionModel.java
index 41331070907..74bd722d36f 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/VersionModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/VersionModel.java
@@ -74,41 +74,31 @@ public class VersionModel implements Serializable, ProtobufMessageHandler {
     if (jerseyVersion == null) jerseyVersion = "";
   }
 
-  /**
-   * @return the REST gateway version
-   */
+  /** Returns the REST gateway version */
   @XmlAttribute(name = "REST")
   public String getRESTVersion() {
     return restVersion;
   }
 
-  /**
-   * @return the JVM vendor and version
-   */
+  /** Returns the JVM vendor and version */
   @XmlAttribute(name = "JVM")
   public String getJVMVersion() {
     return jvmVersion;
   }
 
-  /**
-   * @return the OS name, version, and hardware architecture
-   */
+  /** Returns the OS name, version, and hardware architecture */
   @XmlAttribute(name = "OS")
   public String getOSVersion() {
     return osVersion;
   }
 
-  /**
-   * @return the servlet container version
-   */
+  /** Returns the servlet container version */
   @XmlAttribute(name = "Server")
   public String getServerVersion() {
     return serverVersion;
   }
 
-  /**
-   * @return the version of the embedded Jersey framework
-   */
+  /** Returns the version of the embedded Jersey framework */
   @XmlAttribute(name = "Jersey")
   public String getJerseyVersion() {
     return jerseyVersion;
diff --git a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupAdminServer.java b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupAdminServer.java
index 16d562bbb9f..bf535188c5b 100644
--- a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupAdminServer.java
+++ b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupAdminServer.java
@@ -119,9 +119,7 @@ public class RSGroupAdminServer implements RSGroupAdmin {
     return rsGroupInfo;
   }
 
-  /**
-   * @return List of Regions associated with this <code>server</code>.
-   */
+  /** Returns List of Regions associated with this <code>server</code>. */
   private List<RegionInfo> getRegions(final Address server) {
     LinkedList<RegionInfo> regions = new LinkedList<>();
     for (Map.Entry<RegionInfo, ServerName> el : master.getAssignmentManager().getRegionStates()
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/HDFSBlocksDistribution.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/HDFSBlocksDistribution.java
index d4aa962f2b4..dcddb74b71b 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/HDFSBlocksDistribution.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/HDFSBlocksDistribution.java
@@ -71,23 +71,17 @@ public class HDFSBlocksDistribution {
       this.weightForSsd += weightForSsd;
     }
 
-    /**
-     * @return the host name
-     */
+    /** Returns the host name */
     public String getHost() {
       return host;
     }
 
-    /**
-     * @return the weight
-     */
+    /** Returns the weight */
     public long getWeight() {
       return weight;
     }
 
-    /**
-     * @return the weight for ssd
-     */
+    /** Returns the weight for ssd */
     public long getWeightForSsd() {
       return weightForSsd;
     }
@@ -186,9 +180,7 @@ public class HDFSBlocksDistribution {
     }
   }
 
-  /**
-   * @return the hosts and their weights
-   */
+  /** Returns the hosts and their weights */
   public Map<String, HostAndWeight> getHostAndWeights() {
     return this.hostAndWeights;
   }
@@ -209,9 +201,7 @@ public class HDFSBlocksDistribution {
     return weight;
   }
 
-  /**
-   * @return the sum of all unique blocks' weight
-   */
+  /** Returns the sum of all unique blocks' weight */
   public long getUniqueBlocksTotalWeight() {
     return uniqueBlocksTotalWeight;
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/LocalHBaseCluster.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/LocalHBaseCluster.java
index bc562b0a532..a803b6aee9d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/LocalHBaseCluster.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/LocalHBaseCluster.java
@@ -247,9 +247,7 @@ public class LocalHBaseCluster {
     return regionThreads.get(serverNumber).getRegionServer();
   }
 
-  /**
-   * @return Read-only list of region server threads.
-   */
+  /** Returns Read-only list of region server threads. */
   public List<JVMClusterUtil.RegionServerThread> getRegionServers() {
     return Collections.unmodifiableList(this.regionThreads);
   }
@@ -268,9 +266,7 @@ public class LocalHBaseCluster {
     return liveServers;
   }
 
-  /**
-   * @return the Configuration used by this LocalHBaseCluster
-   */
+  /** Returns the Configuration used by this LocalHBaseCluster */
   public Configuration getConfiguration() {
     return this.conf;
   }
@@ -306,9 +302,7 @@ public class LocalHBaseCluster {
     return rst.getName();
   }
 
-  /**
-   * @return the HMaster thread
-   */
+  /** Returns the HMaster thread */
   public HMaster getMaster(int serverNumber) {
     return masterThreads.get(serverNumber).getMaster();
   }
@@ -328,9 +322,7 @@ public class LocalHBaseCluster {
     return null;
   }
 
-  /**
-   * @return Read-only list of master threads.
-   */
+  /** Returns Read-only list of master threads. */
   public List<JVMClusterUtil.MasterThread> getMasters() {
     return Collections.unmodifiableList(this.masterThreads);
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/MetaRegionLocationCache.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/MetaRegionLocationCache.java
index c1ebb6fb9b4..6f539b3d173 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/MetaRegionLocationCache.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/MetaRegionLocationCache.java
@@ -201,9 +201,7 @@ public class MetaRegionLocationCache extends ZKListener {
     cachedMetaLocations.put(replicaId, location);
   }
 
-  /**
-   * @return Optional list of HRegionLocations for meta replica(s), null if the cache is empty.
-   */
+  /** Returns Optional list of HRegionLocations for meta replica(s), null if the cache is empty. */
   public Optional<List<HRegionLocation>> getMetaRegionLocations() {
     ConcurrentNavigableMap<Integer, HRegionLocation> snapshot =
       cachedMetaLocations.tailMap(cachedMetaLocations.firstKey());
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/Server.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/Server.java
index 0476a2acb7a..69ffd64e02d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/Server.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/Server.java
@@ -58,9 +58,7 @@ public interface Server extends Abortable, Stoppable {
    */
   ClusterConnection getClusterConnection();
 
-  /**
-   * @return The unique server name for this server.
-   */
+  /** Returns The unique server name for this server. */
   ServerName getServerName();
 
   /**
@@ -68,14 +66,10 @@ public interface Server extends Abortable, Stoppable {
    */
   CoordinatedStateManager getCoordinatedStateManager();
 
-  /**
-   * @return The {@link ChoreService} instance for this server
-   */
+  /** Returns The {@link ChoreService} instance for this server */
   ChoreService getChoreService();
 
-  /**
-   * @return Return the FileSystem object used (can return null!).
-   */
+  /** Returns Return the FileSystem object used (can return null!). */
   // TODO: Distinguish between "dataFs" and "walFs".
   default FileSystem getFileSystem() {
     // This default is pretty dodgy!
@@ -91,9 +85,7 @@ public interface Server extends Abortable, Stoppable {
     return fs;
   };
 
-  /**
-   * @return True is the server is Stopping
-   */
+  /** Returns True is the server is Stopping */
   // Note: This method is not part of the Stoppable Interface.
   default boolean isStopping() {
     return false;
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java
index 1dc17eff0d0..7ef131e9e6a 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java
@@ -35,9 +35,7 @@ public interface TableDescriptors {
     return get(tableName) != null;
   }
 
-  /**
-   * @return TableDescriptor for tablename
-   */
+  /** Returns TableDescriptor for tablename */
   TableDescriptor get(TableName tableName) throws IOException;
 
   /**
@@ -72,8 +70,6 @@ public interface TableDescriptors {
    */
   void update(TableDescriptor htd, boolean cacheOnly) throws IOException;
 
-  /**
-   * @return Instance of table descriptor or null if none found.
-   */
+  /** Returns Instance of table descriptor or null if none found. */
   TableDescriptor remove(TableName tablename) throws IOException;
 }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/HFileArchiver.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/HFileArchiver.java
index 8e666308d4f..68dc87502e0 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/HFileArchiver.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/HFileArchiver.java
@@ -80,9 +80,7 @@ public class HFileArchiver {
     // hidden ctor since this is just a util
   }
 
-  /**
-   * @return True if the Region exits in the filesystem.
-   */
+  /** Returns True if the Region exits in the filesystem. */
   public static boolean exists(Configuration conf, FileSystem fs, RegionInfo info)
     throws IOException {
     Path rootDir = CommonFSUtils.getRootDir(conf);
@@ -703,14 +701,10 @@ public class HFileArchiver {
      */
     abstract void close() throws IOException;
 
-    /**
-     * @return the name of the file (not the full fs path, just the individual file name)
-     */
+    /** Returns the name of the file (not the full fs path, just the individual file name) */
     abstract String getName();
 
-    /**
-     * @return the path to this file
-     */
+    /** Returns the path to this file */
     abstract Path getPath();
 
     /**
@@ -722,9 +716,7 @@ public class HFileArchiver {
       return CommonFSUtils.renameAndSetModifyTime(fs, p, dest);
     }
 
-    /**
-     * @return the {@link FileSystem} on which this file resides
-     */
+    /** Returns the {@link FileSystem} on which this file resides */
     public FileSystem getFileSystem() {
       return this.fs;
     }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/example/TableHFileArchiveTracker.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/example/TableHFileArchiveTracker.java
index 1896199ede2..9ae06f2840c 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/example/TableHFileArchiveTracker.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/example/TableHFileArchiveTracker.java
@@ -222,9 +222,7 @@ public final class TableHFileArchiveTracker extends ZKListener {
     return getMonitor().shouldArchiveTable(tableName);
   }
 
-  /**
-   * @return the tracker for which tables should be archived.
-   */
+  /** Returns the tracker for which tables should be archived. */
   public final HFileArchiveTableMonitor getMonitor() {
     return this.monitor;
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/client/VersionInfoUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/client/VersionInfoUtil.java
index ec01bc5b51b..e9703cb8cb2 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/client/VersionInfoUtil.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/client/VersionInfoUtil.java
@@ -65,9 +65,7 @@ public final class VersionInfoUtil {
     return false;
   }
 
-  /**
-   * @return the versionInfo extracted from the current RpcCallContext
-   */
+  /** Returns the versionInfo extracted from the current RpcCallContext */
   public static HBaseProtos.VersionInfo getCurrentClientVersionInfo() {
     return RpcServer.getCurrentCall().map(RpcCallContext::getClientVersionInfo).orElse(null);
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/client/locking/EntityLock.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/client/locking/EntityLock.java
index 50c6dc27657..747ccb54e83 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/client/locking/EntityLock.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/client/locking/EntityLock.java
@@ -208,9 +208,7 @@ public class EntityLock {
       setDaemon(true);
     }
 
-    /**
-     * @return Shuts down the thread clean and quietly.
-     */
+    /** Returns Shuts down the thread clean and quietly. */
     Thread shutdown() {
       shutdown = true;
       interrupt();
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/SplitLogManagerCoordination.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/SplitLogManagerCoordination.java
index 324a6f37b99..29aa273b2b3 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/SplitLogManagerCoordination.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/SplitLogManagerCoordination.java
@@ -56,30 +56,22 @@ public interface SplitLogManagerCoordination {
       this.failedDeletions = failedDeletions;
     }
 
-    /**
-     * @return the master value
-     */
+    /** Returns the master value */
     public MasterServices getMaster() {
       return master;
     }
 
-    /**
-     * @return map of tasks
-     */
+    /** Returns map of tasks */
     public ConcurrentMap<String, Task> getTasks() {
       return tasks;
     }
 
-    /**
-     * @return a set of failed deletions
-     */
+    /** Returns a set of failed deletions */
     public Set<String> getFailedDeletions() {
       return failedDeletions;
     }
 
-    /**
-     * @return server name
-     */
+    /** Returns server name */
     public ServerName getServerName() {
       return master.getServerName();
     }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/SplitLogWorkerCoordination.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/SplitLogWorkerCoordination.java
index ff2fbfbe7e7..3ab8ee83531 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/SplitLogWorkerCoordination.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/SplitLogWorkerCoordination.java
@@ -64,9 +64,7 @@ public interface SplitLogWorkerCoordination {
    */
   void stopProcessingTasks();
 
-  /**
-   * @return the current value of exitWorker
-   */
+  /** Returns the current value of exitWorker */
   boolean isStop();
 
   /**
@@ -123,9 +121,7 @@ public interface SplitLogWorkerCoordination {
    */
   interface SplitTaskDetails {
 
-    /**
-     * @return full file path in HDFS for the WAL file to be split.
-     */
+    /** Returns full file path in HDFS for the WAL file to be split. */
     String getWALFile();
   }
 }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/ZkSplitLogWorkerCoordination.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/ZkSplitLogWorkerCoordination.java
index 1b255b25e17..7acb0891dbc 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/ZkSplitLogWorkerCoordination.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/ZkSplitLogWorkerCoordination.java
@@ -317,9 +317,7 @@ public class ZkSplitLogWorkerCoordination extends ZKListener implements SplitLog
     server.getExecutorService().submit(hsh);
   }
 
-  /**
-   * @return true if more splitters are available, otherwise false.
-   */
+  /** Returns true if more splitters are available, otherwise false. */
   private boolean areSplittersAvailable() {
     return maxConcurrentTasks - tasksInProgress.get() > 0;
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseEnvironment.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseEnvironment.java
index b8cabe8cfed..b5f0809c046 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseEnvironment.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseEnvironment.java
@@ -118,13 +118,13 @@ public class BaseEnvironment<C extends Coprocessor> implements CoprocessorEnviro
     return seq;
   }
 
-  /** @return the coprocessor environment version */
+  /** Returns the coprocessor environment version */
   @Override
   public int getVersion() {
     return Coprocessor.VERSION;
   }
 
-  /** @return the HBase release */
+  /** Returns the HBase release */
   @Override
   public String getHBaseVersion() {
     return VersionInfo.getVersion();
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/HasMasterServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/HasMasterServices.java
index 2682b78fd51..dd8babf21dd 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/HasMasterServices.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/HasMasterServices.java
@@ -30,8 +30,6 @@ import org.apache.yetus.audience.InterfaceAudience;
 @Deprecated
 @InterfaceAudience.Private
 public interface HasMasterServices {
-  /**
-   * @return An instance of RegionServerServices, an object NOT for Coprocessor consumption.
-   */
+  /** Returns An instance of RegionServerServices, an object NOT for Coprocessor consumption. */
   MasterServices getMasterServices();
 }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/HasRegionServerServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/HasRegionServerServices.java
index cef03390acb..21301efeafa 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/HasRegionServerServices.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/HasRegionServerServices.java
@@ -30,8 +30,6 @@ import org.apache.yetus.audience.InterfaceAudience;
 @Deprecated
 @InterfaceAudience.Private
 public interface HasRegionServerServices {
-  /**
-   * @return An instance of RegionServerServices, an object NOT for Coprocessor consumption.
-   */
+  /** Returns An instance of RegionServerServices, an object NOT for Coprocessor consumption. */
   RegionServerServices getRegionServerServices();
 }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterCoprocessorEnvironment.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterCoprocessorEnvironment.java
index c83b9da4308..61c98346887 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterCoprocessorEnvironment.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterCoprocessorEnvironment.java
@@ -30,9 +30,7 @@ import org.apache.yetus.audience.InterfaceStability;
 @InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.COPROC)
 @InterfaceStability.Evolving
 public interface MasterCoprocessorEnvironment extends CoprocessorEnvironment<MasterCoprocessor> {
-  /**
-   * @return Hosting Server's ServerName
-   */
+  /** Returns Hosting Server's ServerName */
   ServerName getServerName();
 
   /**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.java
index 97d572d3727..8967f596379 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.java
@@ -36,23 +36,19 @@ import org.apache.yetus.audience.InterfaceStability;
 @InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.COPROC)
 @InterfaceStability.Evolving
 public interface RegionCoprocessorEnvironment extends CoprocessorEnvironment<RegionCoprocessor> {
-  /** @return the region associated with this coprocessor */
+  /** Returns the region associated with this coprocessor */
   Region getRegion();
 
-  /** @return region information for the region this coprocessor is running on */
+  /** Returns region information for the region this coprocessor is running on */
   RegionInfo getRegionInfo();
 
-  /**
-   * @return Interface to Map of regions online on this RegionServer {@link #getServerName()}}.
-   */
+  /** Returns Interface to Map of regions online on this RegionServer {@link #getServerName()}}. */
   OnlineRegions getOnlineRegions();
 
-  /** @return shared data between all instances of this coprocessor */
+  /** Returns shared data between all instances of this coprocessor */
   ConcurrentMap<String, Object> getSharedData();
 
-  /**
-   * @return Hosting Server's ServerName
-   */
+  /** Returns Hosting Server's ServerName */
   ServerName getServerName();
 
   /**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionServerCoprocessorEnvironment.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionServerCoprocessorEnvironment.java
index 4b2e8b5791b..39d186aba29 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionServerCoprocessorEnvironment.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionServerCoprocessorEnvironment.java
@@ -32,14 +32,10 @@ import org.apache.yetus.audience.InterfaceStability;
 @InterfaceStability.Evolving
 public interface RegionServerCoprocessorEnvironment
   extends CoprocessorEnvironment<RegionServerCoprocessor> {
-  /**
-   * @return Hosting Server's ServerName
-   */
+  /** Returns Hosting Server's ServerName */
   ServerName getServerName();
 
-  /**
-   * @return Interface to Map of regions online on this RegionServer {@link #getServerName()}}.
-   */
+  /** Returns Interface to Map of regions online on this RegionServer {@link #getServerName()}}. */
   OnlineRegions getOnlineRegions();
 
   /**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/WALCoprocessorEnvironment.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/WALCoprocessorEnvironment.java
index 1774481f210..2d414111db5 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/WALCoprocessorEnvironment.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/WALCoprocessorEnvironment.java
@@ -27,7 +27,7 @@ import org.apache.yetus.audience.InterfaceStability;
 @InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.COPROC)
 @InterfaceStability.Evolving
 public interface WALCoprocessorEnvironment extends CoprocessorEnvironment<WALCoprocessor> {
-  /** @return reference to the region server's WAL */
+  /** Returns reference to the region server's WAL */
   WAL getWAL();
 
   /**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/EventHandler.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/EventHandler.java
index fcb60a35c4c..94418f0c381 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/EventHandler.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/EventHandler.java
@@ -134,9 +134,7 @@ public abstract class EventHandler implements Runnable, Comparable<EventHandler>
     return Integer.MAX_VALUE;
   }
 
-  /**
-   * @return This events' sequence id.
-   */
+  /** Returns This events' sequence id. */
   public long getSeqid() {
     return this.seqid;
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/ExecutorService.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/ExecutorService.java
index eafbad1ecba..a60a1ed7512 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/ExecutorService.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/ExecutorService.java
@@ -214,7 +214,7 @@ public class ExecutorService {
     }
 
     /**
-     * @return the executor name inferred from the type and the servername on which this is running.
+     * Returns the executor name inferred from the type and the servername on which this is running.
      */
     public String getName() {
       return getExecutorType().getExecutorName(servername);
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/ExecutorType.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/ExecutorType.java
index a8c62818f15..320842985d4 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/ExecutorType.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/executor/ExecutorType.java
@@ -58,9 +58,7 @@ public enum ExecutorType {
   ExecutorType(int value) {
   }
 
-  /**
-   * @return Conflation of the executor type and the passed {@code serverName}.
-   */
+  /** Returns Conflation of the executor type and the passed {@code serverName}. */
   String getExecutorName(String serverName) {
     return this.toString() + "-" + serverName.replace("%", "%%");
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/favored/FavoredNodesPlan.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/favored/FavoredNodesPlan.java
index 9fbcf288e92..0f0304ab9db 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/favored/FavoredNodesPlan.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/favored/FavoredNodesPlan.java
@@ -71,9 +71,7 @@ public class FavoredNodesPlan {
     return favoredNodesMap.remove(region.getRegionNameAsString());
   }
 
-  /**
-   * @return the list of favored region server for this region based on the plan
-   */
+  /** Returns the list of favored region server for this region based on the plan */
   public List<ServerName> getFavoredNodes(RegionInfo region) {
     return favoredNodesMap.get(region.getRegionNameAsString());
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/filter/FilterWrapper.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/filter/FilterWrapper.java
index f2807ea023b..c31fc25ec82 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/filter/FilterWrapper.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/filter/FilterWrapper.java
@@ -45,9 +45,7 @@ final public class FilterWrapper extends Filter {
     this.filter = filter;
   }
 
-  /**
-   * @return The filter serialized using pb
-   */
+  /** Returns The filter serialized using pb */
   @Override
   public byte[] toByteArray() throws IOException {
     FilterProtos.FilterWrapper.Builder builder = FilterProtos.FilterWrapper.newBuilder();
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/FSDataInputStreamWrapper.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/FSDataInputStreamWrapper.java
index 819c5651081..3ad07622a4f 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/FSDataInputStreamWrapper.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/FSDataInputStreamWrapper.java
@@ -186,9 +186,7 @@ public class FSDataInputStreamWrapper implements Closeable {
     readahead = 0;
   }
 
-  /**
-   * @return Whether we are presently using HBase checksum.
-   */
+  /** Returns Whether we are presently using HBase checksum. */
   public boolean shouldUseHBaseChecksum() {
     return this.useHBaseChecksum;
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/FileLink.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/FileLink.java
index 2d12fd88c11..8fbaa775391 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/FileLink.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/FileLink.java
@@ -335,9 +335,7 @@ public class FileLink {
     this.locations = locations.toArray(new Path[locations.size()]);
   }
 
-  /**
-   * @return the locations to look for the linked file.
-   */
+  /** Returns the locations to look for the linked file. */
   public Path[] getLocations() {
     return locations;
   }
@@ -354,9 +352,7 @@ public class FileLink {
     return str.toString();
   }
 
-  /**
-   * @return true if the file pointed by the link exists
-   */
+  /** Returns true if the file pointed by the link exists */
   public boolean exists(final FileSystem fs) throws IOException {
     for (int i = 0; i < locations.length; ++i) {
       if (fs.exists(locations[i])) {
@@ -366,9 +362,7 @@ public class FileLink {
     return false;
   }
 
-  /**
-   * @return the path of the first available link.
-   */
+  /** Returns the path of the first available link. */
   public Path getAvailablePath(FileSystem fs) throws IOException {
     for (int i = 0; i < locations.length; ++i) {
       if (fs.exists(locations[i])) {
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/HFileLink.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/HFileLink.java
index 4cf350004a3..4184d6ca21e 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/HFileLink.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/HFileLink.java
@@ -160,23 +160,17 @@ public class HFileLink extends FileLink {
     return HFileLink.buildFromHFileLinkPattern(conf, createPath(table, region, family, hfile));
   }
 
-  /**
-   * @return the origin path of the hfile.
-   */
+  /** Returns the origin path of the hfile. */
   public Path getOriginPath() {
     return this.originPath;
   }
 
-  /**
-   * @return the path of the archived hfile.
-   */
+  /** Returns the path of the archived hfile. */
   public Path getArchivePath() {
     return this.archivePath;
   }
 
-  /**
-   * @return the path of the mob hfiles.
-   */
+  /** Returns the path of the mob hfiles. */
   public Path getMobPath() {
     return this.mobPath;
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java
index 7cbfad4b0ac..52eaa30317a 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java
@@ -118,15 +118,11 @@ public interface BlockCache extends Iterable<CachedBlock> {
    */
   long getDataBlockCount();
 
-  /**
-   * @return Iterator over the blocks in the cache.
-   */
+  /** Returns Iterator over the blocks in the cache. */
   @Override
   Iterator<CachedBlock> iterator();
 
-  /**
-   * @return The list of sub blockcaches that make up this one; returns null if no sub caches.
-   */
+  /** Returns The list of sub blockcaches that make up this one; returns null if no sub caches. */
   BlockCache[] getBlockCaches();
 
   /**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCacheKey.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCacheKey.java
index f9c52e4acbc..81da24f9341 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCacheKey.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCacheKey.java
@@ -82,9 +82,7 @@ public class BlockCacheKey implements HeapSize, java.io.Serializable {
   }
 
   // can't avoid this unfortunately
-  /**
-   * @return The hfileName portion of this cache key
-   */
+  /** Returns The hfileName portion of this cache key */
   public String getHfileName() {
     return hfileName;
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCacheUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCacheUtil.java
index 2a2faae3eae..1bad6c5023d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCacheUtil.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCacheUtil.java
@@ -116,9 +116,7 @@ public class BlockCacheUtil {
     }
   }
 
-  /**
-   * @return A JSON String of <code>filename</code> and counts of <code>blocks</code>
-   */
+  /** Returns A JSON String of <code>filename</code> and counts of <code>blocks</code> */
   public static String toJSON(String filename, NavigableSet<CachedBlock> blocks)
     throws IOException {
     CachedBlockCountsPerFile counts = new CachedBlockCountsPerFile(filename);
@@ -134,16 +132,12 @@ public class BlockCacheUtil {
     return GSON.toJson(counts);
   }
 
-  /**
-   * @return JSON string of <code>cbsf</code> aggregated
-   */
+  /** Returns JSON string of <code>cbsf</code> aggregated */
   public static String toJSON(CachedBlocksByFile cbsbf) throws IOException {
     return GSON.toJson(cbsbf);
   }
 
-  /**
-   * @return JSON string of <code>bc</code> content.
-   */
+  /** Returns JSON string of <code>bc</code> content. */
   public static String toJSON(BlockCache bc) throws IOException {
     return GSON.toJson(bc);
   }
@@ -318,9 +312,7 @@ public class BlockCacheUtil {
       return this.cachedBlockByFile;
     }
 
-    /**
-     * @return count of blocks in the cache
-     */
+    /** Returns count of blocks in the cache */
     public int getCount() {
       return count;
     }
@@ -329,16 +321,12 @@ public class BlockCacheUtil {
       return dataBlockCount;
     }
 
-    /**
-     * @return size of blocks in the cache
-     */
+    /** Returns size of blocks in the cache */
     public long getSize() {
       return size;
     }
 
-    /**
-     * @return Size of data.
-     */
+    /** Returns Size of data. */
     public long getDataSize() {
       return dataSize;
     }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheConfig.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheConfig.java
index 9e456afcebc..32c6dfd6c2a 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheConfig.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheConfig.java
@@ -255,9 +255,7 @@ public class CacheConfig {
       || (prefetchOnOpen && (category != BlockCategory.META && category != BlockCategory.UNKNOWN));
   }
 
-  /**
-   * @return true if blocks in this file should be flagged as in-memory
-   */
+  /** Returns true if blocks in this file should be flagged as in-memory */
   public boolean isInMemory() {
     return this.inMemory;
   }
@@ -320,15 +318,13 @@ public class CacheConfig {
     this.evictOnClose = evictOnClose;
   }
 
-  /**
-   * @return true if data blocks should be compressed in the cache, false if not
-   */
+  /** Returns true if data blocks should be compressed in the cache, false if not */
   public boolean shouldCacheDataCompressed() {
     return this.cacheDataOnRead && this.cacheDataCompressed;
   }
 
   /**
-   * @return true if this {@link BlockCategory} should be compressed in blockcache, false otherwise
+   * Returns true if this {@link BlockCategory} should be compressed in blockcache, false otherwise
    */
   public boolean shouldCacheCompressed(BlockCategory category) {
     switch (category) {
@@ -339,23 +335,17 @@ public class CacheConfig {
     }
   }
 
-  /**
-   * @return true if blocks should be prefetched into the cache on open, false if not
-   */
+  /** Returns true if blocks should be prefetched into the cache on open, false if not */
   public boolean shouldPrefetchOnOpen() {
     return this.prefetchOnOpen;
   }
 
-  /**
-   * @return true if blocks should be cached while writing during compaction, false if not
-   */
+  /** Returns true if blocks should be cached while writing during compaction, false if not */
   public boolean shouldCacheCompactedBlocksOnWrite() {
     return this.cacheCompactedDataOnWrite;
   }
 
-  /**
-   * @return total file size in bytes threshold for caching while writing during compaction
-   */
+  /** Returns total file size in bytes threshold for caching while writing during compaction */
   public long getCacheCompactedBlocksOnWriteThreshold() {
     return this.cacheCompactedDataOnWriteThreshold;
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/Cacheable.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/Cacheable.java
index 0a41b53a804..43456d7857d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/Cacheable.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/Cacheable.java
@@ -50,9 +50,7 @@ public interface Cacheable extends HeapSize, HBaseReferenceCounted {
    */
   CacheableDeserializer<Cacheable> getDeserializer();
 
-  /**
-   * @return the block type of this cached HFile block
-   */
+  /** Returns the block type of this cached HFile block */
   BlockType getBlockType();
 
   /******************************* ReferenceCounted Interfaces ***********************************/
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java
index ba7ff84996f..89a407b3389 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java
@@ -211,7 +211,7 @@ public final class HFile {
     /** Add an element to the file info map. */
     void appendFileInfo(byte[] key, byte[] value) throws IOException;
 
-    /** @return the path to this {@link HFile} */
+    /** Returns the path to this {@link HFile} */
     Path getPath();
 
     /**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java
index e9bc08f7d15..7bc2d689f87 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java
@@ -434,7 +434,7 @@ public class HFileBlock implements Cacheable {
     return this;
   }
 
-  /** @return get data block encoding id that was used to encode this block */
+  /** Returns get data block encoding id that was used to encode this block */
   short getDataBlockEncodingId() {
     if (blockType != BlockType.ENCODED_DATA) {
       throw new IllegalArgumentException("Querying encoder ID of a block " + "of type other than "
@@ -443,30 +443,22 @@ public class HFileBlock implements Cacheable {
     return buf.getShort(headerSize());
   }
 
-  /**
-   * @return the on-disk size of header + data part + checksum.
-   */
+  /** Returns the on-disk size of header + data part + checksum. */
   public int getOnDiskSizeWithHeader() {
     return onDiskSizeWithoutHeader + headerSize();
   }
 
-  /**
-   * @return the on-disk size of the data part + checksum (header excluded).
-   */
+  /** Returns the on-disk size of the data part + checksum (header excluded). */
   int getOnDiskSizeWithoutHeader() {
     return onDiskSizeWithoutHeader;
   }
 
-  /**
-   * @return the uncompressed size of data part (header and checksum excluded).
-   */
+  /** Returns the uncompressed size of data part (header and checksum excluded). */
   int getUncompressedSizeWithoutHeader() {
     return uncompressedSizeWithoutHeader;
   }
 
-  /**
-   * @return the offset of the previous block of the same type in the file, or -1 if unknown
-   */
+  /** Returns the offset of the previous block of the same type in the file, or -1 if unknown */
   long getPrevBlockOffset() {
     return prevBlockOffset;
   }
@@ -692,9 +684,7 @@ public class HFileBlock implements Cacheable {
     return offset;
   }
 
-  /**
-   * @return a byte stream reading the data + checksum of this block
-   */
+  /** Returns a byte stream reading the data + checksum of this block */
   DataInputStream getByteStream() {
     ByteBuff dup = this.buf.duplicate();
     dup.position(this.headerSize());
@@ -1081,7 +1071,7 @@ public class HFileBlock implements Cacheable {
       return baosInMemory.size();
     }
 
-    /** @return true if a block is being written */
+    /** Returns true if a block is being written */
     boolean isWriting() {
       return state == State.WRITING;
     }
@@ -1545,7 +1535,7 @@ public class HFileBlock implements Cacheable {
     }
 
     /**
-     * @return Check <code>onDiskSizeWithHeaderL</code> size is healthy and then return it as an int
+     * Returns Check <code>onDiskSizeWithHeaderL</code> size is healthy and then return it as an int
      */
     private static int checkAndGetSizeAsInt(final long onDiskSizeWithHeaderL, final int hdrSize)
       throws IOException {
@@ -1922,7 +1912,7 @@ public class HFileBlock implements Cacheable {
     return this.fileContext.getBytesPerChecksum();
   }
 
-  /** @return the size of data on disk + header. Excludes checksum. */
+  /** Returns the size of data on disk + header. Excludes checksum. */
   int getOnDiskDataSizeWithHeader() {
     return this.onDiskDataSizeWithHeader;
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java
index 782383d697b..0e886d9873c 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java
@@ -508,9 +508,7 @@ public class HFileBlockIndex {
      */
     protected int searchTreeLevel;
 
-    /**
-     * @return true if the block index is empty.
-     */
+    /** Returns true if the block index is empty. */
     public abstract boolean isEmpty();
 
     /**
@@ -585,9 +583,7 @@ public class HFileBlockIndex {
       return blockDataSizes[i];
     }
 
-    /**
-     * @return the number of root-level blocks in this block index
-     */
+    /** Returns the number of root-level blocks in this block index */
     public int getRootBlockCount() {
       return rootCount;
     }
@@ -1111,16 +1107,12 @@ public class HFileBlockIndex {
       curFirstKey = null;
     }
 
-    /**
-     * @return how many block index entries there are in the root level
-     */
+    /** Returns how many block index entries there are in the root level */
     public final int getNumRootEntries() {
       return rootChunk.getNumEntries();
     }
 
-    /**
-     * @return the number of levels in this block index.
-     */
+    /** Returns the number of levels in this block index. */
     public int getNumLevels() {
       return numLevels;
     }
@@ -1462,9 +1454,7 @@ public class HFileBlockIndex {
       }
     }
 
-    /**
-     * @return the size of this chunk if stored in the non-root index block format
-     */
+    /** Returns the size of this chunk if stored in the non-root index block format */
     int getNonRootSize() {
       return Bytes.SIZEOF_INT // Number of entries
         + Bytes.SIZEOF_INT * (blockKeys.size() + 1) // Secondary index
@@ -1486,16 +1476,12 @@ public class HFileBlockIndex {
       }
     }
 
-    /**
-     * @return the size of this chunk if stored in the root index block format
-     */
+    /** Returns the size of this chunk if stored in the root index block format */
     int getRootSize() {
       return curTotalRootSize;
     }
 
-    /**
-     * @return the number of entries in this block index chunk
-     */
+    /** Returns the number of entries in this block index chunk */
     public int getNumEntries() {
       return blockKeys.size();
     }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileDataBlockEncoder.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileDataBlockEncoder.java
index cf253e20bd8..cb2d5bbcfb6 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileDataBlockEncoder.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileDataBlockEncoder.java
@@ -70,7 +70,7 @@ public interface HFileDataBlockEncoder {
    */
   void saveMetadata(HFile.Writer writer) throws IOException;
 
-  /** @return the data block encoding */
+  /** Returns the data block encoding */
   DataBlockEncoding getDataBlockEncoding();
 
   /**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.java
index c74bbb1bdb7..fc2ffaac876 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.java
@@ -214,13 +214,13 @@ public abstract class HFileReaderImpl implements HFile.Reader, Configurable {
     return getLastKey().map(CellUtil::cloneRow);
   }
 
-  /** @return number of KV entries in this HFile */
+  /** Returns number of KV entries in this HFile */
   @Override
   public long getEntries() {
     return trailer.getEntryCount();
   }
 
-  /** @return comparator */
+  /** Returns comparator */
   @Override
   public CellComparator getComparator() {
     return this.hfileContext.getCellComparator();
@@ -961,16 +961,12 @@ public abstract class HFileReaderImpl implements HFile.Reader, Configurable {
       return blockSeek(key, seekBefore);
     }
 
-    /**
-     * @return True if v &lt;= 0 or v &gt; current block buffer limit.
-     */
+    /** Returns True if v &lt;= 0 or v &gt; current block buffer limit. */
     protected final boolean checkKeyLen(final int v) {
       return v <= 0 || v > this.blockBuffer.limit();
     }
 
-    /**
-     * @return True if v &lt; 0 or v &gt; current block buffer limit.
-     */
+    /** Returns True if v &lt; 0 or v &gt; current block buffer limit. */
     protected final boolean checkLen(final int v) {
       return v < 0 || v > this.blockBuffer.limit();
     }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileScanner.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileScanner.java
index 3e5ada1442f..e77b133523f 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileScanner.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileScanner.java
@@ -101,9 +101,7 @@ public interface HFileScanner extends Shipper, Closeable {
    */
   ByteBuffer getValue();
 
-  /**
-   * @return Instance of {@link org.apache.hadoop.hbase.Cell}.
-   */
+  /** Returns Instance of {@link org.apache.hadoop.hbase.Cell}. */
   Cell getCell();
 
   /**
@@ -124,9 +122,7 @@ public interface HFileScanner extends Shipper, Closeable {
   @Deprecated
   String getValueString();
 
-  /**
-   * @return Reader that underlies this Scanner instance.
-   */
+  /** Returns Reader that underlies this Scanner instance. */
   HFile.Reader getReader();
 
   /**
@@ -135,9 +131,7 @@ public interface HFileScanner extends Shipper, Closeable {
    */
   boolean isSeeked();
 
-  /**
-   * @return the next key in the index (the key to seek to the next block)
-   */
+  /** Returns the next key in the index (the key to seek to the next block) */
   Cell getNextIndexedKey();
 
   /**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterImpl.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterImpl.java
index 9170cbef4e7..a638a443fc4 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterImpl.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterImpl.java
@@ -259,9 +259,7 @@ public class HFileWriterImpl implements HFile.Writer {
     }
   }
 
-  /**
-   * @return Path or null if we were passed a stream rather than a Path.
-   */
+  /** Returns Path or null if we were passed a stream rather than a Path. */
   @Override
   public Path getPath() {
     return path;
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/InlineBlockWriter.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/InlineBlockWriter.java
index 040ca6b9164..6b0c913ca1b 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/InlineBlockWriter.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/InlineBlockWriter.java
@@ -58,8 +58,6 @@ public interface InlineBlockWriter {
    */
   BlockType getInlineBlockType();
 
-  /**
-   * @return true if inline blocks produced by this writer should be cached
-   */
+  /** Returns true if inline blocks produced by this writer should be cached */
   boolean getCacheOnWrite();
 }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/LruCachedBlock.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/LruCachedBlock.java
index f60e4300e4e..ddeee207f1f 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/LruCachedBlock.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/LruCachedBlock.java
@@ -80,9 +80,7 @@ public class LruCachedBlock implements HeapSize, Comparable<LruCachedBlock> {
     }
   }
 
-  /**
-   * @return Time we were cached at in nano seconds.
-   */
+  /** Returns Time we were cached at in nano seconds. */
   public long getCachedTime() {
     return this.cachedTime;
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/LruCachedBlockQueue.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/LruCachedBlockQueue.java
index 8e45cb772d2..272b5b0b0cd 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/LruCachedBlockQueue.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/LruCachedBlockQueue.java
@@ -89,16 +89,12 @@ public class LruCachedBlockQueue implements HeapSize {
     }
   }
 
-  /**
-   * @return The next element in this queue, or {@code null} if the queue is empty.
-   */
+  /** Returns The next element in this queue, or {@code null} if the queue is empty. */
   public LruCachedBlock poll() {
     return queue.poll();
   }
 
-  /**
-   * @return The last element in this queue, or {@code null} if the queue is empty.
-   */
+  /** Returns The last element in this queue, or {@code null} if the queue is empty. */
   public LruCachedBlock pollLast() {
     return queue.pollLast();
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/CachedEntryQueue.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/CachedEntryQueue.java
index daa25cee1de..446cea55f14 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/CachedEntryQueue.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/CachedEntryQueue.java
@@ -92,16 +92,12 @@ public class CachedEntryQueue {
     }
   }
 
-  /**
-   * @return The next element in this queue, or {@code null} if the queue is empty.
-   */
+  /** Returns The next element in this queue, or {@code null} if the queue is empty. */
   public Map.Entry<BlockCacheKey, BucketEntry> poll() {
     return queue.poll();
   }
 
-  /**
-   * @return The last element in this queue, or {@code null} if the queue is empty.
-   */
+  /** Returns The last element in this queue, or {@code null} if the queue is empty. */
   public Map.Entry<BlockCacheKey, BucketEntry> pollLast() {
     return queue.pollLast();
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/IOEngine.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/IOEngine.java
index 62ff19878fc..a7b73a5d886 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/IOEngine.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/IOEngine.java
@@ -28,9 +28,7 @@ import org.apache.yetus.audience.InterfaceAudience;
  */
 @InterfaceAudience.Private
 public interface IOEngine {
-  /**
-   * @return true if persistent storage is supported for the cache when shutdown
-   */
+  /** Returns true if persistent storage is supported for the cache when shutdown */
   boolean isPersistent();
 
   /**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/util/MemorySizeUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/util/MemorySizeUtil.java
index f5f9d760a79..85bcc643558 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/util/MemorySizeUtil.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/util/MemorySizeUtil.java
@@ -153,9 +153,7 @@ public class MemorySizeUtil {
     return DEFAULT_MEMSTORE_SIZE_LOWER_LIMIT;
   }
 
-  /**
-   * @return Pair of global memstore size and memory type(ie. on heap or off heap).
-   */
+  /** Returns Pair of global memstore size and memory type(ie. on heap or off heap). */
   public static Pair<Long, MemoryType> getGlobalMemStoreSize(Configuration conf) {
     long offheapMSGlobal = conf.getLong(OFFHEAP_MEMSTORE_SIZE_KEY, 0);// Size in MBs
     if (offheapMSGlobal > 0) {
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/FastPathBalancedQueueRpcExecutor.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/FastPathBalancedQueueRpcExecutor.java
index 7dcf5c1361a..440f4d3a819 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/FastPathBalancedQueueRpcExecutor.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/FastPathBalancedQueueRpcExecutor.java
@@ -73,9 +73,7 @@ public class FastPathBalancedQueueRpcExecutor extends BalancedQueueRpcExecutor {
     return handler != null ? handler.loadCallRunner(callTask) : super.dispatch(callTask);
   }
 
-  /**
-   * @return Pop a Handler instance if one available ready-to-go or else return null.
-   */
+  /** Returns Pop a Handler instance if one available ready-to-go or else return null. */
   private FastPathRpcHandler popReadyHandler() {
     return this.fastPathHandlerStack.poll();
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/QueueBalancer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/QueueBalancer.java
index a13f5d85823..2111d5b78ca 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/QueueBalancer.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/QueueBalancer.java
@@ -27,8 +27,6 @@ import org.apache.yetus.audience.InterfaceStability;
 @InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CONFIG)
 @InterfaceStability.Stable
 public interface QueueBalancer {
-  /**
-   * @return the index of the next queue to which a request should be inserted
-   */
+  /** Returns the index of the next queue to which a request should be inserted */
   int getNextQueue(CallRunner callRunner);
 }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcCall.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcCall.java
index e12bcf6964f..59008cb08a4 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcCall.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcCall.java
@@ -36,34 +36,22 @@ import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader
 @InterfaceStability.Evolving
 public interface RpcCall extends RpcCallContext {
 
-  /**
-   * @return The service of this call.
-   */
+  /** Returns The service of this call. */
   BlockingService getService();
 
-  /**
-   * @return The service method.
-   */
+  /** Returns The service method. */
   MethodDescriptor getMethod();
 
-  /**
-   * @return The call parameter message.
-   */
+  /** Returns The call parameter message. */
   Message getParam();
 
-  /**
-   * @return The CellScanner that can carry input and result payload.
-   */
+  /** Returns The CellScanner that can carry input and result payload. */
   CellScanner getCellScanner();
 
-  /**
-   * @return The timestamp when the call is constructed.
-   */
+  /** Returns The timestamp when the call is constructed. */
   long getReceiveTime();
 
-  /**
-   * @return The time when the call starts to be executed.
-   */
+  /** Returns The time when the call starts to be executed. */
   long getStartTime();
 
   /**
@@ -71,14 +59,10 @@ public interface RpcCall extends RpcCallContext {
    */
   void setStartTime(long startTime);
 
-  /**
-   * @return The timeout of this call.
-   */
+  /** Returns The timeout of this call. */
   int getTimeout();
 
-  /**
-   * @return The Priority of this call.
-   */
+  /** Returns The Priority of this call. */
   int getPriority();
 
   /**
@@ -95,14 +79,10 @@ public interface RpcCall extends RpcCallContext {
    */
   long getSize();
 
-  /**
-   * @return The request header of this call.
-   */
+  /** Returns The request header of this call. */
   RequestHeader getHeader();
 
-  /**
-   * @return Port of remote address in this call
-   */
+  /** Returns Port of remote address in this call */
   int getRemotePort();
 
   /**
@@ -125,8 +105,6 @@ public interface RpcCall extends RpcCallContext {
    */
   void cleanup();
 
-  /**
-   * @return A short string format of this call without possibly lengthy params
-   */
+  /** Returns A short string format of this call without possibly lengthy params */
   String toShortString();
 }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcCallContext.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcCallContext.java
index 6f045731ecb..95843652abe 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcCallContext.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcCallContext.java
@@ -55,21 +55,15 @@ public interface RpcCallContext {
    */
   Optional<User> getRequestUser();
 
-  /**
-   * @return Current request's user name or not present if none ongoing.
-   */
+  /** Returns Current request's user name or not present if none ongoing. */
   default Optional<String> getRequestUserName() {
     return getRequestUser().map(User::getShortName);
   }
 
-  /**
-   * @return Address of remote client in this call
-   */
+  /** Returns Address of remote client in this call */
   InetAddress getRemoteAddress();
 
-  /**
-   * @return the client version info, or null if the information is not present
-   */
+  /** Returns the client version info, or null if the information is not present */
   VersionInfo getClientVersionInfo();
 
   /**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java
index fa3bc2cc63a..2ca86f57f74 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java
@@ -708,9 +708,7 @@ public abstract class RpcServer implements RpcServerInterface, ConfigurationObse
     return getRequestUser().map(User::getShortName);
   }
 
-  /**
-   * @return Address of remote client if a request is ongoing, else null
-   */
+  /** Returns Address of remote client if a request is ongoing, else null */
   public static Optional<InetAddress> getRemoteAddress() {
     return getCurrentCall().map(RpcCall::getRemoteAddress);
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ActiveMasterManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ActiveMasterManager.java
index 50628f8717b..3f8d6d2a702 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ActiveMasterManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ActiveMasterManager.java
@@ -314,9 +314,7 @@ public class ActiveMasterManager extends ZKListener {
     return false;
   }
 
-  /**
-   * @return True if cluster has an active master.
-   */
+  /** Returns True if cluster has an active master. */
   boolean hasActiveMaster() {
     try {
       if (ZKUtil.checkExists(watcher, watcher.getZNodePaths().masterAddressZNode) >= 0) {
@@ -353,9 +351,7 @@ public class ActiveMasterManager extends ZKListener {
     }
   }
 
-  /**
-   * @return list of registered backup masters.
-   */
+  /** Returns list of registered backup masters. */
   public List<ServerName> getBackupMasters() {
     return backupMasters;
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
index 969fb4f5583..b29a8a76f5c 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
@@ -1811,9 +1811,7 @@ public class HMaster extends HRegionServer implements MasterServices {
     }
   }
 
-  /**
-   * @return Get remote side's InetAddress
-   */
+  /** Returns Get remote side's InetAddress */
   InetAddress getRemoteInetAddress(final int port, final long serverStartCode)
     throws UnknownHostException {
     // Do it out here in its own little method so can fake an address when
@@ -1831,9 +1829,7 @@ public class HMaster extends HRegionServer implements MasterServices {
     return ia;
   }
 
-  /**
-   * @return Maximum time we should run balancer for
-   */
+  /** Returns Maximum time we should run balancer for */
   private int getMaxBalancingTime() {
     // if max balancing time isn't set, defaulting it to period time
     int maxBalancingTime =
@@ -1842,9 +1838,7 @@ public class HMaster extends HRegionServer implements MasterServices {
     return maxBalancingTime;
   }
 
-  /**
-   * @return Maximum number of regions in transition
-   */
+  /** Returns Maximum number of regions in transition */
   private int getMaxRegionsInTransition() {
     int numRegions = this.assignmentManager.getRegionStates().getRegionAssignments().size();
     return Math.max((int) Math.floor(numRegions * this.maxRitPercent), 1);
@@ -2122,9 +2116,7 @@ public class HMaster extends HRegionServer implements MasterServices {
     return regionNormalizerManager.normalizeRegions(targetTables, isHighPriority);
   }
 
-  /**
-   * @return Client info for use as prefix on an audit log string; who did an action
-   */
+  /** Returns Client info for use as prefix on an audit log string; who did an action */
   @Override
   public String getClientIdAuditPrefix() {
     return "Client=" + RpcServer.getRequestUserName().orElse(null) + "/"
@@ -2925,9 +2917,7 @@ public class HMaster extends HRegionServer implements MasterServices {
     return null;
   }
 
-  /**
-   * @return cluster status
-   */
+  /** Returns cluster status */
   public ClusterMetrics getClusterMetrics() throws IOException {
     return getClusterMetrics(EnumSet.allOf(Option.class));
   }
@@ -2953,9 +2943,7 @@ public class HMaster extends HRegionServer implements MasterServices {
     return activeMasterManager.getBackupMasters();
   }
 
-  /**
-   * @return info port of active master or 0 if any exception occurs.
-   */
+  /** Returns info port of active master or 0 if any exception occurs. */
   public int getActiveMasterInfoPort() {
     return activeMasterManager.getActiveMasterInfoPort();
   }
@@ -2982,23 +2970,17 @@ public class HMaster extends HRegionServer implements MasterServices {
     return CoprocessorHost.getLoadedCoprocessors().toString();
   }
 
-  /**
-   * @return timestamp in millis when HMaster was started.
-   */
+  /** Returns timestamp in millis when HMaster was started. */
   public long getMasterStartTime() {
     return startcode;
   }
 
-  /**
-   * @return timestamp in millis when HMaster became the active master.
-   */
+  /** Returns timestamp in millis when HMaster became the active master. */
   public long getMasterActiveTime() {
     return masterActiveTime;
   }
 
-  /**
-   * @return timestamp in millis when HMaster finished becoming the active master
-   */
+  /** Returns timestamp in millis when HMaster finished becoming the active master */
   public long getMasterFinishedInitializationTime() {
     return masterFinishedInitializationTime;
   }
@@ -3030,9 +3012,7 @@ public class HMaster extends HRegionServer implements MasterServices {
     assignmentManager.checkIfShouldMoveSystemRegionAsync();
   }
 
-  /**
-   * @return array of coprocessor SimpleNames.
-   */
+  /** Returns array of coprocessor SimpleNames. */
   public String[] getMasterCoprocessors() {
     Set<String> masterCoprocessors = getMasterCoprocessorHost().getCoprocessors();
     return masterCoprocessors.toArray(new String[masterCoprocessors.size()]);
@@ -3305,17 +3285,13 @@ public class HMaster extends HRegionServer implements MasterServices {
     return this.logCleaner;
   }
 
-  /**
-   * @return the underlying snapshot manager
-   */
+  /** Returns the underlying snapshot manager */
   @Override
   public SnapshotManager getSnapshotManager() {
     return this.snapshotManager;
   }
 
-  /**
-   * @return the underlying MasterProcedureManagerHost
-   */
+  /** Returns the underlying MasterProcedureManagerHost */
   @Override
   public MasterProcedureManagerHost getMasterProcedureManagerHost() {
     return mpmHost;
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java
index 5b210df4b26..7803fbd9f11 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java
@@ -185,37 +185,27 @@ public class MasterFileSystem {
     return this.conf;
   }
 
-  /**
-   * @return HBase root dir.
-   */
+  /** Returns HBase root dir. */
   public Path getRootDir() {
     return this.rootdir;
   }
 
-  /**
-   * @return HBase root log dir.
-   */
+  /** Returns HBase root log dir. */
   public Path getWALRootDir() {
     return this.walRootDir;
   }
 
-  /**
-   * @return the directory for a give {@code region}.
-   */
+  /** Returns the directory for a give {@code region}. */
   public Path getRegionDir(RegionInfo region) {
     return FSUtils.getRegionDirFromRootDir(getRootDir(), region);
   }
 
-  /**
-   * @return HBase temp dir.
-   */
+  /** Returns HBase temp dir. */
   public Path getTempDir() {
     return this.tempdir;
   }
 
-  /**
-   * @return The unique identifier generated for this cluster
-   */
+  /** Returns The unique identifier generated for this cluster */
   public ClusterId getClusterId() {
     return clusterId;
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java
index d48671b79f5..bc3b4af2bc3 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java
@@ -502,9 +502,7 @@ public class MasterRpcServices extends RSRpcServices
     return switchBalancer(b, BalanceSwitchMode.SYNC);
   }
 
-  /**
-   * @return list of blocking services and their security info classes that this server supports
-   */
+  /** Returns list of blocking services and their security info classes that this server supports */
   @Override
   protected List<BlockingServiceAndInterface> getServices() {
     List<BlockingServiceAndInterface> bssi = new ArrayList<>(5);
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java
index cfaeb139331..6251bdd8eaf 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java
@@ -61,84 +61,52 @@ import org.apache.yetus.audience.InterfaceAudience;
  */
 @InterfaceAudience.Private
 public interface MasterServices extends Server {
-  /**
-   * @return the underlying snapshot manager
-   */
+  /** Returns the underlying snapshot manager */
   SnapshotManager getSnapshotManager();
 
-  /**
-   * @return the underlying MasterProcedureManagerHost
-   */
+  /** Returns the underlying MasterProcedureManagerHost */
   MasterProcedureManagerHost getMasterProcedureManagerHost();
 
-  /**
-   * @return Master's instance of {@link ClusterSchema}
-   */
+  /** Returns Master's instance of {@link ClusterSchema} */
   ClusterSchema getClusterSchema();
 
-  /**
-   * @return Master's instance of the {@link AssignmentManager}
-   */
+  /** Returns Master's instance of the {@link AssignmentManager} */
   AssignmentManager getAssignmentManager();
 
-  /**
-   * @return Master's filesystem {@link MasterFileSystem} utility class.
-   */
+  /** Returns Master's filesystem {@link MasterFileSystem} utility class. */
   MasterFileSystem getMasterFileSystem();
 
-  /**
-   * @return Master's WALs {@link MasterWalManager} utility class.
-   */
+  /** Returns Master's WALs {@link MasterWalManager} utility class. */
   MasterWalManager getMasterWalManager();
 
-  /**
-   * @return Master's {@link ServerManager} instance.
-   */
+  /** Returns Master's {@link ServerManager} instance. */
   ServerManager getServerManager();
 
-  /**
-   * @return Master's instance of {@link ExecutorService}
-   */
+  /** Returns Master's instance of {@link ExecutorService} */
   ExecutorService getExecutorService();
 
-  /**
-   * @return Master's instance of {@link TableStateManager}
-   */
+  /** Returns Master's instance of {@link TableStateManager} */
   TableStateManager getTableStateManager();
 
-  /**
-   * @return Master's instance of {@link MasterCoprocessorHost}
-   */
+  /** Returns Master's instance of {@link MasterCoprocessorHost} */
   MasterCoprocessorHost getMasterCoprocessorHost();
 
-  /**
-   * @return Master's instance of {@link MasterQuotaManager}
-   */
+  /** Returns Master's instance of {@link MasterQuotaManager} */
   MasterQuotaManager getMasterQuotaManager();
 
-  /**
-   * @return Master's instance of {@link RegionNormalizerManager}
-   */
+  /** Returns Master's instance of {@link RegionNormalizerManager} */
   RegionNormalizerManager getRegionNormalizerManager();
 
-  /**
-   * @return Master's instance of {@link CatalogJanitor}
-   */
+  /** Returns Master's instance of {@link CatalogJanitor} */
   CatalogJanitor getCatalogJanitor();
 
-  /**
-   * @return Master's instance of {@link ProcedureExecutor}
-   */
+  /** Returns Master's instance of {@link ProcedureExecutor} */
   ProcedureExecutor<MasterProcedureEnv> getMasterProcedureExecutor();
 
-  /**
-   * @return Tripped when Master has finished initialization.
-   */
+  /** Returns Tripped when Master has finished initialization. */
   public ProcedureEvent<?> getInitializedEvent();
 
-  /**
-   * @return Master's instance of {@link MetricsMaster}
-   */
+  /** Returns Master's instance of {@link MetricsMaster} */
   MetricsMaster getMasterMetrics();
 
   /**
@@ -259,9 +227,7 @@ public interface MasterServices extends Server {
   long splitRegion(final RegionInfo regionInfo, final byte[] splitRow, final long nonceGroup,
     final long nonce) throws IOException;
 
-  /**
-   * @return Return table descriptors implementation.
-   */
+  /** Returns Return table descriptors implementation. */
   TableDescriptors getTableDescriptors();
 
   /**
@@ -277,14 +243,10 @@ public interface MasterServices extends Server {
    */
   boolean registerService(Service instance);
 
-  /**
-   * @return true if master is the active one
-   */
+  /** Returns true if master is the active one */
   boolean isActiveMaster();
 
-  /**
-   * @return true if master is initialized
-   */
+  /** Returns true if master is initialized */
   boolean isInitialized();
 
   /**
@@ -347,16 +309,12 @@ public interface MasterServices extends Server {
    */
   public long getLastMajorCompactionTimestampForRegion(byte[] regionName) throws IOException;
 
-  /**
-   * @return load balancer
-   */
+  /** Returns load balancer */
   public LoadBalancer getLoadBalancer();
 
   boolean isSplitOrMergeEnabled(MasterSwitchType switchType);
 
-  /**
-   * @return Favored Nodes Manager
-   */
+  /** Returns Favored Nodes Manager */
   public FavoredNodesManager getFavoredNodesManager();
 
   /**
@@ -415,9 +373,7 @@ public interface MasterServices extends Server {
   List<ReplicationPeerDescription> listReplicationPeers(String regex)
     throws ReplicationException, IOException;
 
-  /**
-   * @return {@link LockManager} to lock namespaces/tables/regions.
-   */
+  /** Returns {@link LockManager} to lock namespaces/tables/regions. */
   LockManager getLockManager();
 
   public String getRegionServerVersion(final ServerName sn);
@@ -430,26 +386,18 @@ public interface MasterServices extends Server {
 
   String getClientIdAuditPrefix();
 
-  /**
-   * @return True if cluster is up; false if cluster is not up (we are shutting down).
-   */
+  /** Returns True if cluster is up; false if cluster is not up (we are shutting down). */
   boolean isClusterUp();
 
-  /**
-   * @return return null if current is zk-based WAL splitting
-   */
+  /** Returns return null if current is zk-based WAL splitting */
   default SplitWALManager getSplitWALManager() {
     return null;
   }
 
-  /**
-   * @return the {@link AccessChecker}
-   */
+  /** Returns the {@link AccessChecker} */
   AccessChecker getAccessChecker();
 
-  /**
-   * @return the {@link ZKPermissionWatcher}
-   */
+  /** Returns the {@link ZKPermissionWatcher} */
   ZKPermissionWatcher getZKPermissionWatcher();
 
   /**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterWalManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterWalManager.java
index e6ab12c0250..125657db187 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterWalManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterWalManager.java
@@ -171,9 +171,7 @@ public class MasterWalManager {
       p -> !p.getName().endsWith(AbstractFSWALProvider.SPLITTING_EXT));
   }
 
-  /**
-   * @return listing of ServerNames found by parsing WAL directory paths in FS.
-   */
+  /** Returns listing of ServerNames found by parsing WAL directory paths in FS. */
   public Set<ServerName> getServerNamesFromWALDirPath(final PathFilter filter) throws IOException {
     FileStatus[] walDirForServerNames = getWALDirPaths(filter);
     return Stream.of(walDirForServerNames).map(s -> {
@@ -189,7 +187,7 @@ public class MasterWalManager {
   }
 
   /**
-   * @return List of all RegionServer WAL dirs; i.e. this.rootDir/HConstants.HREGION_LOGDIR_NAME.
+   * Returns List of all RegionServer WAL dirs; i.e. this.rootDir/HConstants.HREGION_LOGDIR_NAME.
    */
   public FileStatus[] getWALDirPaths(final PathFilter filter) throws IOException {
     Path walDirPath = new Path(CommonFSUtils.getWALRootDir(conf), HConstants.HREGION_LOGDIR_NAME);
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MetricsAssignmentManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MetricsAssignmentManager.java
index 41575c5a7bd..eaf335e52ef 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MetricsAssignmentManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MetricsAssignmentManager.java
@@ -127,58 +127,42 @@ public class MetricsAssignmentManager {
     assignmentManagerSource.setEmptyRegionInfoRegions(emptyRegionInfoRegions);
   }
 
-  /**
-   * @return Set of common metrics for assign procedure
-   */
+  /** Returns Set of common metrics for assign procedure */
   public ProcedureMetrics getAssignProcMetrics() {
     return assignProcMetrics;
   }
 
-  /**
-   * @return Set of common metrics for unassign procedure
-   */
+  /** Returns Set of common metrics for unassign procedure */
   public ProcedureMetrics getUnassignProcMetrics() {
     return unassignProcMetrics;
   }
 
-  /**
-   * @return Set of common metrics for move procedure
-   */
+  /** Returns Set of common metrics for move procedure */
   public ProcedureMetrics getMoveProcMetrics() {
     return moveProcMetrics;
   }
 
-  /**
-   * @return Set of common metrics for reopen procedure
-   */
+  /** Returns Set of common metrics for reopen procedure */
   public ProcedureMetrics getReopenProcMetrics() {
     return reopenProcMetrics;
   }
 
-  /**
-   * @return Set of common metrics for OpenRegionProcedure
-   */
+  /** Returns Set of common metrics for OpenRegionProcedure */
   public ProcedureMetrics getOpenProcMetrics() {
     return openProcMetrics;
   }
 
-  /**
-   * @return Set of common metrics for CloseRegionProcedure
-   */
+  /** Returns Set of common metrics for CloseRegionProcedure */
   public ProcedureMetrics getCloseProcMetrics() {
     return closeProcMetrics;
   }
 
-  /**
-   * @return Set of common metrics for split procedure
-   */
+  /** Returns Set of common metrics for split procedure */
   public ProcedureMetrics getSplitProcMetrics() {
     return splitProcMetrics;
   }
 
-  /**
-   * @return Set of common metrics for merge procedure
-   */
+  /** Returns Set of common metrics for merge procedure */
   public ProcedureMetrics getMergeProcMetrics() {
     return mergeProcMetrics;
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MetricsMaster.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MetricsMaster.java
index bf6589bf4e8..671291c62e2 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MetricsMaster.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MetricsMaster.java
@@ -116,9 +116,7 @@ public class MetricsMaster {
     masterQuotaSource.incrementSpaceQuotaObserverChoreTime(executionTime);
   }
 
-  /**
-   * @return Set of metrics for assign procedure
-   */
+  /** Returns Set of metrics for assign procedure */
   public ProcedureMetrics getServerCrashProcMetrics() {
     return serverCrashProcMetrics;
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionPlacementMaintainer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionPlacementMaintainer.java
index 17b87848733..c9876aeb83c 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionPlacementMaintainer.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionPlacementMaintainer.java
@@ -128,9 +128,7 @@ public class RegionPlacementMaintainer {
     }
   }
 
-  /**
-   * @return the new RegionAssignmentSnapshot n
-   */
+  /** Returns the new RegionAssignmentSnapshot n */
   public SnapshotOfRegionAssignmentFromMeta getRegionAssignmentSnapshot() throws IOException {
     SnapshotOfRegionAssignmentFromMeta currentAssignmentShapshot =
       new SnapshotOfRegionAssignmentFromMeta(ConnectionFactory.createConnection(conf));
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
index faae49ff7e8..9471a9ff317 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
@@ -439,15 +439,13 @@ public class ServerManager {
     return numServers == 0 ? 0 : (double) totalLoad / (double) numServers;
   }
 
-  /** @return the count of active regionservers */
+  /** Returns the count of active regionservers */
   public int countOfRegionServers() {
     // Presumes onlineServers is a concurrent map
     return this.onlineServers.size();
   }
 
-  /**
-   * @return Read-only map of servers to serverinfo
-   */
+  /** Returns Read-only map of servers to serverinfo */
   public Map<ServerName, ServerMetrics> getOnlineServers() {
     // Presumption is that iterating the returned Map is OK.
     synchronized (this.onlineServers) {
@@ -835,9 +833,7 @@ public class ServerManager {
     return max == Integer.MAX_VALUE ? "NO_LIMIT" : Integer.toString(max);
   }
 
-  /**
-   * @return A copy of the internal list of online servers.
-   */
+  /** Returns A copy of the internal list of online servers. */
   public List<ServerName> getOnlineServersList() {
     // TODO: optimize the load balancer call so we don't need to make a new list
     // TODO: FIX. THIS IS POPULAR CALL.
@@ -865,9 +861,7 @@ public class ServerManager {
     return names;
   }
 
-  /**
-   * @return A copy of the internal list of draining servers.
-   */
+  /** Returns A copy of the internal list of draining servers. */
   public List<ServerName> getDrainingServersList() {
     return new ArrayList<>(this.drainingServers);
   }
@@ -882,9 +876,7 @@ public class ServerManager {
     UNKNOWN
   }
 
-  /**
-   * @return whether the server is online, dead, or unknown.
-   */
+  /** Returns whether the server is online, dead, or unknown. */
   public synchronized ServerLiveState isServerKnownAndOnline(ServerName serverName) {
     return onlineServers.containsKey(serverName)
       ? ServerLiveState.LIVE
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SnapshotSentinel.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SnapshotSentinel.java
index 0a564b9b9c6..e14027bf173 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SnapshotSentinel.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SnapshotSentinel.java
@@ -37,9 +37,7 @@ public interface SnapshotSentinel {
    */
   boolean isFinished();
 
-  /**
-   * @return -1 if the snapshot is in progress, otherwise the completion timestamp.
-   */
+  /** Returns -1 if the snapshot is in progress, otherwise the completion timestamp. */
   long getCompletionTimestamp();
 
   /**
@@ -48,9 +46,7 @@ public interface SnapshotSentinel {
    */
   void cancel(String why);
 
-  /**
-   * @return the description of the snapshot being run
-   */
+  /** Returns the description of the snapshot being run */
   SnapshotDescription getSnapshot();
 
   /**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignmentManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignmentManager.java
index 71dad35c0f3..6257edf04f7 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignmentManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignmentManager.java
@@ -2386,9 +2386,7 @@ public class AssignmentManager {
     return master;
   }
 
-  /**
-   * @return a snapshot of rsReports
-   */
+  /** Returns a snapshot of rsReports */
   public Map<ServerName, Set<byte[]>> getRSReports() {
     Map<ServerName, Set<byte[]>> rsReportsSnapshot = new HashMap<>();
     synchronized (rsReports) {
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.java
index d00394a6fe3..882716a17ac 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.java
@@ -745,9 +745,7 @@ public class MergeTableRegionsProcedure
     }
   }
 
-  /**
-   * @return The merged region. Maybe be null if called to early or we failed.
-   */
+  /** Returns The merged region. Maybe be null if called to early or we failed. */
   RegionInfo getMergedRegion() {
     return this.mergedRegion;
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/RegionStates.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/RegionStates.java
index 855731982ba..3053aece59d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/RegionStates.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/RegionStates.java
@@ -177,12 +177,12 @@ public class RegionStates {
     return regions;
   }
 
-  /** @return A view of region state nodes for all the regions. */
+  /** Returns A view of region state nodes for all the regions. */
   public Collection<RegionStateNode> getRegionStateNodes() {
     return Collections.unmodifiableCollection(regionsMap.values());
   }
 
-  /** @return A snapshot of region state nodes for all the regions. */
+  /** Returns A snapshot of region state nodes for all the regions. */
   public ArrayList<RegionState> getRegionStates() {
     final ArrayList<RegionState> regions = new ArrayList<>(regionsMap.size());
     for (RegionStateNode node : regionsMap.values()) {
@@ -217,9 +217,7 @@ public class RegionStates {
     return !getTableRegionStates(tableName).isEmpty();
   }
 
-  /**
-   * @return Return online regions of table; does not include OFFLINE or SPLITTING regions.
-   */
+  /** Returns Return online regions of table; does not include OFFLINE or SPLITTING regions. */
   public List<RegionInfo> getRegionsOfTable(TableName table) {
     return getRegionsOfTable(table, regionNode -> !regionNode.isInState(State.OFFLINE, State.SPLIT)
       && !regionNode.getRegionInfo().isSplitParent());
@@ -352,9 +350,7 @@ public class RegionStates {
       .collect(Collectors.toList());
   }
 
-  /**
-   * @return Return the regions of the table and filter them.
-   */
+  /** Returns Return the regions of the table and filter them. */
   private List<RegionInfo> getRegionsOfTable(TableName table, Predicate<RegionStateNode> filter) {
     return getTableRegionStateNodes(table).stream().filter(filter).map(n -> n.getRegionInfo())
       .collect(Collectors.toList());
@@ -465,9 +461,7 @@ public class RegionStates {
     return isRegionInState(regionInfo, State.OPEN);
   }
 
-  /**
-   * @return True if region is offline (In OFFLINE or CLOSED state).
-   */
+  /** Returns True if region is offline (In OFFLINE or CLOSED state). */
   public boolean isRegionOffline(final RegionInfo regionInfo) {
     return isRegionInState(regionInfo, State.OFFLINE, State.CLOSED);
   }
@@ -762,9 +756,7 @@ public class RegionStates {
     serverMap.remove(serverName);
   }
 
-  /**
-   * @return Pertinent ServerStateNode or NULL if none found (Do not make modifications).
-   */
+  /** Returns Pertinent ServerStateNode or NULL if none found (Do not make modifications). */
   public ServerStateNode getServerNode(final ServerName serverName) {
     return serverMap.get(serverName);
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.java
index 60c9609d3a2..59003412a8d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.java
@@ -130,7 +130,7 @@ public class UnassignProcedure extends RegionTransitionProcedure {
   }
 
   /**
-   * @return If true, we will re-wake up this procedure; if false, the procedure stays suspended.
+   * Returns If true, we will re-wake up this procedure; if false, the procedure stays suspended.
    */
   @Override
   protected boolean remoteCallFailed(final MasterProcedureEnv env, final RegionStateNode regionNode,
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/FavoredStochasticBalancer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/FavoredStochasticBalancer.java
index d2c93b8ead1..7fc867bf1b9 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/FavoredStochasticBalancer.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/FavoredStochasticBalancer.java
@@ -83,9 +83,7 @@ public class FavoredStochasticBalancer extends StochasticLoadBalancer
     return fnPickers;
   }
 
-  /**
-   * @return any candidate generator in random
-   */
+  /** Returns any candidate generator in random */
   @Override
   protected CandidateGenerator getRandomGenerator() {
     return candidateGenerators.get(ThreadLocalRandom.current().nextInt(candidateGenerators.size()));
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/hbck/HbckChore.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/hbck/HbckChore.java
index 0a3d310c98c..85e978014e5 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/hbck/HbckChore.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/hbck/HbckChore.java
@@ -81,7 +81,7 @@ public class HbckChore extends ScheduledChore {
   }
 
   /**
-   * @return Returns last published Report that comes of last successful execution of this chore.
+   * Returns Returns last published Report that comes of last successful execution of this chore.
    */
   public HbckReport getLastReport() {
     return lastReport;
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/janitor/CatalogJanitor.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/janitor/CatalogJanitor.java
index 4bc54aae969..0d96d584d7d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/janitor/CatalogJanitor.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/janitor/CatalogJanitor.java
@@ -237,9 +237,7 @@ public class CatalogJanitor extends ScheduledChore {
     return visitor.getReport();
   }
 
-  /**
-   * @return Returns last published Report that comes of last successful scan of hbase:meta.
-   */
+  /** Returns Returns last published Report that comes of last successful scan of hbase:meta. */
   public CatalogJanitorReport getLastReport() {
     return this.lastReport;
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/janitor/CatalogJanitorReport.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/janitor/CatalogJanitorReport.java
index 6a9eb7e1b4e..57ecb626d1e 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/janitor/CatalogJanitorReport.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/janitor/CatalogJanitorReport.java
@@ -85,9 +85,7 @@ public class CatalogJanitorReport {
     return emptyRegionInfo;
   }
 
-  /**
-   * @return True if an 'empty' lastReport -- no problems found.
-   */
+  /** Returns True if an 'empty' lastReport -- no problems found. */
   public boolean isEmpty() {
     return this.holes.isEmpty() && this.overlaps.isEmpty() && this.unknownServers.isEmpty()
       && this.emptyRegionInfo.isEmpty();
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/janitor/ReportMakingVisitor.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/janitor/ReportMakingVisitor.java
index a61c9010685..20b24ada2a5 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/janitor/ReportMakingVisitor.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/janitor/ReportMakingVisitor.java
@@ -195,9 +195,7 @@ class ReportMakingVisitor implements MetaTableAccessor.CloseableVisitor {
     this.report.holes.add(new Pair<>(a, b));
   }
 
-  /**
-   * @return True if table is disabled or disabling; defaults false!
-   */
+  /** Returns True if table is disabled or disabling; defaults false! */
   boolean isTableDisabled(RegionInfo ri) {
     if (ri == null) {
       return false;
@@ -274,9 +272,7 @@ class ReportMakingVisitor implements MetaTableAccessor.CloseableVisitor {
     }
   }
 
-  /**
-   * @return True iff first row in hbase:meta or if we've broached a new table in hbase:meta
-   */
+  /** Returns True iff first row in hbase:meta or if we've broached a new table in hbase:meta */
   private boolean isTableTransition(RegionInfo ri) {
     return this.previous == null || !this.previous.getTable().equals(ri.getTable());
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/NormalizationPlan.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/NormalizationPlan.java
index 435553eb8cf..0e0d39d10b1 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/NormalizationPlan.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/NormalizationPlan.java
@@ -32,8 +32,6 @@ public interface NormalizationPlan {
     NONE
   }
 
-  /**
-   * @return the type of this plan
-   */
+  /** Returns the type of this plan */
   PlanType getType();
 }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java
index c4da9c049ab..ceb8851645d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java
@@ -234,9 +234,7 @@ class SimpleRegionNormalizer implements RegionNormalizer, ConfigurationObserver
     return plans;
   }
 
-  /**
-   * @return size of region in MB and if region is not found than -1
-   */
+  /** Returns size of region in MB and if region is not found than -1 */
   private long getRegionSizeMB(RegionInfo hri) {
     ServerName sn =
       masterServices.getAssignmentManager().getRegionStates().getRegionServerOfRegion(hri);
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/AbstractStateMachineRegionProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/AbstractStateMachineRegionProcedure.java
index a6c2e52aef6..fe98a78b4d7 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/AbstractStateMachineRegionProcedure.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/AbstractStateMachineRegionProcedure.java
@@ -46,9 +46,7 @@ public abstract class AbstractStateMachineRegionProcedure<TState>
     super();
   }
 
-  /**
-   * @return The RegionInfo of the region we are operating on.
-   */
+  /** Returns The RegionInfo of the region we are operating on. */
   public RegionInfo getRegion() {
     return this.hri;
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/EnableTableProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/EnableTableProcedure.java
index b42cc1a2f6f..3466bf68843 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/EnableTableProcedure.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/EnableTableProcedure.java
@@ -342,9 +342,7 @@ public class EnableTableProcedure extends AbstractStateMachineTableProcedure<Ena
     }
   }
 
-  /**
-   * @return Maximum region replica id found in passed list of regions.
-   */
+  /** Returns Maximum region replica id found in passed list of regions. */
   private static int getMaxReplicaId(List<RegionInfo> regions) {
     int max = 0;
     for (RegionInfo regionInfo : regions) {
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.java
index 8691675972e..dff3de7b5be 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.java
@@ -267,9 +267,7 @@ public class ServerCrashProcedure extends
     return Flow.HAS_MORE_STATE;
   }
 
-  /**
-   * @return List of Regions on crashed server.
-   */
+  /** Returns List of Regions on crashed server. */
   List<RegionInfo> getRegionsOnCrashedServer(MasterProcedureEnv env) {
     return env.getMasterServices().getAssignmentManager().getRegionsOnServer(serverName);
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ServerProcedureInterface.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ServerProcedureInterface.java
index 6578754a0db..1d7dcb77e09 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ServerProcedureInterface.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ServerProcedureInterface.java
@@ -58,14 +58,10 @@ public interface ServerProcedureInterface {
     VERIFY_SNAPSHOT
   }
 
-  /**
-   * @return Name of this server instance.
-   */
+  /** Returns Name of this server instance. */
   ServerName getServerName();
 
-  /**
-   * @return True if this server has an hbase:meta table region.
-   */
+  /** Returns True if this server has an hbase:meta table region. */
   boolean hasMetaTableRegion();
 
   /**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/TableProcedureInterface.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/TableProcedureInterface.java
index 1a362e108f8..4b3ae5ae138 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/TableProcedureInterface.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/TableProcedureInterface.java
@@ -44,9 +44,7 @@ public interface TableProcedureInterface {
     MERGED_REGIONS_GC/* region operations */
   };
 
-  /**
-   * @return the name of the table the procedure is operating on
-   */
+  /** Returns the name of the table the procedure is operating on */
   TableName getTableName();
 
   /**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.java
index 5d9701d9bcc..e9a33f6cb28 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.java
@@ -349,9 +349,7 @@ public class SnapshotFileCache implements Stoppable {
       this.files = files;
     }
 
-    /**
-     * @return the hfiles in the snapshot when <tt>this</tt> was made.
-     */
+    /** Returns the hfiles in the snapshot when <tt>this</tt> was made. */
     public Collection<String> getFiles() {
       return this.files;
     }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotManager.java
index 259987ca66d..56c4af7a85e 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotManager.java
@@ -824,9 +824,7 @@ public class SnapshotManager extends MasterProcedureManager implements Stoppable
     }
   }
 
-  /**
-   * @return distributed commit coordinator for all running snapshots
-   */
+  /** Returns distributed commit coordinator for all running snapshots */
   ProcedureCoordinator getCoordinator() {
     return coordinator;
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/zksyncer/ClientZKSyncer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/zksyncer/ClientZKSyncer.java
index a9881ef9b64..ed5b78e211d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/zksyncer/ClientZKSyncer.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/zksyncer/ClientZKSyncer.java
@@ -271,9 +271,7 @@ public abstract class ClientZKSyncer extends ZKListener {
    */
   protected abstract boolean validate(String path);
 
-  /**
-   * @return the zk path(s) to watch
-   */
+  /** Returns the zk path(s) to watch */
   protected abstract Set<String> getPathsToWatch();
 
   protected final void refreshWatchingList() {
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/Procedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/Procedure.java
index cd8ae58e53a..8eb477fb846 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/Procedure.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/Procedure.java
@@ -161,7 +161,7 @@ public class Procedure implements Callable<Void>, ForeignExceptionListener {
   }
 
   /**
-   * @return String of the procedure members both trying to enter the barrier and already in barrier
+   * Returns String of the procedure members both trying to enter the barrier and already in barrier
    */
   public String getStatus() {
     String waiting, done;
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinator.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinator.java
index 2e7d7e207ea..908419a42b4 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinator.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinator.java
@@ -289,9 +289,7 @@ public class ProcedureCoordinator {
     proc.barrierReleasedByMember(member, dataFromMember);
   }
 
-  /**
-   * @return the rpcs implementation for all current procedures
-   */
+  /** Returns the rpcs implementation for all current procedures */
   ProcedureCoordinatorRpcs getRpcs() {
     return rpcs;
   }
@@ -306,9 +304,7 @@ public class ProcedureCoordinator {
     return procedures.get(name);
   }
 
-  /**
-   * @return Return set of all procedure names.
-   */
+  /** Returns Return set of all procedure names. */
   public Set<String> getProcedureNames() {
     return new HashSet<>(procedures.keySet());
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/MasterQuotaManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/MasterQuotaManager.java
index 1132c4fcfe7..644ae6c887c 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/MasterQuotaManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/MasterQuotaManager.java
@@ -523,9 +523,7 @@ public class MasterQuotaManager implements RegionStateListener {
     }
   }
 
-  /**
-   * @return cached region count, or -1 if quota manager is disabled or table status not found
-   */
+  /** Returns cached region count, or -1 if quota manager is disabled or table status not found */
   public int getRegionCountOfTable(TableName tName) throws IOException {
     if (initialized) {
       return namespaceQuotaManager.getRegionCountOfTable(tName);
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/OperationQuota.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/OperationQuota.java
index 7ba88e51fd4..aaae64b6184 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/OperationQuota.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/OperationQuota.java
@@ -67,6 +67,6 @@ public interface OperationQuota {
    */
   void addMutation(Mutation mutation);
 
-  /** @return the number of bytes available to read to avoid exceeding the quota */
+  /** Returns the number of bytes available to read to avoid exceeding the quota */
   long getReadAvailable();
 }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaLimiter.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaLimiter.java
index bbe2161a327..14326e4e0d2 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaLimiter.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaLimiter.java
@@ -70,12 +70,12 @@ public interface QuotaLimiter {
    */
   void consumeRead(long size, long capacityUnit);
 
-  /** @return true if the limiter is a noop */
+  /** Returns true if the limiter is a noop */
   boolean isBypass();
 
-  /** @return the number of bytes available to read to avoid exceeding the quota */
+  /** Returns the number of bytes available to read to avoid exceeding the quota */
   long getReadAvailable();
 
-  /** @return the number of bytes available to write to avoid exceeding the quota */
+  /** Returns the number of bytes available to write to avoid exceeding the quota */
   long getWriteAvailable();
 }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaState.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaState.java
index 6f7b351ef09..7c9445e1558 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaState.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaState.java
@@ -69,9 +69,7 @@ public class QuotaState {
     return builder.toString();
   }
 
-  /**
-   * @return true if there is no quota information associated to this object
-   */
+  /** Returns true if there is no quota information associated to this object */
   public synchronized boolean isBypass() {
     return globalLimiter == NoopQuotaLimiter.get();
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/RateLimiter.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/RateLimiter.java
index 9e258e70951..b83e45e9407 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/RateLimiter.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/RateLimiter.java
@@ -199,15 +199,13 @@ public abstract class RateLimiter {
     }
   }
 
-  /**
-   * @return estimate of the ms required to wait before being able to provide 1 resource.
-   */
+  /** Returns estimate of the ms required to wait before being able to provide 1 resource. */
   public long waitInterval() {
     return waitInterval(1);
   }
 
   /**
-   * @return estimate of the ms required to wait before being able to provide "amount" resources.
+   * Returns estimate of the ms required to wait before being able to provide "amount" resources.
    */
   public synchronized long waitInterval(final long amount) {
     // TODO Handle over quota?
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/RegionServerSpaceQuotaManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/RegionServerSpaceQuotaManager.java
index ea194da1430..0233e33312f 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/RegionServerSpaceQuotaManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/RegionServerSpaceQuotaManager.java
@@ -108,9 +108,7 @@ public class RegionServerSpaceQuotaManager {
     started = false;
   }
 
-  /**
-   * @return if the {@code Chore} has been started.
-   */
+  /** Returns if the {@code Chore} has been started. */
   public boolean isStarted() {
     return started;
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/UserQuotaState.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/UserQuotaState.java
index 4759672570e..6078c687d3d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/UserQuotaState.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/UserQuotaState.java
@@ -83,9 +83,7 @@ public class UserQuotaState extends QuotaState {
     return builder.toString();
   }
 
-  /**
-   * @return true if there is no quota information associated to this object
-   */
+  /** Returns true if there is no quota information associated to this object */
   @Override
   public synchronized boolean isBypass() {
     return !bypassGlobals && getGlobalLimiterWithoutUpdatingLastQuery() == NoopQuotaLimiter.get()
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMemStore.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMemStore.java
index af954474014..179c617462f 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMemStore.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMemStore.java
@@ -218,9 +218,7 @@ public abstract class AbstractMemStore implements MemStore {
     }
   }
 
-  /**
-   * @return Oldest timestamp of all the Cells in the MemStore
-   */
+  /** Returns Oldest timestamp of all the Cells in the MemStore */
   @Override
   public long timeOfOldestEdit() {
     return timeOfOldestEdit;
@@ -356,7 +354,7 @@ public abstract class AbstractMemStore implements MemStore {
   }
 
   /**
-   * @return The total size of cells in this memstore. We will not consider cells in the snapshot
+   * Returns The total size of cells in this memstore. We will not consider cells in the snapshot
    */
   protected abstract long keySize();
 
@@ -378,9 +376,7 @@ public abstract class AbstractMemStore implements MemStore {
     return snapshot;
   }
 
-  /**
-   * @return an ordered list of segments from most recent to oldest in memstore
-   */
+  /** Returns an ordered list of segments from most recent to oldest in memstore */
   protected abstract List<Segment> getSegments() throws IOException;
 
 }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ChangedReadersObserver.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ChangedReadersObserver.java
index a7a3dc800ce..d45a8046873 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ChangedReadersObserver.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ChangedReadersObserver.java
@@ -27,9 +27,7 @@ import org.apache.yetus.audience.InterfaceAudience;
 @InterfaceAudience.Private
 public interface ChangedReadersObserver {
 
-  /**
-   * @return the read point of the current scan
-   */
+  /** Returns the read point of the current scan */
   long getReadPoint();
 
   /**
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Chunk.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Chunk.java
index eb6282bf163..5e37ce4ee72 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Chunk.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Chunk.java
@@ -171,9 +171,7 @@ public abstract class Chunk {
     }
   }
 
-  /**
-   * @return This chunk's backing data.
-   */
+  /** Returns This chunk's backing data. */
   ByteBuffer getData() {
     return this.data;
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplit.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplit.java
index 6719db72c8c..8a6e806292a 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplit.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplit.java
@@ -497,9 +497,7 @@ public class CompactSplit implements CompactionRequester, PropagatingConfigurati
     return (regionSplitLimit > server.getNumberOfOnlineRegions());
   }
 
-  /**
-   * @return the regionSplitLimit
-   */
+  /** Returns the regionSplitLimit */
   public int getRegionSplitLimit() {
     return this.regionSplitLimit;
   }
@@ -856,16 +854,12 @@ public class CompactSplit implements CompactionRequester, PropagatingConfigurati
     this.conf.set(HBASE_REGION_SERVER_ENABLE_COMPACTION, String.valueOf(compactionsEnabled));
   }
 
-  /**
-   * @return the longCompactions thread pool executor
-   */
+  /** Returns the longCompactions thread pool executor */
   ThreadPoolExecutor getLongCompactions() {
     return longCompactions;
   }
 
-  /**
-   * @return the shortCompactions thread pool executor
-   */
+  /** Returns the shortCompactions thread pool executor */
   ThreadPoolExecutor getShortCompactions() {
     return shortCompactions;
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompositeImmutableSegment.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompositeImmutableSegment.java
index 57cc0c6f21a..f955eb5d582 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompositeImmutableSegment.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompositeImmutableSegment.java
@@ -58,9 +58,7 @@ public class CompositeImmutableSegment extends ImmutableSegment {
     return segments.size();
   }
 
-  /**
-   * @return whether the segment has any cells
-   */
+  /** Returns whether the segment has any cells */
   @Override
   public boolean isEmpty() {
     for (ImmutableSegment s : segments) {
@@ -69,9 +67,7 @@ public class CompositeImmutableSegment extends ImmutableSegment {
     return true;
   }
 
-  /**
-   * @return number of cells in segment
-   */
+  /** Returns number of cells in segment */
   @Override
   public int getCellsCount() {
     int result = 0;
@@ -160,17 +156,13 @@ public class CompositeImmutableSegment extends ImmutableSegment {
     return false;
   }
 
-  /**
-   * @return Sum of all cell sizes.
-   */
+  /** Returns Sum of all cell sizes. */
   @Override
   public long getDataSize() {
     return this.keySize;
   }
 
-  /**
-   * @return The heap size of this segment.
-   */
+  /** Returns The heap size of this segment. */
   @Override
   public long getHeapSize() {
     long result = 0;
@@ -224,9 +216,7 @@ public class CompositeImmutableSegment extends ImmutableSegment {
     throw new IllegalStateException("Not supported by CompositeImmutableScanner");
   }
 
-  /**
-   * @return a set of all cells in the segment
-   */
+  /** Returns a set of all cells in the segment */
   @Override
   protected CellSet getCellSet() {
     throw new IllegalStateException("Not supported by CompositeImmutableScanner");
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ConstantSizeRegionSplitPolicy.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ConstantSizeRegionSplitPolicy.java
index e0d7627a929..be8fbb1b320 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ConstantSizeRegionSplitPolicy.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ConstantSizeRegionSplitPolicy.java
@@ -90,9 +90,7 @@ public class ConstantSizeRegionSplitPolicy extends RegionSplitPolicy {
     return this.jitterRate > 0;
   }
 
-  /**
-   * @return true if region size exceed the sizeToCheck
-   */
+  /** Returns true if region size exceed the sizeToCheck */
   protected final boolean isExceedSize(long sizeToCheck) {
     if (overallHRegionFiles) {
       long sumSize = 0;
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/FlushNonSloppyStoresFirstPolicy.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/FlushNonSloppyStoresFirstPolicy.java
index 75277fdcab6..b130373ab74 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/FlushNonSloppyStoresFirstPolicy.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/FlushNonSloppyStoresFirstPolicy.java
@@ -33,9 +33,7 @@ public class FlushNonSloppyStoresFirstPolicy extends FlushLargeStoresPolicy {
   private Collection<HStore> regularStores = new HashSet<>();
   private Collection<HStore> sloppyStores = new HashSet<>();
 
-  /**
-   * @return the stores need to be flushed.
-   */
+  /** Returns the stores need to be flushed. */
   @Override
   public Collection<HStore> selectStoresToFlush() {
     Collection<HStore> specificStoresToFlush = new HashSet<>();
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/FlushPolicy.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/FlushPolicy.java
index fb9f8c8cb7a..401f79ce383 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/FlushPolicy.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/FlushPolicy.java
@@ -40,8 +40,6 @@ public abstract class FlushPolicy extends Configured {
     this.region = region;
   }
 
-  /**
-   * @return the stores need to be flushed.
-   */
+  /** Returns the stores need to be flushed. */
   public abstract Collection<HStore> selectStoresToFlush();
 }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
index 7ee26d2ec92..89117fc79a0 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
@@ -622,9 +622,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
     private boolean failedSanityCheck = false;
     private boolean wrongFamily = false;
 
-    /**
-     * @return If a {@link WrongRegionException} has been observed.
-     */
+    /** Returns If a {@link WrongRegionException} has been observed. */
     boolean hasSeenWrongRegion() {
       return wrongRegion;
     }
@@ -636,9 +634,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
       wrongRegion = true;
     }
 
-    /**
-     * @return If a {@link FailedSanityCheckException} has been observed.
-     */
+    /** Returns If a {@link FailedSanityCheckException} has been observed. */
     boolean hasSeenFailedSanityCheck() {
       return failedSanityCheck;
     }
@@ -650,9 +646,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
       failedSanityCheck = true;
     }
 
-    /**
-     * @return If a {@link NoSuchColumnFamilyException} has been observed.
-     */
+    /** Returns If a {@link NoSuchColumnFamilyException} has been observed. */
     boolean hasSeenNoSuchFamily() {
       return wrongFamily;
     }
@@ -1174,9 +1168,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
     }
   }
 
-  /**
-   * @return Map of StoreFiles by column family
-   */
+  /** Returns Map of StoreFiles by column family */
   private NavigableMap<byte[], List<Path>> getStoreFiles() {
     NavigableMap<byte[], List<Path>> allStoreFiles = new TreeMap<>(Bytes.BYTES_COMPARATOR);
     for (HStore store : stores.values()) {
@@ -1219,9 +1211,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
     }
   }
 
-  /**
-   * @return True if this region has references.
-   */
+  /** Returns True if this region has references. */
   public boolean hasReferences() {
     return stores.values().stream().anyMatch(HStore::hasReferences);
   }
@@ -1344,9 +1334,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
     return this.fs.getRegionInfo();
   }
 
-  /**
-   * @return Instance of {@link RegionServerServices} used by this HRegion. Can be null.
-   */
+  /** Returns Instance of {@link RegionServerServices} used by this HRegion. Can be null. */
   RegionServerServices getRegionServerServices() {
     return this.rsServices;
   }
@@ -1381,7 +1369,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
     return memStoreSizing.getOffHeapSize();
   }
 
-  /** @return store services for this region, to access services required by store level needs */
+  /** Returns store services for this region, to access services required by store level needs */
   public RegionServicesForStores getRegionServicesForStores() {
     return regionServicesForStores;
   }
@@ -1472,9 +1460,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
     return maxFlushedSeqId;
   }
 
-  /**
-   * @return readpoint considering given IsolationLevel. Pass {@code null} for default
-   */
+  /** Returns readpoint considering given IsolationLevel. Pass {@code null} for default */
   public long getReadPoint(IsolationLevel isolationLevel) {
     if (isolationLevel != null && isolationLevel == IsolationLevel.READ_UNCOMMITTED) {
       // This scan can read even uncommitted transactions
@@ -1955,9 +1941,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
       });
   }
 
-  /**
-   * @return True if its worth doing a flush before we put up the close flag.
-   */
+  /** Returns True if its worth doing a flush before we put up the close flag. */
   private boolean worthPreFlushing() {
     return this.memStoreSizing.getDataSize()
         > this.conf.getLong("hbase.hregion.preclose.flush.size", 1024 * 1024 * 5);
@@ -1978,7 +1962,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
     htableDescriptor = desc;
   }
 
-  /** @return WAL in use for this region */
+  /** Returns WAL in use for this region */
   public WAL getWAL() {
     return this.wal;
   }
@@ -2005,9 +1989,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
     this.mobFileCache = mobFileCache;
   }
 
-  /**
-   * @return split policy for this region.
-   */
+  /** Returns split policy for this region. */
   RegionSplitPolicy getSplitPolicy() {
     return this.splitPolicy;
   }
@@ -2023,23 +2005,23 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
     return this.baseConf;
   }
 
-  /** @return {@link FileSystem} being used by this region */
+  /** Returns {@link FileSystem} being used by this region */
   public FileSystem getFilesystem() {
     return fs.getFileSystem();
   }
 
-  /** @return the {@link HRegionFileSystem} used by this region */
+  /** Returns the {@link HRegionFileSystem} used by this region */
   public HRegionFileSystem getRegionFileSystem() {
     return this.fs;
   }
 
-  /** @return the WAL {@link HRegionFileSystem} used by this region */
+  /** Returns the WAL {@link HRegionFileSystem} used by this region */
   HRegionWALFileSystem getRegionWALFileSystem() throws IOException {
     return new HRegionWALFileSystem(conf, getWalFileSystem(),
       CommonFSUtils.getWALTableDir(conf, htableDescriptor.getTableName()), fs.getRegionInfo());
   }
 
-  /** @return the WAL {@link FileSystem} being used by this region */
+  /** Returns the WAL {@link FileSystem} being used by this region */
   FileSystem getWalFileSystem() throws IOException {
     if (walFS == null) {
       walFS = CommonFSUtils.getWALFileSystem(conf);
@@ -2336,13 +2318,13 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
       CANNOT_FLUSH
     }
 
-    /** @return the detailed result code */
+    /** Returns the detailed result code */
     Result getResult();
 
-    /** @return true if the memstores were flushed, else false */
+    /** Returns true if the memstores were flushed, else false */
     boolean isFlushSucceeded();
 
-    /** @return True if the flush requested a compaction, else false */
+    /** Returns True if the flush requested a compaction, else false */
     boolean isCompactionNeeded();
   }
 
@@ -2792,9 +2774,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
     }
   }
 
-  /**
-   * @return True if passed Set is all families in the region.
-   */
+  /** Returns True if passed Set is all families in the region. */
   private boolean isAllFamilies(Collection<HStore> families) {
     return families == null || this.stores.size() == families.size();
   }
@@ -4010,9 +3990,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
       }
     }
 
-    /**
-     * @return Get the long out of the passed in Cell
-     */
+    /** Returns Get the long out of the passed in Cell */
     private static long getLongValue(final Cell cell) throws DoNotRetryIOException {
       int len = cell.getValueLength();
       if (len != Bytes.SIZEOF_LONG) {
@@ -7486,9 +7464,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
     });
   }
 
-  /**
-   * @return statistics about the current load of the region
-   */
+  /** Returns statistics about the current load of the region */
   public ClientProtos.RegionLoadStats getLoadStatistics() {
     if (!regionStatsEnabled) {
       return null;
@@ -7779,9 +7755,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
       SequenceId.NO_SEQUENCE_ID);
   }
 
-  /**
-   * @return writeEntry associated with this append
-   */
+  /** Returns writeEntry associated with this append */
   private WriteEntry doWALAppend(WALEdit walEdit, Durability durability, List<UUID> clusterIds,
     long now, long nonceGroup, long nonce, long origLogSeqNum) throws IOException {
     Preconditions.checkArgument(walEdit != null && !walEdit.isEmpty(), "WALEdit is null or empty!");
@@ -8002,9 +7976,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
     }
   }
 
-  /**
-   * @return The priority that this region should have in the compaction queue
-   */
+  /** Returns The priority that this region should have in the compaction queue */
   public int getCompactPriority() {
     if (conf.getBoolean(SPLIT_IGNORE_BLOCKING_ENABLED_KEY, false) && checkSplit().isPresent()) {
       // if a region should split, split it before compact
@@ -8014,7 +7986,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
       .orElse(Store.NO_PRIORITY);
   }
 
-  /** @return the coprocessor host */
+  /** Returns the coprocessor host */
   public RegionCoprocessorHost getCoprocessorHost() {
     return coprocessorHost;
   }
@@ -8243,7 +8215,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
     return regionDurability.ordinal() > Durability.ASYNC_WAL.ordinal();
   }
 
-  /** @return the latest sequence number that was read from storage when this region was opened */
+  /** Returns the latest sequence number that was read from storage when this region was opened */
   public long getOpenSeqNum() {
     return this.openSeqNum;
   }
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionFileSystem.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionFileSystem.java
index 663e2b73ae2..2d2965d7237 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionFileSystem.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionFileSystem.java
@@ -124,12 +124,12 @@ public class HRegionFileSystem {
       conf.getInt("hdfs.client.sleep.before.retries", DEFAULT_BASE_SLEEP_BEFORE_RETRIES);
   }
 
-  /** @return the underlying {@link FileSystem} */
+  /** Returns the underlying {@link FileSystem} */
   public FileSystem getFileSystem() {
     return this.fs;
   }
 
-  /** @return the {@link RegionInfo} that describe this on-disk region view */
+  /** Returns the {@link RegionInfo} that describe this on-disk region view */
   public RegionInfo getRegionInfo() {
     return this.regionInfo;
   }
@@ -138,12 +138,12 @@ public class HRegionFileSystem {
     return this.regionInfoForFs;
   }
 
-  /** @return {@link Path} to the region's root directory. */
+  /** Returns {@link Path} to the region's root directory. */
   public Path getTableDir() {
     return this.tableDir;
   }
 
-  /** @return {@link Path} to the region directory. */
+  /** Returns {@link Path} to the region directory. */
   public Path getRegionDir() {
     return regionDir;
   }
@@ -151,7 +151,7 @@ public class HRegionFileSystem {
   // ===========================================================================
   // Temp Helpers
   // ===========================================================================
-  /** @return {@link Path} to the region's temp directory, used for file creations */
+  /** Returns {@link Path} to the region's temp directory, used for file creations */
   public Path getTempDir() {
     return new Path(getRegionDir(), REGION_TEMP_DIR);
   }
@@ -379,9 +379,7 @@ public class HRegionFileSystem {
     return false;
   }
 
-  /**
-   * @return the set of families present on disk n
-   */
+  /** Returns the set of families present on disk n */
   public Collection<String> getFamilies() throws IOException {
     FileStatus[] fds =
       CommonFSUtils.listStatus(fs, getRegionDir(), new FSUtils.FamilyDirFilter(fs));
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java
index 4e74034e51b..5408685311e 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java
@@ -1020,9 +1020,7 @@ public class HRegionServer extends Thread
     }
   }
 
-  /**
-   * @return True if the cluster is up.
-   */
+  /** Returns True if the cluster is up. */
   @Override
   public boolean isClusterUp() {
     return this.masterless
@@ -1314,9 +1312,7 @@ public class HRegionServer extends Thread
     return allUserRegionsOffline;
   }
 
-  /**
-   * @return Current write count for all online regions.
-   */
+  /** Returns Current write count for all online regions. */
   private long getWriteRequestCount() {
     long writeCount = 0;
     for (Map.Entry<String, HRegion> e : this.onlineRegions.entrySet()) {
@@ -2066,9 +2062,7 @@ public class HRegionServer extends Thread
     }
   }
 
-  /**
-   * @return Master address tracker instance.
-   */
+  /** Returns Master address tracker instance. */
   public MasterAddressTracker getMasterAddressTracker() {
     return this.masterAddressTracker;
   }
@@ -2821,17 +2815,13 @@ public class HRegionServer extends Thread
     }
   }
 
-  /**
-   * @return Return the object that implements the replication source executorService.
-   */
+  /** Returns Return the object that implements the replication source executorService. */
   @InterfaceAudience.Private
   public ReplicationSourceService getReplicationSourceService() {
     return replicationSourceHandler;
   }
... 3617 lines suppressed ...