You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by zh...@apache.org on 2022/04/09 15:55:12 UTC

[hbase] branch branch-2 updated: HBASE-26922 Fix LineLength warnings as much as possible if it can not be fixed by spotless (#4324)

This is an automated email from the ASF dual-hosted git repository.

zhangduo pushed a commit to branch branch-2
in repository https://gitbox.apache.org/repos/asf/hbase.git


The following commit(s) were added to refs/heads/branch-2 by this push:
     new e7eb628025f HBASE-26922 Fix LineLength warnings as much as possible if it can not be fixed by spotless (#4324)
e7eb628025f is described below

commit e7eb628025f9894dda5290a5950e76306c517f84
Author: Duo Zhang <zh...@apache.org>
AuthorDate: Sat Apr 9 21:38:41 2022 +0800

    HBASE-26922 Fix LineLength warnings as much as possible if it can not be fixed by spotless (#4324)
    
    Signed-off-by: Yulin Niu <niuyulin@apache.org
    (cherry picked from commit 3ae0d9012c6a0132147f2cd1a96337213bbf2b8d)
---
 .../io/asyncfs/monitor/StreamSlowMonitor.java      |  10 +-
 .../CatalogReplicaLoadBalanceSimpleSelector.java   |   4 +-
 .../hadoop/hbase/filter/ColumnCountGetFilter.java  |   1 -
 .../apache/hadoop/hbase/filter/FuzzyRowFilter.java |   4 +-
 .../hbase/filter/SingleColumnValueFilter.java      |   1 -
 .../hadoop/hbase/shaded/protobuf/ProtobufUtil.java |  29 ++----
 .../apache/hadoop/hbase/CellComparatorImpl.java    |   6 +-
 .../java/org/apache/hadoop/hbase/ExtendedCell.java |   7 +-
 .../java/org/apache/hadoop/hbase/KeyValue.java     |  22 +++--
 .../org/apache/hadoop/hbase/PrivateCellUtil.java   |  68 ++++++++-----
 .../hadoop/hbase/util/CoprocessorClassLoader.java  |  31 +++---
 .../org/apache/hadoop/hbase/thrift/DemoClient.java |   4 +-
 .../regionserver/MetricsRegionServerSource.java    |   4 +-
 .../hadoop/hbase/IntegrationTestIngestWithACL.java |  19 ++--
 .../ServerAndDependenciesKillingMonkeyFactory.java |   8 +-
 .../factories/ServerKillingMonkeyFactory.java      |  10 +-
 .../hadoop/hbase/mapred/TableInputFormatBase.java  |  11 +--
 .../hbase/mapreduce/MultiTableInputFormatBase.java |   6 +-
 .../mapreduce/MultiTableSnapshotInputFormat.java   |   3 +-
 .../MultiTableSnapshotInputFormatImpl.java         |   5 +-
 .../hbase/mapreduce/MultithreadedTableMapper.java  |   5 +-
 .../hadoop/hbase/mob/mapreduce/MobRefReporter.java | 109 ++++++++++-----------
 .../hadoop/hbase/mapreduce/TestImportExport.java   |  28 +++---
 .../hadoop/hbase/rest/model/TestRowModel.java      |   5 +-
 .../hbase/errorhandling/ForeignException.java      |   1 -
 .../org/apache/hadoop/hbase/io/hfile/HFile.java    |   4 +-
 .../hadoop/hbase/master/HMasterCommandLine.java    |   9 +-
 .../hadoop/hbase/master/MasterWalManager.java      |   6 +-
 .../hadoop/hbase/master/SplitLogManager.java       |   5 +-
 .../master/normalizer/SimpleRegionNormalizer.java  |   6 +-
 .../master/procedure/CloneSnapshotProcedure.java   |  29 +++---
 .../master/procedure/RestoreSnapshotProcedure.java |  30 +++---
 .../impl/BalancerRejectionQueueService.java        |   4 +-
 .../apache/hadoop/hbase/quotas/UserQuotaState.java |   7 +-
 .../hbase/regionserver/AbstractMemStore.java       |   6 +-
 .../hadoop/hbase/regionserver/RSRpcServices.java   |   5 +-
 .../regionserver/ReplicationSource.java            |   8 +-
 .../regionserver/ReplicationSourceShipper.java     |  13 +--
 .../regionserver/ReplicationSourceWALReader.java   |   8 +-
 .../hadoop/hbase/security/token/TokenUtil.java     |  10 +-
 .../org/apache/hadoop/hbase/wal/EntryBuffers.java  |   4 +-
 .../apache/hadoop/hbase/HBaseTestingUtility.java   |   5 +-
 .../TestRegionObserverScannerOpenHook.java         |  30 +++---
 .../favored/TestFavoredNodeAssignmentHelper.java   |   5 +-
 .../hbase/io/hfile/TestFixedFileTrailer.java       |   7 +-
 .../hadoop/hbase/io/hfile/TestHFileBlock.java      |  45 ++++-----
 .../TestSnapshotProcedureSnapshotCorrupted.java    |   7 +-
 .../hbase/regionserver/TestMetricsRegion.java      |  46 ++++-----
 .../TestRegionMergeTransactionOnCluster.java       |  35 ++++---
 .../hbase/regionserver/TestStoreScanner.java       |  58 ++++++-----
 .../TestCompactionWithThroughputController.java    |  27 ++---
 .../TestRefreshPeerWhileRegionServerRestarts.java  |   6 +-
 .../apache/hadoop/hbase/zookeeper/TestZKMulti.java |  14 ++-
 53 files changed, 435 insertions(+), 405 deletions(-)

diff --git a/hbase-asyncfs/src/main/java/org/apache/hadoop/hbase/io/asyncfs/monitor/StreamSlowMonitor.java b/hbase-asyncfs/src/main/java/org/apache/hadoop/hbase/io/asyncfs/monitor/StreamSlowMonitor.java
index 7ee04f8eebd..73cce189574 100644
--- a/hbase-asyncfs/src/main/java/org/apache/hadoop/hbase/io/asyncfs/monitor/StreamSlowMonitor.java
+++ b/hbase-asyncfs/src/main/java/org/apache/hadoop/hbase/io/asyncfs/monitor/StreamSlowMonitor.java
@@ -69,8 +69,8 @@ public class StreamSlowMonitor implements ConfigurationObserver {
    */
   private static final String DATANODE_PACKET_FLUSH_CHECK_SPEED_MIN_DATA_LENGTH_KEY =
     "hbase.regionserver.async.wal.datanode.slow.check.speed.packet.data.length.min";
-  private static final long DEFAULT_DATANODE_PACKET_FLUSH_CHECK_SPEED_MIN_DATA_LENGTH =
-    64 * 1024; //64KB
+  // 64KB
+  private static final long DEFAULT_DATANODE_PACKET_FLUSH_CHECK_SPEED_MIN_DATA_LENGTH = 64 * 1024;
 
   /**
    * Configure for the slow packet process time, a duration from send to ACK.
@@ -79,7 +79,8 @@ public class StreamSlowMonitor implements ConfigurationObserver {
    */
   public static final String DATANODE_SLOW_PACKET_PROCESS_TIME_KEY =
     "hbase.regionserver.async.wal.datanode.slow.packet.process.time.millis";
-  private static final long DEFAULT_DATANODE_SLOW_PACKET_PROCESS_TIME = 6000; // 6s in ms
+  // 6s in ms
+  private static final long DEFAULT_DATANODE_SLOW_PACKET_PROCESS_TIME = 6000;
 
   /**
    * Configure for the check of large packet(which is configured by
@@ -89,7 +90,8 @@ public class StreamSlowMonitor implements ConfigurationObserver {
    */
   private static final String DATANODE_SLOW_PACKET_FLUSH_MIN_SPEED_KEY =
     "hbase.regionserver.async.wal.datanode.slow.packet.speed.min.kbs";
-  private static final double DEFAULT_DATANODE_SLOW_PACKET_FLUSH_MIN_SPEED = 20; // 20KB/s
+  // 20KB/s
+  private static final double DEFAULT_DATANODE_SLOW_PACKET_FLUSH_MIN_SPEED = 20;
 
   private final String name;
   // this is a map of datanodeInfo->queued slow PacketAckData
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CatalogReplicaLoadBalanceSimpleSelector.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CatalogReplicaLoadBalanceSimpleSelector.java
index f78dfb199c1..bbafc650b29 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CatalogReplicaLoadBalanceSimpleSelector.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CatalogReplicaLoadBalanceSimpleSelector.java
@@ -103,8 +103,8 @@ class CatalogReplicaLoadBalanceSimpleSelector implements
     }
   }
 
-  private final ConcurrentMap<TableName, ConcurrentNavigableMap<byte[], StaleLocationCacheEntry>>
-    staleCache = new ConcurrentHashMap<>();
+  private final ConcurrentMap<TableName,
+    ConcurrentNavigableMap<byte[], StaleLocationCacheEntry>> staleCache = new ConcurrentHashMap<>();
   private volatile int numOfReplicas;
   private final ChoreService choreService;
   private final TableName tableName;
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java
index 4ecf2d7bf4d..f880d5f2aac 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java
@@ -101,7 +101,6 @@ public class ColumnCountGetFilter extends FilterBase {
   /**
    * @param pbBytes A pb serialized {@link ColumnCountGetFilter} instance
    * @return An instance of {@link ColumnCountGetFilter} made from <code>bytes</code>
-   * @throws org.apache.hadoop.hbase.exceptions.DeserializationException
    * @see #toByteArray
    */
   public static ColumnCountGetFilter parseFrom(final byte [] pbBytes)
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java
index 5abcee2a9fd..6f8e9916255 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java
@@ -241,8 +241,8 @@ public class FuzzyRowFilter extends FilterBase {
 
     byte[] nextRow() {
       if (nextRows.isEmpty()) {
-        throw new IllegalStateException(
-            "NextRows should not be empty, make sure to call nextRow() after updateTracker() return true");
+        throw new IllegalStateException("NextRows should not be empty, "
+          + "make sure to call nextRow() after updateTracker() return true");
       } else {
         return nextRows.peek().getFirst();
       }
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java
index b4303cdb76d..f314bede082 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java
@@ -399,7 +399,6 @@ public class SingleColumnValueFilter extends FilterBase {
   /**
    * @param pbBytes A pb serialized {@link SingleColumnValueFilter} instance
    * @return An instance of {@link SingleColumnValueFilter} made from <code>bytes</code>
-   * @throws org.apache.hadoop.hbase.exceptions.DeserializationException
    * @see #toByteArray
    */
   public static SingleColumnValueFilter parseFrom(final byte [] pbBytes)
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java
index 523b2242414..436d9eebbe3 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java
@@ -2958,9 +2958,7 @@ public final class ProtobufUtil {
   }
 
   /**
-   * Creates {@link CompactionState} from
-   * {@link org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState}
-   * state
+   * Creates {@link CompactionState} from {@link GetRegionInfoResponse.CompactionState} state
    * @param state the protobuf CompactionState
    * @return CompactionState
    */
@@ -2973,9 +2971,7 @@ public final class ProtobufUtil {
   }
 
   /**
-   * Creates {@link CompactionState} from
-   * {@link org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos
-   * .RegionLoad.CompactionState} state
+   * Creates {@link CompactionState} from {@link RegionLoad.CompactionState} state
    * @param state the protobuf CompactionState
    * @return CompactionState
    */
@@ -2995,8 +2991,7 @@ public final class ProtobufUtil {
   }
 
   /**
-   * Creates {@link org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Type}
-   * from {@link SnapshotType}
+   * Creates {@link SnapshotProtos.SnapshotDescription.Type} from {@link SnapshotType}
    * @param type the SnapshotDescription type
    * @return the protobuf SnapshotDescription type
    */
@@ -3006,8 +3001,8 @@ public final class ProtobufUtil {
   }
 
   /**
-   * Creates {@link org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Type}
-   * from the type of SnapshotDescription string
+   * Creates {@link SnapshotProtos.SnapshotDescription.Type} from the type of SnapshotDescription
+   * string
    * @param snapshotDesc string representing the snapshot description type
    * @return the protobuf SnapshotDescription type
    */
@@ -3017,18 +3012,16 @@ public final class ProtobufUtil {
   }
 
   /**
-   * Creates {@link SnapshotType} from the type of
-   * {@link org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription}
-   * @param type the snapshot description type
-   * @return the protobuf SnapshotDescription type
+   * Creates {@link SnapshotType} from the {@link SnapshotProtos.SnapshotDescription.Type}
+   * @param  type the snapshot description type
+   * @return      the protobuf SnapshotDescription type
    */
   public static SnapshotType createSnapshotType(SnapshotProtos.SnapshotDescription.Type type) {
     return SnapshotType.valueOf(type.toString());
   }
 
   /**
-   * Convert from {@link SnapshotDescription} to
-   * {@link org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription}
+   * Convert from {@link SnapshotDescription} to {@link SnapshotProtos.SnapshotDescription}
    * @param snapshotDesc the POJO SnapshotDescription
    * @return the protobuf SnapshotDescription
    */
@@ -3062,9 +3055,7 @@ public final class ProtobufUtil {
   }
 
   /**
-   * Convert from
-   * {@link org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription} to
-   * {@link SnapshotDescription}
+   * Convert from {@link SnapshotProtos.SnapshotDescription} to {@link SnapshotDescription}
    * @param snapshotDesc the protobuf SnapshotDescription
    * @return the POJO SnapshotDescription
    */
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellComparatorImpl.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellComparatorImpl.java
index a6be2378560..7b995b0b8cd 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellComparatorImpl.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellComparatorImpl.java
@@ -38,9 +38,9 @@ import org.apache.yetus.audience.InterfaceStability;
  * {@link BBKVComparator}. Use it when mostly {@link ByteBufferKeyValue}s.
  * </p>
  */
-@edu.umd.cs.findbugs.annotations.SuppressWarnings(
-    value="UNKNOWN",
-    justification="Findbugs doesn't like the way we are negating the result of a compare in below")
+@edu.umd.cs.findbugs.annotations.SuppressWarnings(value = "UNKNOWN",
+    justification = "Findbugs doesn't like the way we are negating the result of"
+      + " a compare in below")
 @InterfaceAudience.Private
 @InterfaceStability.Evolving
 public class CellComparatorImpl implements CellComparator {
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/ExtendedCell.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/ExtendedCell.java
index 20f8e59ef08..bdb457a0063 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/ExtendedCell.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/ExtendedCell.java
@@ -73,13 +73,14 @@ public interface ExtendedCell extends RawCell, HeapSize {
   }
 
   /**
-   * @param withTags Whether to write tags.
-   * @return Bytes count required to serialize this Cell in a {@link KeyValue} format.
-   * <br> KeyValue format <br>
+   * KeyValue format
+   * <p/>
    * <code>&lt;4 bytes keylength&gt; &lt;4 bytes valuelength&gt; &lt;2 bytes rowlength&gt;
    * &lt;row&gt; &lt;1 byte columnfamilylength&gt; &lt;columnfamily&gt; &lt;columnqualifier&gt;
    * &lt;8 bytes timestamp&gt; &lt;1 byte keytype&gt; &lt;value&gt; &lt;2 bytes tagslength&gt;
    * &lt;tags&gt;</code>
+   * @param withTags Whether to write tags.
+   * @return Bytes count required to serialize this Cell in a {@link KeyValue} format.
    */
   // TODO remove the boolean param once HBASE-16706 is done.
   default int getSerializedSize(boolean withTags) {
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
index c05d0be3e56..14455ba110b 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
@@ -2369,21 +2369,23 @@ public class KeyValue implements ExtendedCell, Cloneable {
 
   /**
    * HeapSize implementation
-   *
+   * <p/>
    * We do not count the bytes in the rowCache because it should be empty for a KeyValue in the
    * MemStore.
    */
   @Override
   public long heapSize() {
-    /*
-     * Deep object overhead for this KV consists of two parts. The first part is the KV object
-     * itself, while the second part is the backing byte[]. We will only count the array overhead
-     * from the byte[] only if this is the first KV in there.
-     */
-    return ClassSize.align(FIXED_OVERHEAD) +
-        (offset == 0
-          ? ClassSize.sizeOfByteArray(length)  // count both length and object overhead
-          : length);                           // only count the number of bytes
+    // Deep object overhead for this KV consists of two parts. The first part is the KV object
+    // itself, while the second part is the backing byte[]. We will only count the array overhead
+    // from the byte[] only if this is the first KV in there.
+    int fixed = ClassSize.align(FIXED_OVERHEAD);
+    if (offset == 0) {
+      // count both length and object overhead
+      return fixed + ClassSize.sizeOfByteArray(length);
+    } else {
+      // only count the number of bytes
+      return fixed + length;
+    }
   }
 
   /**
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/PrivateCellUtil.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/PrivateCellUtil.java
index af61256cad0..54b95e3aa53 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/PrivateCellUtil.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/PrivateCellUtil.java
@@ -1589,11 +1589,13 @@ public final class PrivateCellUtil {
   }
 
   private static class FirstOnRowCell extends EmptyCell {
+    // @formatter:off
     private static final int FIXED_HEAPSIZE =
         ClassSize.OBJECT // object
       + ClassSize.REFERENCE // row array
       + Bytes.SIZEOF_INT // row offset
       + Bytes.SIZEOF_SHORT;  // row length
+    // @formatter:on
     private final byte[] rowArray;
     private final int roffset;
     private final short rlength;
@@ -1643,11 +1645,13 @@ public final class PrivateCellUtil {
   }
 
   private static class FirstOnRowByteBufferExtendedCell extends EmptyByteBufferExtendedCell {
+    // @formatter:off
     private static final int FIXED_OVERHEAD =
         ClassSize.OBJECT // object
         + ClassSize.REFERENCE // row buffer
         + Bytes.SIZEOF_INT // row offset
         + Bytes.SIZEOF_SHORT; // row length
+    // @formatter:on
     private final ByteBuffer rowBuff;
     private final int roffset;
     private final short rlength;
@@ -1698,11 +1702,12 @@ public final class PrivateCellUtil {
   }
 
   private static class LastOnRowByteBufferExtendedCell extends EmptyByteBufferExtendedCell {
-    private static final int FIXED_OVERHEAD =
-        ClassSize.OBJECT // object
+    // @formatter:off
+    private static final int FIXED_OVERHEAD = ClassSize.OBJECT // object
       + ClassSize.REFERENCE // rowBuff
       + Bytes.SIZEOF_INT // roffset
       + Bytes.SIZEOF_SHORT; // rlength
+    // @formatter:on
     private final ByteBuffer rowBuff;
     private final int roffset;
     private final short rlength;
@@ -1754,11 +1759,12 @@ public final class PrivateCellUtil {
 
   private static class FirstOnRowColByteBufferExtendedCell
       extends FirstOnRowByteBufferExtendedCell {
-    private static final int FIXED_OVERHEAD =
-        FirstOnRowByteBufferExtendedCell.FIXED_OVERHEAD
-        + ClassSize.REFERENCE * 2 // family buffer and column buffer
-        + Bytes.SIZEOF_INT * 3 // famOffset, colOffset, colLength
-        + Bytes.SIZEOF_BYTE; // famLength
+    // @formatter:off
+    private static final int FIXED_OVERHEAD = FirstOnRowByteBufferExtendedCell.FIXED_OVERHEAD
+      + ClassSize.REFERENCE * 2 // family buffer and column buffer
+      + Bytes.SIZEOF_INT * 3 // famOffset, colOffset, colLength
+      + Bytes.SIZEOF_BYTE; // famLength
+    // @formatter:on
     private final ByteBuffer famBuff;
     private final int famOffset;
     private final byte famLength;
@@ -1823,11 +1829,12 @@ public final class PrivateCellUtil {
   }
 
   private static class FirstOnRowColCell extends FirstOnRowCell {
-    private static final long FIXED_HEAPSIZE =
-        FirstOnRowCell.FIXED_HEAPSIZE
+    // @formatter:off
+    private static final long FIXED_HEAPSIZE = FirstOnRowCell.FIXED_HEAPSIZE
       + Bytes.SIZEOF_BYTE // flength
       + Bytes.SIZEOF_INT * 3 // foffset, qoffset, qlength
       + ClassSize.REFERENCE * 2; // fArray, qArray
+    // @formatter:on
     private final byte[] fArray;
     private final int foffset;
     private final byte flength;
@@ -1886,10 +1893,11 @@ public final class PrivateCellUtil {
   }
 
   private static class FirstOnRowColTSCell extends FirstOnRowColCell {
-    private static final long FIXED_HEAPSIZE =
-        FirstOnRowColCell.FIXED_HEAPSIZE
-            + Bytes.SIZEOF_LONG; // ts
+    // @formatter:off
+    private static final long FIXED_HEAPSIZE = FirstOnRowColCell.FIXED_HEAPSIZE
+      + Bytes.SIZEOF_LONG; // ts
     private long ts;
+    // @formatter:on
 
     public FirstOnRowColTSCell(byte[] rArray, int roffset, short rlength, byte[] fArray,
         int foffset, byte flength, byte[] qArray, int qoffset, int qlength, long ts) {
@@ -1910,10 +1918,11 @@ public final class PrivateCellUtil {
 
   private static class FirstOnRowColTSByteBufferExtendedCell
       extends FirstOnRowColByteBufferExtendedCell {
-    private static final int FIXED_OVERHEAD =
-        FirstOnRowColByteBufferExtendedCell.FIXED_OVERHEAD
-            + Bytes.SIZEOF_LONG; // ts
+    // @formatter:off
+    private static final int FIXED_OVERHEAD = FirstOnRowColByteBufferExtendedCell.FIXED_OVERHEAD
+      + Bytes.SIZEOF_LONG; // ts
     private long ts;
+    // @formatter:on
 
     public FirstOnRowColTSByteBufferExtendedCell(ByteBuffer rBuffer, int roffset, short rlength,
         ByteBuffer fBuffer, int foffset, byte flength, ByteBuffer qBuffer, int qoffset, int qlength,
@@ -1934,11 +1943,12 @@ public final class PrivateCellUtil {
   }
 
   private static class LastOnRowCell extends EmptyCell {
-    private static final int FIXED_OVERHEAD =
-        ClassSize.OBJECT // object
+    // @formatter:off
+    private static final int FIXED_OVERHEAD = ClassSize.OBJECT // object
       + ClassSize.REFERENCE // row array
       + Bytes.SIZEOF_INT // row offset
       + Bytes.SIZEOF_SHORT; // row length
+    // @formatter:on
     private final byte[] rowArray;
     private final int roffset;
     private final short rlength;
@@ -1988,10 +1998,12 @@ public final class PrivateCellUtil {
   }
 
   private static class LastOnRowColCell extends LastOnRowCell {
+    // @formatter:off
     private static final long FIXED_OVERHEAD = LastOnRowCell.FIXED_OVERHEAD
-        + ClassSize.REFERENCE * 2 // fArray and qArray
-        + Bytes.SIZEOF_INT * 3 // foffset, qoffset, qlength
-        + Bytes.SIZEOF_BYTE; // flength
+      + ClassSize.REFERENCE * 2 // fArray and qArray
+      + Bytes.SIZEOF_INT * 3 // foffset, qoffset, qlength
+      + Bytes.SIZEOF_BYTE; // flength
+    // @formatter:on
     private final byte[] fArray;
     private final int foffset;
     private final byte flength;
@@ -2050,11 +2062,12 @@ public final class PrivateCellUtil {
   }
 
   private static class LastOnRowColByteBufferExtendedCell extends LastOnRowByteBufferExtendedCell {
-    private static final int FIXED_OVERHEAD =
-        LastOnRowByteBufferExtendedCell.FIXED_OVERHEAD
-            + ClassSize.REFERENCE * 2 // fBuffer and qBuffer
-            + Bytes.SIZEOF_INT * 3 // foffset, qoffset, qlength
-            + Bytes.SIZEOF_BYTE; // flength
+    // @formatter:off
+    private static final int FIXED_OVERHEAD = LastOnRowByteBufferExtendedCell.FIXED_OVERHEAD
+      + ClassSize.REFERENCE * 2 // fBuffer and qBuffer
+      + Bytes.SIZEOF_INT * 3 // foffset, qoffset, qlength
+      + Bytes.SIZEOF_BYTE; // flength
+    // @formatter:on
     private final ByteBuffer fBuffer;
     private final int foffset;
     private final byte flength;
@@ -2119,11 +2132,12 @@ public final class PrivateCellUtil {
   }
 
   private static class FirstOnRowDeleteFamilyCell extends EmptyCell {
-    private static final int FIXED_OVERHEAD =
-        ClassSize.OBJECT // object
+    // @formatter:off
+    private static final int FIXED_OVERHEAD = ClassSize.OBJECT // object
       + ClassSize.REFERENCE * 2 // fBuffer and qBuffer
       + Bytes.SIZEOF_INT * 3 // foffset, qoffset, qlength
       + Bytes.SIZEOF_BYTE; // flength
+    // @formatter:on
     private final byte[] row;
     private final byte[] fam;
 
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CoprocessorClassLoader.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CoprocessorClassLoader.java
index f36ff1387d0..cb8a8e7f7b6 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CoprocessorClassLoader.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CoprocessorClassLoader.java
@@ -164,17 +164,22 @@ public class CoprocessorClassLoader extends ClassLoaderBase {
     }
 
     FileSystem fs = pathPattern.getFileSystem(conf);
-    Path pathPattern1 = fs.isDirectory(pathPattern) ?
-      new Path(pathPattern, "*.jar") : pathPattern;  // append "*.jar" if a directory is specified
-    FileStatus[] fileStatuses = fs.globStatus(pathPattern1);  // return all files that match the pattern
-    if (fileStatuses == null || fileStatuses.length == 0) {  // if no one matches
+    // append "*.jar" if a directory is specified
+    Path pathPattern1 = fs.isDirectory(pathPattern) ? new Path(pathPattern, "*.jar") : pathPattern;
+    // return all files that match the pattern
+    FileStatus[] fileStatuses = fs.globStatus(pathPattern1);
+    if (fileStatuses == null || fileStatuses.length == 0) {
+      // if no one matches
       throw new FileNotFoundException(pathPattern1.toString());
     } else {
       boolean validFileEncountered = false;
-      for (Path path : FileUtil.stat2Paths(fileStatuses)) {  // for each file that match the pattern
-        if (fs.isFile(path)) {  // only process files, skip for directories
-          File dst = new File(parentDirStr, "." + pathPrefix + "."
-            + path.getName() + "." + EnvironmentEdgeManager.currentTime() + ".jar");
+      // for each file that match the pattern
+      for (Path path : FileUtil.stat2Paths(fileStatuses)) {
+        if (fs.isFile(path)) {
+          // only process files, skip for directories
+          File dst = new File(parentDirStr,
+            "." + pathPrefix + "." + path.getName() + "." + EnvironmentEdgeManager.currentTime()
+              + ".jar");
           fs.copyToLocalFile(path, new Path(dst.toString()));
           dst.deleteOnExit();
 
@@ -182,7 +187,8 @@ public class CoprocessorClassLoader extends ClassLoaderBase {
 
           JarFile jarFile = new JarFile(dst.toString());
           try {
-            Enumeration<JarEntry> entries = jarFile.entries();  // get entries inside a jar file
+            // get entries inside a jar file
+            Enumeration<JarEntry> entries = jarFile.entries();
             while (entries.hasMoreElements()) {
               JarEntry entry = entries.nextElement();
               Matcher m = libJarPattern.matcher(entry.getName());
@@ -200,11 +206,12 @@ public class CoprocessorClassLoader extends ClassLoaderBase {
           } finally {
             jarFile.close();
           }
-
-          validFileEncountered = true;  // Set to true when encountering a file
+          // Set to true when encountering a file
+          validFileEncountered = true;
         }
       }
-      if (validFileEncountered == false) {  // all items returned by globStatus() are directories
+      if (validFileEncountered == false) {
+        // all items returned by globStatus() are directories
         throw new FileNotFoundException("No file found matching " + pathPattern1.toString());
       }
     }
diff --git a/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/DemoClient.java b/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/DemoClient.java
index 58075056f38..92c80a9036d 100644
--- a/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/DemoClient.java
+++ b/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/DemoClient.java
@@ -303,8 +303,8 @@ public class DemoClient {
       m = new Mutation();
       m.column = ByteBuffer.wrap(bytes("entry:sqr"));
       m.isDelete = true;
-      client.mutateRowTs(demoTable, ByteBuffer.wrap(row), mutations, 1,
-              dummyAttributes); // shouldn't override latest
+      // shouldn't override latest
+      client.mutateRowTs(demoTable, ByteBuffer.wrap(row), mutations, 1, dummyAttributes);
       printRow(client.getRow(demoTable, ByteBuffer.wrap(row), dummyAttributes));
 
       List<TCell> versions = client.getVer(demoTable, ByteBuffer.wrap(row),
diff --git a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionServerSource.java b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionServerSource.java
index 834a3387306..179e53f0199 100644
--- a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionServerSource.java
+++ b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionServerSource.java
@@ -483,8 +483,8 @@ public interface MetricsRegionServerSource extends BaseSource, JvmPauseMonitorSo
   String HEDGED_READ_WINS_DESC =
       "The number of times we started a hedged read and a hedged read won";
   String HEDGED_READ_IN_CUR_THREAD = "hedgedReadOpsInCurThread";
-  String HEDGED_READ_IN_CUR_THREAD_DESC =
-    "The number of times we execute a hedged read in current thread as a fallback for task rejection";
+  String HEDGED_READ_IN_CUR_THREAD_DESC = "The number of times we execute a hedged read"
+    + " in current thread as a fallback for task rejection";
 
   String TOTAL_BYTES_READ = "totalBytesRead";
   String TOTAL_BYTES_READ_DESC = "The total number of bytes read from HDFS";
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithACL.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithACL.java
index 90086732d59..927bf59ce34 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithACL.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithACL.java
@@ -86,21 +86,20 @@ public class IntegrationTestIngestWithACL extends IntegrationTestIngest {
     tmp.add(sb.toString());
     return tmp.toArray(new String[tmp.size()]);
   }
+
   @Override
   protected void addOptions() {
     super.addOptions();
-    super.addOptWithArg(OPT_SUPERUSER,
-        "Super user name used to add the ACL permissions");
+    super.addOptWithArg(OPT_SUPERUSER, "Super user name used to add the ACL permissions");
     super.addOptWithArg(OPT_USERS,
       "List of users to be added with the ACLs.  Should be comma seperated.");
-    super
-        .addOptWithArg(
-          OPT_AUTHN,
-          "The name of the properties file that contains kerberos key tab file and principal definitions. " +
-          "The principal key in the file should be of the form hbase.<username>.kerberos.principal." +
-          " The keytab key in the file should be of the form hbase.<username>.keytab.file. Example:  " +
-          "hbase.user1.kerberos.principal=user1/fully.qualified.domain.name@YOUR-REALM.COM, " +
-          "hbase.user1.keytab.file=<filelocation>.");
+    super.addOptWithArg(OPT_AUTHN,
+      "The name of the properties file that contains"
+        + " kerberos key tab file and principal definitions. The principal key in the file"
+        + " should be of the form hbase.<username>.kerberos.principal. The keytab key in the"
+        + " file should be of the form hbase.<username>.keytab.file. Example:"
+        + "  hbase.user1.kerberos.principal=user1/fully.qualified.domain.name@YOUR-REALM.COM,"
+        + " hbase.user1.keytab.file=<filelocation>.");
   }
 
   @Override
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/ServerAndDependenciesKillingMonkeyFactory.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/ServerAndDependenciesKillingMonkeyFactory.java
index 676b04d3fee..cc8dfcfcad3 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/ServerAndDependenciesKillingMonkeyFactory.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/ServerAndDependenciesKillingMonkeyFactory.java
@@ -49,11 +49,12 @@ public class ServerAndDependenciesKillingMonkeyFactory extends MonkeyFactory {
     loadProperties();
 
     // Destructive actions to mess things around. Cannot run batch restart.
-    Action[] actions1 = new Action[]{
+    // @formatter:off
+    Action[] actions1 = new Action[] {
       new RestartRandomRsExceptMetaAction(60000),
       new RestartActiveMasterAction(5000),
-      new RollingBatchRestartRsAction(5000, 1.0f, 2,
-        true), // only allow 2 servers to be dead.
+      // only allow 2 servers to be dead.
+      new RollingBatchRestartRsAction(5000, 1.0f, 2, true),
       new ForceBalancerAction(),
       new RestartRandomDataNodeAction(60000),
       new RestartRandomZKNodeAction(60000),
@@ -61,6 +62,7 @@ public class ServerAndDependenciesKillingMonkeyFactory extends MonkeyFactory {
       new RollingBatchSuspendResumeRsAction(rollingBatchSuspendRSSleepTime,
           rollingBatchSuspendtRSRatio)
     };
+    // @formatter:on
 
     // Action to log more info for debugging
     Action[] actions2 = new Action[]{
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/ServerKillingMonkeyFactory.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/ServerKillingMonkeyFactory.java
index 6a5efeda845..cfbab8eb955 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/ServerKillingMonkeyFactory.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/ServerKillingMonkeyFactory.java
@@ -47,16 +47,18 @@ public class ServerKillingMonkeyFactory extends MonkeyFactory {
     loadProperties();
 
     // Destructive actions to mess things around. Cannot run batch restart
+    // @formatter:off
     Action[] actions1 = new Action[] {
-        new RestartRandomRsExceptMetaAction(60000),
-        new RestartActiveMasterAction(5000),
-      new RollingBatchRestartRsAction(5000, 1.0f, 2,
-          true), //only allow 2 servers to be dead
+      new RestartRandomRsExceptMetaAction(60000),
+      new RestartActiveMasterAction(5000),
+      // only allow 2 servers to be dead
+      new RollingBatchRestartRsAction(5000, 1.0f, 2, true),
       new ForceBalancerAction(),
       new GracefulRollingRestartRsAction(gracefulRollingRestartTSSLeepTime),
       new RollingBatchSuspendResumeRsAction(rollingBatchSuspendRSSleepTime,
           rollingBatchSuspendtRSRatio)
     };
+    // @formatter:on
 
     // Action to log more info for debugging
     Action[] actions2 = new Action[] {
diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapred/TableInputFormatBase.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapred/TableInputFormatBase.java
index 509972e92aa..cef8e8a8176 100644
--- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapred/TableInputFormatBase.java
+++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapred/TableInputFormatBase.java
@@ -42,7 +42,7 @@ import org.apache.hadoop.mapred.Reporter;
  * A Base for {@link TableInputFormat}s. Receives a {@link Table}, a
  * byte[] of input columns and optionally a {@link Filter}.
  * Subclasses may use other TableRecordReader implementations.
- *
+ * <p/>
  * Subclasses MUST ensure initializeTable(Connection, TableName) is called for an instance to
  * function properly. Each of the entry points to this class used by the MapReduce framework,
  * {@link #getRecordReader(InputSplit, JobConf, Reporter)} and {@link #getSplits(JobConf, int)},
@@ -98,8 +98,7 @@ implements InputFormat<ImmutableBytesWritable, Result> {
    * Builds a TableRecordReader. If no TableRecordReader was provided, uses
    * the default.
    *
-   * @see org.apache.hadoop.mapred.InputFormat#getRecordReader(InputSplit,
-   *      JobConf, Reporter)
+   * @see InputFormat#getRecordReader(InputSplit, JobConf, Reporter)
    */
   public RecordReader<ImmutableBytesWritable, Result> getRecordReader(
       InputSplit split, JobConf job, Reporter reporter)
@@ -165,7 +164,7 @@ implements InputFormat<ImmutableBytesWritable, Result> {
 
   /**
    * Calculates the splits that will serve as input for the map tasks.
-   *
+   * <p/>
    * Splits are created in number equal to the smallest between numSplits and
    * the number of {@link org.apache.hadoop.hbase.regionserver.HRegion}s in the table.
    * If the number of splits is smaller than the number of
@@ -180,7 +179,7 @@ implements InputFormat<ImmutableBytesWritable, Result> {
    *
    * @return the input splits
    *
-   * @see org.apache.hadoop.mapred.InputFormat#getSplits(org.apache.hadoop.mapred.JobConf, int)
+   * @see InputFormat#getSplits(org.apache.hadoop.mapred.JobConf, int)
    */
   public InputSplit[] getSplits(JobConf job, int numSplits) throws IOException {
     if (this.table == null) {
@@ -282,7 +281,7 @@ implements InputFormat<ImmutableBytesWritable, Result> {
    * will call {@link #initialize(JobConf)} as a convenient centralized location to handle
    * retrieving the necessary configuration information and calling
    * {@link #initializeTable(Connection, TableName)}.
-   *
+   * <p/>
    * Subclasses should implement their initialize call such that it is safe to call multiple times.
    * The current TableInputFormatBase implementation relies on a non-null table reference to decide
    * if an initialize call is needed, but this behavior may change in the future. In particular,
diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatBase.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatBase.java
index d8205c1a0ca..ec789f2bdcd 100644
--- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatBase.java
+++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatBase.java
@@ -72,9 +72,7 @@ public abstract class MultiTableInputFormatBase extends
    * @return The newly created record reader.
    * @throws IOException When creating the reader fails.
    * @throws InterruptedException when record reader initialization fails
-   * @see org.apache.hadoop.mapreduce.InputFormat#createRecordReader(
-   *      org.apache.hadoop.mapreduce.InputSplit,
-   *      org.apache.hadoop.mapreduce.TaskAttemptContext)
+   * @see InputFormat#createRecordReader(InputSplit, TaskAttemptContext)
    */
   @Override
   public RecordReader<ImmutableBytesWritable, Result> createRecordReader(
@@ -152,7 +150,7 @@ public abstract class MultiTableInputFormatBase extends
    * @param context The current job context.
    * @return The list of input splits.
    * @throws IOException When creating the list of splits fails.
-   * @see org.apache.hadoop.mapreduce.InputFormat#getSplits(org.apache.hadoop.mapreduce.JobContext)
+   * @see InputFormat#getSplits(org.apache.hadoop.mapreduce.JobContext)
    */
   @Override
   public List<InputSplit> getSplits(JobContext context) throws IOException {
diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableSnapshotInputFormat.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableSnapshotInputFormat.java
index fa712903040..07c0820c1ae 100644
--- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableSnapshotInputFormat.java
+++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableSnapshotInputFormat.java
@@ -49,8 +49,7 @@ import java.util.Map;
  * included in each snapshot/scan
  * pair.
  * {@link TableMapReduceUtil#initMultiTableSnapshotMapperJob
- * (java.util.Map, Class, Class, Class, org.apache.hadoop.mapreduce.Job, boolean, org.apache
- * .hadoop.fs.Path)}
+ * (Map, Class, Class, Class, org.apache.hadoop.mapreduce.Job, boolean, Path)}
  * can be used to configure the job.
  * <pre>{@code
  * Job job = new Job(conf);
diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableSnapshotInputFormatImpl.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableSnapshotInputFormatImpl.java
index ef5c161d4cd..76254c3f6ba 100644
--- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableSnapshotInputFormatImpl.java
+++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableSnapshotInputFormatImpl.java
@@ -78,8 +78,7 @@ public class MultiTableSnapshotInputFormatImpl {
 
   /**
    * Return the list of splits extracted from the scans/snapshots pushed to conf by
-   * {@link
-   * #setInput(org.apache.hadoop.conf.Configuration, java.util.Map, org.apache.hadoop.fs.Path)}
+   * {@link #setInput(Configuration, Map, Path)}
    *
    * @param conf Configuration to determine splits from
    * @return Return the list of splits extracted from the scans/snapshots pushed to conf
@@ -115,7 +114,7 @@ public class MultiTableSnapshotInputFormatImpl {
 
   /**
    * Retrieve the snapshot name -&gt; list&lt;scan&gt; mapping pushed to configuration by
-   * {@link #setSnapshotToScans(org.apache.hadoop.conf.Configuration, java.util.Map)}
+   * {@link #setSnapshotToScans(Configuration, Map)}
    *
    * @param conf Configuration to extract name -&gt; list&lt;scan&gt; mappings from.
    * @return the snapshot name -&gt; list&lt;scan&gt; mapping pushed to configuration
diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultithreadedTableMapper.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultithreadedTableMapper.java
index ca82e2a58ee..4ad1935f37a 100644
--- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultithreadedTableMapper.java
+++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultithreadedTableMapper.java
@@ -231,8 +231,9 @@ public class MultithreadedTableMapper<K2, V2> extends TableMapper<K2, V2> {
     }
   }
 
-  @edu.umd.cs.findbugs.annotations.SuppressWarnings(value="REC_CATCH_EXCEPTION",
-      justification="Don't understand why FB is complaining about this one. We do throw exception")
+  @edu.umd.cs.findbugs.annotations.SuppressWarnings(value = "REC_CATCH_EXCEPTION",
+      justification = "Don't understand why FB is complaining about this one."
+        + " We do throw exception")
   private class MapRunner implements Runnable {
     private Mapper<ImmutableBytesWritable, Result, K2,V2> mapper;
     private Context subcontext;
diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mob/mapreduce/MobRefReporter.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mob/mapreduce/MobRefReporter.java
index 4152182a6d3..3b9a13879e8 100644
--- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mob/mapreduce/MobRefReporter.java
+++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mob/mapreduce/MobRefReporter.java
@@ -65,10 +65,10 @@ import org.slf4j.LoggerFactory;
 
 
 /**
- * Scans a given table + CF for all mob reference cells to get the list of backing mob files.
- * For each referenced file we attempt to verify that said file is on the FileSystem in a place
- * that the MOB system will look when attempting to resolve the actual value.
- *
+ * Scans a given table + CF for all mob reference cells to get the list of backing mob files. For
+ * each referenced file we attempt to verify that said file is on the FileSystem in a place that the
+ * MOB system will look when attempting to resolve the actual value.
+ * <p/>
  * The job includes counters that can help provide a rough sketch of the mob data.
  *
  * <pre>
@@ -94,31 +94,31 @@ import org.slf4j.LoggerFactory;
  *         Number of rows with total size in the 100,000s of bytes=6838
  *         Number of rows with total size in the 1,000,000s of bytes=3162
  * </pre>
- *
- *   * Map-Reduce Framework:Map input records - the number of rows with mob references
- *   * Map-Reduce Framework:Reduce output records - the number of unique hfiles referenced
- *   * MOB:NUM_CELLS - the total number of mob reference cells
- *   * PROBLEM:Affected rows - the number of rows that reference hfiles with an issue
- *   * PROBLEM:Problem MOB files - the number of unique hfiles that have an issue
- *   * CELLS PER ROW: - this counter group gives a histogram of the order of magnitude of the
- *         number of cells in a given row by grouping by the number of digits used in each count.
- *         This allows us to see more about the distribution of cells than what we can determine
- *         with just the cell count and the row count. In this particular example we can see that
- *         all of our rows have somewhere between 1 - 9 cells.
- *   * ROWS WITH PROBLEMS PER FILE: - this counter group gives a histogram of the order of
- *         magnitude of the number of rows in each of the hfiles with a problem. e.g. in the
- *         example there are 2 hfiles and they each have the same order of magnitude number of rows,
- *         specifically between 100 and 999.
- *   * SIZES OF CELLS: - this counter group gives a histogram of the order of magnitude of
- *         the size of mob values according to our reference cells. e.g. in the example above we
- *         have cell sizes that are all between 10,000 bytes and 9,999,999 bytes. From this
- *         histogram we can also see that _most_ cells are 100,000 - 999,000 bytes and the smaller
- *         and bigger ones are outliers making up less than 2% of mob cells.
- *   * SIZES OF ROWS: - this counter group gives a histogram of the order of magnitude of the
- *         size of mob values across each row according to our reference cells. In the example above
- *         we have rows that are are between 100,000 bytes and 9,999,999 bytes. We can also see that
- *         about 2/3rd of our rows are 100,000 - 999,999 bytes.
- *
+ * <ol>
+ * <li>Map-Reduce Framework:Map input records - the number of rows with mob references</li>
+ * <li>Map-Reduce Framework:Reduce output records - the number of unique hfiles referenced</li>
+ * <li>MOB:NUM_CELLS - the total number of mob reference cells</li>
+ * <li>PROBLEM:Affected rows - the number of rows that reference hfiles with an issue</li>
+ * <li>PROBLEM:Problem MOB files - the number of unique hfiles that have an issue</li>
+ * <li>CELLS PER ROW: - this counter group gives a histogram of the order of magnitude of the number
+ * of cells in a given row by grouping by the number of digits used in each count. This allows us to
+ * see more about the distribution of cells than what we can determine with just the cell count and
+ * the row count. In this particular example we can see that all of our rows have somewhere between
+ * 1 - 9 cells.</li>
+ * <li>ROWS WITH PROBLEMS PER FILE: - this counter group gives a histogram of the order of magnitude
+ * of the number of rows in each of the hfiles with a problem. e.g. in the example there are 2
+ * hfiles and they each have the same order of magnitude number of rows, specifically between 100
+ * and 999.</li>
+ * <li>SIZES OF CELLS: - this counter group gives a histogram of the order of magnitude of the size
+ * of mob values according to our reference cells. e.g. in the example above we have cell sizes that
+ * are all between 10,000 bytes and 9,999,999 bytes. From this histogram we can also see that _most_
+ * cells are 100,000 - 999,000 bytes and the smaller and bigger ones are outliers making up less
+ * than 2% of mob cells.</li>
+ * <li>SIZES OF ROWS: - this counter group gives a histogram of the order of magnitude of the size
+ * of mob values across each row according to our reference cells. In the example above we have rows
+ * that are are between 100,000 bytes and 9,999,999 bytes. We can also see that about 2/3rd of our
+ * rows are 100,000 - 999,999 bytes.</li>
+ * </ol>
  * Generates a report that gives one file status per line, with tabs dividing fields.
  *
  * <pre>
@@ -133,32 +133,31 @@ import org.slf4j.LoggerFactory;
  * </pre>
  *
  * Possible results are listed; the first three indicate things are working properly.
- *   * MOB DIR - the reference is in the normal MOB area for the given table and CF
- *   * HLINK TO ARCHIVE FOR SAME TABLE - the reference is present in the archive area for this
- *         table and CF
- *   * HLINK TO ARCHIVE FOR OTHER TABLE - the reference is present in a different table and CF,
- *         either in the MOB or archive areas (e.g. from a snapshot restore or clone)
- *   * ARCHIVE WITH HLINK BUT NOT FROM OUR TABLE - the reference is currently present in the archive
- *         area for this table and CF, but it is kept there because a _different_ table has a
- *         reference to it (e.g. from a snapshot clone). If these other tables are removed then
- *         the file will likely be deleted unless there is a snapshot also referencing it.
- *   * ARCHIVE BUT NO HLINKS - the reference is currently present in the archive for this table and
- *         CF, but there are no references present to prevent its removal. Unless it is newer than
- *         the general TTL (default 5 minutes) or referenced in a snapshot it will be subject to
- *         cleaning.
- *   * ARCHIVE BUT FAILURE WHILE CHECKING HLINKS - Check the job logs to see why things failed while
- *         looking for why this file is being kept around.
- *   * MISSING FILE - We couldn't find the reference on the FileSystem. Either there is dataloss due
- *         to a bug in the MOB storage system or the MOB storage is damaged but in an edge case that
- *         allows it to work for now. You can verify which by doing a raw reference scan to get the
- *         referenced hfile and check the underlying filesystem. See the ref guide section on mob
- *         for details.
- *   * HLINK BUT POINT TO MISSING FILE - There is a pointer in our mob area for this table and CF
- *         to a file elsewhere on the FileSystem, however the file it points to no longer exists.
- *   * MISSING FILE BUT FAILURE WHILE CHECKING HLINKS - We could not find the referenced file,
- *         however you should check the job logs to see why we couldn't check to see if there is a
- *         pointer to the referenced file in our archive or another table's archive or mob area.
- *
+ * <ol>
+ * <li>MOB DIR - the reference is in the normal MOB area for the given table and CF</li>
+ * <li>HLINK TO ARCHIVE FOR SAME TABLE - the reference is present in the archive area for this table
+ * and CF</li>
+ * <li>HLINK TO ARCHIVE FOR OTHER TABLE - the reference is present in a different table and CF,
+ * either in the MOB or archive areas (e.g. from a snapshot restore or clone)</li>
+ * <li>ARCHIVE WITH HLINK BUT NOT FROM OUR TABLE - the reference is currently present in the archive
+ * area for this table and CF, but it is kept there because a _different_ table has a reference to
+ * it (e.g. from a snapshot clone). If these other tables are removed then the file will likely be
+ * deleted unless there is a snapshot also referencing it.</li>
+ * <li>ARCHIVE BUT NO HLINKS - the reference is currently present in the archive for this table and
+ * CF, but there are no references present to prevent its removal. Unless it is newer than the
+ * general TTL (default 5 minutes) or referenced in a snapshot it will be subject to cleaning.</li>
+ * <li>ARCHIVE BUT FAILURE WHILE CHECKING HLINKS - Check the job logs to see why things failed while
+ * looking for why this file is being kept around.</li>
+ * <li>MISSING FILE - We couldn't find the reference on the FileSystem. Either there is dataloss due
+ * to a bug in the MOB storage system or the MOB storage is damaged but in an edge case that allows
+ * it to work for now. You can verify which by doing a raw reference scan to get the referenced
+ * hfile and check the underlying filesystem. See the ref guide section on mob for details.</li>
+ * <li>HLINK BUT POINT TO MISSING FILE - There is a pointer in our mob area for this table and CF to
+ * a file elsewhere on the FileSystem, however the file it points to no longer exists.</li>
+ * <li>MISSING FILE BUT FAILURE WHILE CHECKING HLINKS - We could not find the referenced file,
+ * however you should check the job logs to see why we couldn't check to see if there is a pointer
+ * to the referenced file in our archive or another table's archive or mob area.</li>
+ * </ol>
  */
 @InterfaceAudience.Private
 public class MobRefReporter extends Configured implements Tool {
diff --git a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java
index 4e56e4face1..fd6d112f10c 100644
--- a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java
+++ b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java
@@ -298,15 +298,16 @@ public class TestImportExport {
               IMPORT_TABLE, FQ_OUTPUT_DIR
       };
       assertTrue(runImport(args));
-      /* exportedTableIn94Format contains 5 rows
-      ROW         COLUMN+CELL
-      r1          column=f1:c1, timestamp=1383766761171, value=val1
-      r2          column=f1:c1, timestamp=1383766771642, value=val2
-      r3          column=f1:c1, timestamp=1383766777615, value=val3
-      r4          column=f1:c1, timestamp=1383766785146, value=val4
-      r5          column=f1:c1, timestamp=1383766791506, value=val5
-      */
-     assertEquals(5, UTIL.countRows(t));
+      // @formatter:off
+      // exportedTableIn94Format contains 5 rows
+      // ROW         COLUMN+CELL
+      // r1          column=f1:c1, timestamp=1383766761171, value=val1
+      // r2          column=f1:c1, timestamp=1383766771642, value=val2
+      // r3          column=f1:c1, timestamp=1383766777615, value=val3
+      // r4          column=f1:c1, timestamp=1383766785146, value=val4
+      // r5          column=f1:c1, timestamp=1383766791506, value=val5
+      // @formatter:on
+      assertEquals(5, UTIL.countRows(t));
     }
   }
 
@@ -331,12 +332,9 @@ public class TestImportExport {
       p.addColumn(FAMILYA, QUAL, now + 3, QUAL);
       p.addColumn(FAMILYA, QUAL, now + 4, QUAL);
       t.put(p);
-
-      String[] args = new String[] {
-          "-D" + ExportUtils.EXPORT_BATCHING + "=" + EXPORT_BATCH_SIZE,  // added scanner batching arg.
-          name.getMethodName(),
-          FQ_OUTPUT_DIR
-      };
+      // added scanner batching arg.
+      String[] args = new String[] { "-D" + ExportUtils.EXPORT_BATCHING + "=" + EXPORT_BATCH_SIZE,
+        name.getMethodName(), FQ_OUTPUT_DIR };
       assertTrue(runExport(args));
 
       FileSystem fs = FileSystem.get(UTIL.getConfiguration());
diff --git a/hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/model/TestRowModel.java b/hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/model/TestRowModel.java
index 99f8e3df447..808b77bc9d6 100644
--- a/hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/model/TestRowModel.java
+++ b/hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/model/TestRowModel.java
@@ -48,8 +48,9 @@ public class TestRowModel extends TestModelBase<RowModel> {
   public TestRowModel() throws Exception {
     super(RowModel.class);
     AS_XML =
-      "<?xml version=\"1.0\" encoding=\"UTF-8\" standalone=\"yes\"?><Row key=\"dGVzdHJvdzE=\">" +
-      "<Cell column=\"dGVzdGNvbHVtbjE=\" timestamp=\"1245219839331\">dGVzdHZhbHVlMQ==</Cell></Row>";
+      "<?xml version=\"1.0\" encoding=\"UTF-8\" standalone=\"yes\"?>" + "<Row key=\"dGVzdHJvdzE=\">"
+        + "<Cell column=\"dGVzdGNvbHVtbjE=\" timestamp=\"1245219839331\">dGVzdHZhbHVlMQ==</Cell>"
+        + "</Row>";
 
     AS_JSON =
       "{\"key\":\"dGVzdHJvdzE=\",\"Cell\":[{\"column\":\"dGVzdGNvbHVtbjE=\"," +
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/errorhandling/ForeignException.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/errorhandling/ForeignException.java
index 85abc722044..142827be70c 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/errorhandling/ForeignException.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/errorhandling/ForeignException.java
@@ -157,7 +157,6 @@ public class ForeignException extends IOException {
    * @param bytes
    * @return the ForeignExcpetion instance
    * @throws InvalidProtocolBufferException if there was deserialization problem this is thrown.
-   * @throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException 
    */
   public static ForeignException deserialize(byte[] bytes)
   throws IOException {
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java
index 2e6c19edfca..0cb100eefb5 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java
@@ -534,13 +534,13 @@ public final class HFile {
   /**
    * @param fs filesystem
    * @param path Path to file to read
-   * @param cacheConf This must not be null. @see
-   *          {@link org.apache.hadoop.hbase.io.hfile.CacheConfig#CacheConfig(Configuration)}
+   * @param cacheConf This must not be null.
    * @param primaryReplicaReader true if this is a reader for primary replica
    * @param conf Configuration
    * @return an active Reader instance
    * @throws IOException Will throw a CorruptHFileException (DoNotRetryIOException subtype) if hfile
    *           is corrupt/invalid.
+   * @see CacheConfig#CacheConfig(Configuration)
    */
   public static Reader createReader(FileSystem fs, Path path, CacheConfig cacheConf,
       boolean primaryReplicaReader, Configuration conf) throws IOException {
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMasterCommandLine.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMasterCommandLine.java
index 83baf156ba9..aeba930cdb8 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMasterCommandLine.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMasterCommandLine.java
@@ -149,11 +149,10 @@ public class HMasterCommandLine extends ServerCommandLine {
       if (shutDownCluster) {
         return stopMaster();
       }
-      System.err.println(
-        "To shutdown the master run " +
-        "hbase-daemon.sh stop master or send a kill signal to " +
-        "the HMaster pid, " +
-        "and to stop HBase Cluster run \"stop-hbase.sh\" or \"hbase master stop --shutDownCluster\"");
+      System.err.println("To shutdown the master run "
+        + "hbase-daemon.sh stop master or send a kill signal to the HMaster pid, "
+        + "and to stop HBase Cluster run \"stop-hbase.sh\" or \"hbase master "
+        + "stop --shutDownCluster\"");
       return 1;
     } else if ("clear".equals(command)) {
       return (ZNodeClearer.clear(getConf()) ? 0 : 1);
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterWalManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterWalManager.java
index c63dec308a9..8093511606a 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterWalManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterWalManager.java
@@ -293,9 +293,9 @@ public class MasterWalManager {
     splitLog(serverNames, META_FILTER);
   }
 
-  @edu.umd.cs.findbugs.annotations.SuppressWarnings(value="UL_UNRELEASED_LOCK", justification=
-      "We only release this lock when we set it. Updates to code that uses it should verify use " +
-      "of the guard boolean.")
+  @edu.umd.cs.findbugs.annotations.SuppressWarnings(value = "UL_UNRELEASED_LOCK",
+      justification = "We only release this lock when we set it. Updates to code "
+        + "that uses it should verify use of the guard boolean.")
   List<Path> getLogDirs(final Set<ServerName> serverNames) throws IOException {
     List<Path> logDirs = new ArrayList<>();
     boolean needReleaseLock = false;
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java
index 186a8ff11bb..16185cb7949 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java
@@ -16,6 +16,7 @@
  * limitations under the License.
  */
 package org.apache.hadoop.hbase.master;
+
 import static org.apache.hadoop.hbase.master.SplitLogManager.ResubmitDirective.CHECK;
 import static org.apache.hadoop.hbase.master.SplitLogManager.ResubmitDirective.FORCE;
 import static org.apache.hadoop.hbase.master.SplitLogManager.TerminationStatus.DELETED;
@@ -155,10 +156,10 @@ public class SplitLogManager {
   /**
    * Get a list of paths that need to be split given a set of server-specific directories and
    * optionally  a filter.
-   *
+   * <p/>
    * See {@link AbstractFSWALProvider#getServerNameFromWALDirectoryName} for more info on directory
    * layout.
-   *
+   * <p/>
    * Should be package-private, but is needed by
    * {@link org.apache.hadoop.hbase.wal.WALSplitter#split(Path, Path, Path, FileSystem,
    *     Configuration, org.apache.hadoop.hbase.wal.WALFactory)} for tests.
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java
index 4b6a8ba139a..5aa09090c44 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java
@@ -274,11 +274,11 @@ class SimpleRegionNormalizer implements RegionNormalizer, ConfigurationObserver
   }
 
   /**
-   * @param tableRegions regions of table to normalize
+   * Also make sure tableRegions contains regions of the same table
+   * @param tableRegions    regions of table to normalize
    * @param tableDescriptor the TableDescriptor
    * @return average region size depending on
-   * @see org.apache.hadoop.hbase.client.TableDescriptor#getNormalizerTargetRegionCount()
-   * Also make sure tableRegions contains regions of the same table
+   * @see TableDescriptor#getNormalizerTargetRegionCount()
    */
   private double getAverageRegionSizeMb(final List<RegionInfo> tableRegions,
     final TableDescriptor tableDescriptor) {
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/CloneSnapshotProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/CloneSnapshotProcedure.java
index f6185d15603..b75feaf0984 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/CloneSnapshotProcedure.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/CloneSnapshotProcedure.java
@@ -64,8 +64,9 @@ import org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
 
 import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
 import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos;
 import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CloneSnapshotState;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CloneSnapshotStateData;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.RestoreParentToChildRegionsPair;
 import org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription;
 
 @InterfaceAudience.Private
@@ -310,11 +311,10 @@ public class CloneSnapshotProcedure
       throws IOException {
     super.serializeStateData(serializer);
 
-    MasterProcedureProtos.CloneSnapshotStateData.Builder cloneSnapshotMsg =
-      MasterProcedureProtos.CloneSnapshotStateData.newBuilder()
-        .setUserInfo(MasterProcedureUtil.toProtoUserInfo(getUser()))
-        .setSnapshot(this.snapshot)
-        .setTableSchema(ProtobufUtil.toTableSchema(tableDescriptor));
+    CloneSnapshotStateData.Builder cloneSnapshotMsg = CloneSnapshotStateData.newBuilder()
+      .setUserInfo(MasterProcedureUtil.toProtoUserInfo(getUser()))
+      .setSnapshot(this.snapshot)
+      .setTableSchema(ProtobufUtil.toTableSchema(tableDescriptor));
 
     cloneSnapshotMsg.setRestoreAcl(restoreAcl);
     if (newRegions != null) {
@@ -328,11 +328,11 @@ public class CloneSnapshotProcedure
       while (it.hasNext()) {
         final Map.Entry<String, Pair<String, String>> entry = it.next();
 
-        MasterProcedureProtos.RestoreParentToChildRegionsPair.Builder parentToChildrenPair =
-          MasterProcedureProtos.RestoreParentToChildRegionsPair.newBuilder()
-          .setParentRegionName(entry.getKey())
-          .setChild1RegionName(entry.getValue().getFirst())
-          .setChild2RegionName(entry.getValue().getSecond());
+        RestoreParentToChildRegionsPair.Builder parentToChildrenPair =
+          RestoreParentToChildRegionsPair.newBuilder()
+            .setParentRegionName(entry.getKey())
+            .setChild1RegionName(entry.getValue().getFirst())
+            .setChild2RegionName(entry.getValue().getSecond());
         cloneSnapshotMsg.addParentToChildRegionsPairList(parentToChildrenPair);
       }
     }
@@ -347,8 +347,7 @@ public class CloneSnapshotProcedure
       throws IOException {
     super.deserializeStateData(serializer);
 
-    MasterProcedureProtos.CloneSnapshotStateData cloneSnapshotMsg =
-        serializer.deserialize(MasterProcedureProtos.CloneSnapshotStateData.class);
+    CloneSnapshotStateData cloneSnapshotMsg = serializer.deserialize(CloneSnapshotStateData.class);
     setUser(MasterProcedureUtil.toUserInfo(cloneSnapshotMsg.getUserInfo()));
     snapshot = cloneSnapshotMsg.getSnapshot();
     tableDescriptor = ProtobufUtil.toTableDescriptor(cloneSnapshotMsg.getTableSchema());
@@ -365,8 +364,8 @@ public class CloneSnapshotProcedure
     }
     if (cloneSnapshotMsg.getParentToChildRegionsPairListCount() > 0) {
       parentsToChildrenPairMap = new HashMap<>();
-      for (MasterProcedureProtos.RestoreParentToChildRegionsPair parentToChildrenPair:
-        cloneSnapshotMsg.getParentToChildRegionsPairListList()) {
+      for (RestoreParentToChildRegionsPair parentToChildrenPair : cloneSnapshotMsg
+        .getParentToChildRegionsPairListList()) {
         parentsToChildrenPairMap.put(
           parentToChildrenPair.getParentRegionName(),
           new Pair<>(
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.java
index 6b28173d4e4..4eaacee4fe2 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.java
@@ -58,8 +58,9 @@ import org.slf4j.LoggerFactory;
 
 import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
 import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.RestoreParentToChildRegionsPair;
 import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.RestoreSnapshotState;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.RestoreSnapshotStateData;
 import org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription;
 
 @InterfaceAudience.Private
@@ -239,11 +240,10 @@ public class RestoreSnapshotProcedure
       throws IOException {
     super.serializeStateData(serializer);
 
-    MasterProcedureProtos.RestoreSnapshotStateData.Builder restoreSnapshotMsg =
-      MasterProcedureProtos.RestoreSnapshotStateData.newBuilder()
-        .setUserInfo(MasterProcedureUtil.toProtoUserInfo(getUser()))
-        .setSnapshot(this.snapshot)
-        .setModifiedTableSchema(ProtobufUtil.toTableSchema(modifiedTableDescriptor));
+    RestoreSnapshotStateData.Builder restoreSnapshotMsg = RestoreSnapshotStateData.newBuilder()
+      .setUserInfo(MasterProcedureUtil.toProtoUserInfo(getUser()))
+      .setSnapshot(this.snapshot)
+      .setModifiedTableSchema(ProtobufUtil.toTableSchema(modifiedTableDescriptor));
 
     if (regionsToRestore != null) {
       for (RegionInfo hri: regionsToRestore) {
@@ -266,11 +266,11 @@ public class RestoreSnapshotProcedure
       while (it.hasNext()) {
         final Map.Entry<String, Pair<String, String>> entry = it.next();
 
-        MasterProcedureProtos.RestoreParentToChildRegionsPair.Builder parentToChildrenPair =
-          MasterProcedureProtos.RestoreParentToChildRegionsPair.newBuilder()
-          .setParentRegionName(entry.getKey())
-          .setChild1RegionName(entry.getValue().getFirst())
-          .setChild2RegionName(entry.getValue().getSecond());
+        RestoreParentToChildRegionsPair.Builder parentToChildrenPair =
+          RestoreParentToChildRegionsPair.newBuilder()
+            .setParentRegionName(entry.getKey())
+            .setChild1RegionName(entry.getValue().getFirst())
+            .setChild2RegionName(entry.getValue().getSecond());
         restoreSnapshotMsg.addParentToChildRegionsPairList (parentToChildrenPair);
       }
     }
@@ -283,8 +283,8 @@ public class RestoreSnapshotProcedure
       throws IOException {
     super.deserializeStateData(serializer);
 
-    MasterProcedureProtos.RestoreSnapshotStateData restoreSnapshotMsg =
-        serializer.deserialize(MasterProcedureProtos.RestoreSnapshotStateData.class);
+    RestoreSnapshotStateData restoreSnapshotMsg =
+        serializer.deserialize(RestoreSnapshotStateData.class);
     setUser(MasterProcedureUtil.toUserInfo(restoreSnapshotMsg.getUserInfo()));
     snapshot = restoreSnapshotMsg.getSnapshot();
     modifiedTableDescriptor =
@@ -315,8 +315,8 @@ public class RestoreSnapshotProcedure
       }
     }
     if (restoreSnapshotMsg.getParentToChildRegionsPairListCount() > 0) {
-      for (MasterProcedureProtos.RestoreParentToChildRegionsPair parentToChildrenPair:
-        restoreSnapshotMsg.getParentToChildRegionsPairListList()) {
+      for (RestoreParentToChildRegionsPair parentToChildrenPair : restoreSnapshotMsg
+        .getParentToChildRegionsPairListList()) {
         parentsToChildrenPairMap.put(
           parentToChildrenPair.getParentRegionName(),
           new Pair<>(
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/namequeues/impl/BalancerRejectionQueueService.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/namequeues/impl/BalancerRejectionQueueService.java
index 6da708381af..9e922ee4393 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/namequeues/impl/BalancerRejectionQueueService.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/namequeues/impl/BalancerRejectionQueueService.java
@@ -79,8 +79,8 @@ public class BalancerRejectionQueueService implements NamedQueueService {
       return;
     }
     if (!(namedQueuePayload instanceof BalancerRejectionDetails)) {
-      LOG.warn(
-        "BalancerRejectionQueueService: NamedQueuePayload is not of type BalancerRejectionDetails.");
+      LOG.warn("BalancerRejectionQueueService: NamedQueuePayload is not of type"
+        + " BalancerRejectionDetails.");
       return;
     }
     BalancerRejectionDetails balancerRejectionDetails = (BalancerRejectionDetails) namedQueuePayload;
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/UserQuotaState.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/UserQuotaState.java
index b285e049bdf..e17f8996431 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/UserQuotaState.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/UserQuotaState.java
@@ -34,9 +34,10 @@ import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
  */
 @InterfaceAudience.Private
 @InterfaceStability.Evolving
-@edu.umd.cs.findbugs.annotations.SuppressWarnings(value="IS2_INCONSISTENT_SYNC",
-  justification="FindBugs seems confused; says bypassGlobals, namepaceLimiters, and " +
-    "tableLimiters are mostly synchronized...but to me it looks like they are totally synchronized")
+@edu.umd.cs.findbugs.annotations.SuppressWarnings(value = "IS2_INCONSISTENT_SYNC",
+    justification = "FindBugs seems confused; says bypassGlobals, namepaceLimiters, and "
+      + "tableLimiters are mostly synchronized..."
+      + "but to me it looks like they are totally synchronized")
 public class UserQuotaState extends QuotaState {
   private Map<String, QuotaLimiter> namespaceLimiters = null;
   private Map<TableName, QuotaLimiter> tableLimiters = null;
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMemStore.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMemStore.java
index 56dab21baf2..4ffb600dd09 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMemStore.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMemStore.java
@@ -55,9 +55,11 @@ public abstract class AbstractMemStore implements MemStore {
 
   protected RegionServicesForStores regionServices;
 
+  // @formatter:off
   public final static long FIXED_OVERHEAD = (long) ClassSize.OBJECT
-          + (5 * ClassSize.REFERENCE)
-          + (2 * Bytes.SIZEOF_LONG); // snapshotId, timeOfOldestEdit
+    + (5 * ClassSize.REFERENCE)
+    + (2 * Bytes.SIZEOF_LONG); // snapshotId, timeOfOldestEdit
+  // @formatter:on
 
   public final static long DEEP_OVERHEAD = FIXED_OVERHEAD;
 
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java
index abbf35f7c4c..c036da4fd38 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java
@@ -2091,8 +2091,9 @@ public class RSRpcServices implements HBaseRPCErrorHandler, AdminService.Blockin
         throw new ServiceException(ie);
       }
       // We are assigning meta, wait a little for regionserver to finish initialization.
-      int timeout = regionServer.getConfiguration().getInt(HConstants.HBASE_RPC_TIMEOUT_KEY,
-        HConstants.DEFAULT_HBASE_RPC_TIMEOUT) >> 2; // Quarter of RPC timeout
+      // Default to quarter of RPC timeout
+      int timeout = regionServer.getConfiguration()
+        .getInt(HConstants.HBASE_RPC_TIMEOUT_KEY, HConstants.DEFAULT_HBASE_RPC_TIMEOUT) >> 2;
       long endTime = EnvironmentEdgeManager.currentTime() + timeout;
       synchronized (regionServer.online) {
         try {
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.java
index a1d082e9701..5ac86970c9d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.java
@@ -200,10 +200,10 @@ public class ReplicationSource implements ReplicationSourceInterface {
     this.waitOnEndpointSeconds =
       this.conf.getInt(WAIT_ON_ENDPOINT_SECONDS, DEFAULT_WAIT_ON_ENDPOINT_SECONDS);
     decorateConf();
-    this.sleepForRetries =
-        this.conf.getLong("replication.source.sleepforretries", 1000);    // 1 second
-    this.maxRetriesMultiplier =
-        this.conf.getInt("replication.source.maxretriesmultiplier", 300); // 5 minutes @ 1 sec per
+    // 1 second
+    this.sleepForRetries = this.conf.getLong("replication.source.sleepforretries", 1000);
+    // 5 minutes @ 1 sec per
+    this.maxRetriesMultiplier = this.conf.getInt("replication.source.maxretriesmultiplier", 300);
     this.queueSizePerGroup = this.conf.getInt("hbase.regionserver.maxlogs", 32);
     this.logQueue = new ReplicationSourceLogQueue(conf, metrics, this);
     this.queueStorage = queueStorage;
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceShipper.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceShipper.java
index 67f9358ed00..c40cf7e00a4 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceShipper.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceShipper.java
@@ -81,14 +81,15 @@ public class ReplicationSourceShipper extends Thread {
     this.walGroupId = walGroupId;
     this.logQueue = logQueue;
     this.source = source;
-    this.sleepForRetries =
-        this.conf.getLong("replication.source.sleepforretries", 1000);    // 1 second
-    this.maxRetriesMultiplier =
-        this.conf.getInt("replication.source.maxretriesmultiplier", 300); // 5 minutes @ 1 sec per
+    // 1 second
+    this.sleepForRetries = this.conf.getLong("replication.source.sleepforretries", 1000);
+    // 5 minutes @ 1 sec per
+    this.maxRetriesMultiplier = this.conf.getInt("replication.source.maxretriesmultiplier", 300);
+    // 20 seconds
     this.getEntriesTimeout =
-        this.conf.getInt("replication.source.getEntries.timeout", DEFAULT_TIMEOUT); // 20 seconds
+      this.conf.getInt("replication.source.getEntries.timeout", DEFAULT_TIMEOUT);
     this.shipEditsTimeout = this.conf.getInt(HConstants.REPLICATION_SOURCE_SHIPEDITS_TIMEOUT,
-        HConstants.REPLICATION_SOURCE_SHIPEDITS_TIMEOUT_DFAULT);
+      HConstants.REPLICATION_SOURCE_SHIPEDITS_TIMEOUT_DFAULT);
   }
 
   @Override
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALReader.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALReader.java
index 11090448c7c..c61494e12c6 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALReader.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALReader.java
@@ -106,10 +106,10 @@ class ReplicationSourceWALReader extends Thread {
     int batchCount = conf.getInt("replication.source.nb.batches", 1);
     this.totalBufferUsed = source.getSourceManager().getTotalBufferUsed();
     this.totalBufferQuota = source.getSourceManager().getTotalBufferLimit();
-    this.sleepForRetries =
-        this.conf.getLong("replication.source.sleepforretries", 1000);    // 1 second
-    this.maxRetriesMultiplier =
-        this.conf.getInt("replication.source.maxretriesmultiplier", 300); // 5 minutes @ 1 sec per
+    // 1 second
+    this.sleepForRetries = this.conf.getLong("replication.source.sleepforretries", 1000);
+    // 5 minutes @ 1 sec per
+    this.maxRetriesMultiplier = this.conf.getInt("replication.source.maxretriesmultiplier", 300);
     this.eofAutoRecovery = conf.getBoolean("replication.source.eof.autorecovery", false);
     this.entryBatchQueue = new LinkedBlockingQueue<>(batchCount);
     this.walGroupId = walGroupId;
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java
index f540b536ee2..3f47a3cba5f 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java
@@ -72,7 +72,7 @@ public class TokenUtil {
 
 
   /**
-   * See {@link ClientTokenUtil#toToken(org.apache.hadoop.security.token.Token)}.
+   * See {@link ClientTokenUtil#toToken(Token)}.
    * @deprecated External users should not use this method. Please post on
    *   the HBase dev mailing list if you need this method. Internal
    *   HBase code should use {@link ClientTokenUtil} instead.
@@ -83,8 +83,7 @@ public class TokenUtil {
   }
 
   /**
-   * See {@link ClientTokenUtil#obtainToken(org.apache.hadoop.hbase.client.Connection,
-   * org.apache.hadoop.hbase.security.User)}.
+   * See {@link ClientTokenUtil#obtainToken(Connection, User)}.
    * @deprecated External users should not use this method. Please post on
    *   the HBase dev mailing list if you need this method. Internal
    *   HBase code should use {@link ClientTokenUtil} instead.
@@ -96,8 +95,7 @@ public class TokenUtil {
   }
 
   /**
-   * See {@link ClientTokenUtil#obtainAndCacheToken(org.apache.hadoop.hbase.client.Connection,
-   * org.apache.hadoop.hbase.security.User)}.
+   * See {@link ClientTokenUtil#obtainAndCacheToken(Connection, User)}.
    */
   public static void obtainAndCacheToken(final Connection conn,
       User user)
@@ -106,7 +104,7 @@ public class TokenUtil {
   }
 
   /**
-   * See {@link ClientTokenUtil#toToken(org.apache.hadoop.security.token.Token)}.
+   * See {@link ClientTokenUtil#toToken(org.apache.hadoop.hbase.shaded.protobuf.generated.AuthenticationProtos.Token)}.
    * @deprecated External users should not use this method. Please post on
    *   the HBase dev mailing list if you need this method. Internal
    *   HBase code should use {@link ClientTokenUtil} instead.
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/EntryBuffers.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/EntryBuffers.java
index 908d3e6db77..4c1e5be695c 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/EntryBuffers.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/EntryBuffers.java
@@ -167,8 +167,8 @@ public class EntryBuffers {
       internify(entry);
       entries.add(entry);
       // TODO linkedlist entry
-      long incrHeap = entry.getEdit().heapSize() +
-          ClassSize.align(2 * ClassSize.REFERENCE); // WALKey pointers
+      // entry size plus WALKey pointers
+      long incrHeap = entry.getEdit().heapSize() + ClassSize.align(2 * ClassSize.REFERENCE);
       heapInBuffer += incrHeap;
       return incrHeap;
     }
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java
index 7d056916326..c7c7d751586 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java
@@ -2924,9 +2924,8 @@ public class HBaseTestingUtility extends HBaseZKTestingUtility {
     if (jobConf == null) {
       jobConf = mrCluster.createJobConf();
     }
-
-    jobConf.set("mapreduce.cluster.local.dir",
-      conf.get("mapreduce.cluster.local.dir")); //Hadoop MiniMR overwrites this while it should not
+    // Hadoop MiniMR overwrites this while it should not
+    jobConf.set("mapreduce.cluster.local.dir", conf.get("mapreduce.cluster.local.dir"));
     LOG.info("Mini mapreduce cluster started");
 
     // In hadoop2, YARN/MR2 starts a mini cluster with its own conf instance and updates settings.
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverScannerOpenHook.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverScannerOpenHook.java
index 9ab021aeb9d..53a84366d1d 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverScannerOpenHook.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverScannerOpenHook.java
@@ -223,8 +223,9 @@ public class TestRegionObserverScannerOpenHook {
     Get get = new Get(ROW);
     Result r = region.get(get);
     assertNull(
-      "Got an unexpected number of rows - no data should be returned with the NoDataFromScan coprocessor. Found: "
-          + r, r.listCells());
+      "Got an unexpected number of rows - "
+        + "no data should be returned with the NoDataFromScan coprocessor. Found: " + r,
+      r.listCells());
     HBaseTestingUtility.closeRegionAndWAL(region);
   }
 
@@ -250,8 +251,9 @@ public class TestRegionObserverScannerOpenHook {
     Get get = new Get(ROW);
     Result r = region.get(get);
     assertNull(
-      "Got an unexpected number of rows - no data should be returned with the NoDataFromScan coprocessor. Found: "
-          + r, r.listCells());
+      "Got an unexpected number of rows - "
+        + "no data should be returned with the NoDataFromScan coprocessor. Found: " + r,
+      r.listCells());
     HBaseTestingUtility.closeRegionAndWAL(region);
   }
 
@@ -269,15 +271,19 @@ public class TestRegionObserverScannerOpenHook {
     }
 
     public CountDownLatch getCompactionStateChangeLatch() {
-      if (compactionStateChangeLatch == null) compactionStateChangeLatch = new CountDownLatch(1);
+      if (compactionStateChangeLatch == null) {
+        compactionStateChangeLatch = new CountDownLatch(1);
+      }
       return compactionStateChangeLatch;
     }
 
     @Override
     public boolean compact(CompactionContext compaction, HStore store,
-        ThroughputController throughputController) throws IOException {
+      ThroughputController throughputController) throws IOException {
       boolean ret = super.compact(compaction, store, throughputController);
-      if (ret) compactionStateChangeLatch.countDown();
+      if (ret) {
+        compactionStateChangeLatch.countDown();
+      }
       return ret;
     }
 
@@ -341,14 +347,16 @@ public class TestRegionObserverScannerOpenHook {
     Get get = new Get(ROW);
     Result r = table.get(get);
     assertNull(
-      "Got an unexpected number of rows - no data should be returned with the NoDataFromScan coprocessor. Found: "
-          + r, r.listCells());
+      "Got an unexpected number of rows - "
+        + "no data should be returned with the NoDataFromScan coprocessor. Found: " + r,
+      r.listCells());
 
     get = new Get(Bytes.toBytes("anotherrow"));
     r = table.get(get);
     assertNull(
-      "Got an unexpected number of rows - no data should be returned with the NoDataFromScan coprocessor Found: "
-          + r, r.listCells());
+      "Got an unexpected number of rows - "
+        + "no data should be returned with the NoDataFromScan coprocessor Found: " + r,
+      r.listCells());
 
     table.close();
     UTIL.shutdownMiniCluster();
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/favored/TestFavoredNodeAssignmentHelper.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/favored/TestFavoredNodeAssignmentHelper.java
index a23f7be0ed1..307a38a0ea4 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/favored/TestFavoredNodeAssignmentHelper.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/favored/TestFavoredNodeAssignmentHelper.java
@@ -193,8 +193,9 @@ public class TestFavoredNodeAssignmentHelper {
     // the primary can be assigned but the secondary/tertiary would be null
     Map<String,Integer> rackToServerCount = new HashMap<>();
     rackToServerCount.put("rack1", 1);
-    Triple<Map<RegionInfo, ServerName>, FavoredNodeAssignmentHelper, List<RegionInfo>>
-      primaryRSMapAndHelper = secondaryAndTertiaryRSPlacementHelper(1, rackToServerCount);
+    Triple<Map<RegionInfo, ServerName>, FavoredNodeAssignmentHelper,
+      List<RegionInfo>> primaryRSMapAndHelper =
+        secondaryAndTertiaryRSPlacementHelper(1, rackToServerCount);
     FavoredNodeAssignmentHelper helper = primaryRSMapAndHelper.getSecond();
     Map<RegionInfo, ServerName> primaryRSMap = primaryRSMapAndHelper.getFirst();
     List<RegionInfo> regions = primaryRSMapAndHelper.getThird();
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestFixedFileTrailer.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestFixedFileTrailer.java
index 6382a0d7470..167e99e30f5 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestFixedFileTrailer.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestFixedFileTrailer.java
@@ -212,9 +212,10 @@ public class TestFixedFileTrailer {
           String msg = ex.getMessage();
           String cleanMsg = msg.replaceAll(
               "^(java(\\.[a-zA-Z]+)+:\\s+)?|\\s+\\(.*\\)\\s*$", "");
-          assertEquals("Actual exception message is \"" + msg + "\".\n" +
-              "Cleaned-up message", // will be followed by " expected: ..."
-              "Invalid HFile version: " + invalidVersion, cleanMsg);
+          // will be followed by " expected: ..."
+          assertEquals("Actual exception message is \"" + msg + "\".\nCleaned-up message",
+            "Invalid HFile version: " + invalidVersion,
+            cleanMsg);
           LOG.info("Got an expected exception: " + msg);
         }
       }
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlock.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlock.java
index 7dc55b88386..7606c20b347 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlock.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlock.java
@@ -309,28 +309,29 @@ public class TestHFileBlock {
 
   @Test
   public void testGzipCompression() throws IOException {
-    final String correctTestBlockStr =
-        "DATABLK*\\x00\\x00\\x00>\\x00\\x00\\x0F\\xA0\\xFF\\xFF\\xFF\\xFF"
-            + "\\xFF\\xFF\\xFF\\xFF"
-            + "\\x0" + ChecksumType.getDefaultChecksumType().getCode()
-            + "\\x00\\x00@\\x00\\x00\\x00\\x00["
-            // gzip-compressed block: http://www.gzip.org/zlib/rfc-gzip.html
-            + "\\x1F\\x8B"  // gzip magic signature
-            + "\\x08"  // Compression method: 8 = "deflate"
-            + "\\x00"  // Flags
-            + "\\x00\\x00\\x00\\x00"  // mtime
-            + "\\x00"  // XFL (extra flags)
-            // OS (0 = FAT filesystems, 3 = Unix). However, this field
-            // sometimes gets set to 0 on Linux and Mac, so we reset it to 3.
-            // This appears to be a difference caused by the availability
-            // (and use) of the native GZ codec.
-            + "\\x03"
-            + "\\xED\\xC3\\xC1\\x11\\x00 \\x08\\xC00DD\\xDD\\x7Fa"
-            + "\\xD6\\xE8\\xA3\\xB9K\\x84`\\x96Q\\xD3\\xA8\\xDB\\xA8e\\xD4c"
-            + "\\xD46\\xEA5\\xEA3\\xEA7\\xE7\\x00LI\\x5Cs\\xA0\\x0F\\x00\\x00"
-            + "\\x00\\x00\\x00\\x00"; //  4 byte checksum (ignored)
-    final int correctGzipBlockLength = 95;
-    final String testBlockStr = createTestBlockStr(GZ, correctGzipBlockLength, false);
+    // @formatter:off
+    String correctTestBlockStr = "DATABLK*\\x00\\x00\\x00>\\x00\\x00\\x0F\\xA0\\xFF\\xFF\\xFF\\xFF"
+      + "\\xFF\\xFF\\xFF\\xFF"
+      + "\\x0" + ChecksumType.getDefaultChecksumType().getCode()
+      + "\\x00\\x00@\\x00\\x00\\x00\\x00["
+      // gzip-compressed block: http://www.gzip.org/zlib/rfc-gzip.html
+      + "\\x1F\\x8B"  // gzip magic signature
+      + "\\x08"  // Compression method: 8 = "deflate"
+      + "\\x00"  // Flags
+      + "\\x00\\x00\\x00\\x00"  // mtime
+      + "\\x00"  // XFL (extra flags)
+      // OS (0 = FAT filesystems, 3 = Unix). However, this field
+      // sometimes gets set to 0 on Linux and Mac, so we reset it to 3.
+      // This appears to be a difference caused by the availability
+      // (and use) of the native GZ codec.
+      + "\\x03"
+      + "\\xED\\xC3\\xC1\\x11\\x00 \\x08\\xC00DD\\xDD\\x7Fa"
+      + "\\xD6\\xE8\\xA3\\xB9K\\x84`\\x96Q\\xD3\\xA8\\xDB\\xA8e\\xD4c"
+      + "\\xD46\\xEA5\\xEA3\\xEA7\\xE7\\x00LI\\x5Cs\\xA0\\x0F\\x00\\x00"
+      + "\\x00\\x00\\x00\\x00"; //  4 byte checksum (ignored)
+    // @formatter:on
+    int correctGzipBlockLength = 95;
+    String testBlockStr = createTestBlockStr(GZ, correctGzipBlockLength, false);
     // We ignore the block checksum because createTestBlockStr can change the
     // gzip header after the block is produced
     assertEquals(correctTestBlockStr.substring(0, correctGzipBlockLength - 4),
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestSnapshotProcedureSnapshotCorrupted.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestSnapshotProcedureSnapshotCorrupted.java
index f177a44e99d..0deef15732a 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestSnapshotProcedureSnapshotCorrupted.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestSnapshotProcedureSnapshotCorrupted.java
@@ -36,7 +36,7 @@ import org.junit.ClassRule;
 import org.junit.Test;
 import org.junit.experimental.categories.Category;
 
-import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.SnapshotState;
 
 @Category({ MasterTests.class, MediumTests.class })
 public class TestSnapshotProcedureSnapshotCorrupted extends TestSnapshotProcedure {
@@ -50,8 +50,9 @@ public class TestSnapshotProcedureSnapshotCorrupted extends TestSnapshotProcedur
     ProcedureExecutor<MasterProcedureEnv> procExec = master.getMasterProcedureExecutor();
     SnapshotProcedure sp = new SnapshotProcedure(procExec.getEnvironment(), snapshotProto);
     procExec.submitProcedure(sp);
-    TEST_UTIL.waitFor(60000, 500, () -> sp.getCurrentStateId() >
-      MasterProcedureProtos.SnapshotState.SNAPSHOT_CONSOLIDATE_SNAPSHOT_VALUE);
+    TEST_UTIL.waitFor(60000,
+      500,
+      () -> sp.getCurrentStateId() > SnapshotState.SNAPSHOT_CONSOLIDATE_SNAPSHOT_VALUE);
     DistributedFileSystem dfs = TEST_UTIL.getDFSCluster().getFileSystem();
     Optional<HRegion> region = TEST_UTIL.getHBaseCluster().getRegions(TABLE_NAME).stream()
       .filter(r -> !r.getStoreFileList(new byte[][] { CF }).isEmpty())
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsRegion.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsRegion.java
index b5fbc96d9f6..46c544ecde5 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsRegion.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsRegion.java
@@ -75,34 +75,24 @@ public class TestMetricsRegion {
     // test region with replica id > 0
     mr = new MetricsRegion(new MetricsRegionWrapperStub(1), new Configuration());
     agg = mr.getSource().getAggregateSource();
-    HELPER.assertGauge(
-      "namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001_metric_storeCount",
-      101, agg);
-    HELPER.assertGauge(
-      "namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001_metric_storeFileCount",
-      102, agg);
-    HELPER.assertGauge(
-      "namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001_metric_memstoreSize",
-      103, agg);
-    HELPER.assertCounter(
-      "namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001_metric_" +
-        "filteredReadRequestCount",
-      107, agg);
-    HELPER.assertCounter(
-      "namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001_metric_replicaid",
-      1, agg);
-    HELPER.assertCounter(
-      "namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001_metric_compactionsQueuedCount",
-      4, agg);
-    HELPER.assertCounter(
-      "namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001_metric_flushesQueuedCount",
-      6, agg);
-    HELPER.assertCounter(
-      "namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001_metric_maxCompactionQueueSize",
-      4, agg);
-    HELPER.assertCounter(
-      "namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001_metric_maxFlushQueueSize",
-      6, agg);
+    HELPER.assertGauge("namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001"
+      + "_metric_storeCount", 101, agg);
+    HELPER.assertGauge("namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001"
+      + "_metric_storeFileCount", 102, agg);
+    HELPER.assertGauge("namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001"
+      + "_metric_memstoreSize", 103, agg);
+    HELPER.assertCounter("namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001"
+      + "_metric_filteredReadRequestCount", 107, agg);
+    HELPER.assertCounter("namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001"
+      + "_metric_replicaid", 1, agg);
+    HELPER.assertCounter("namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001"
+      + "_metric_compactionsQueuedCount", 4, agg);
+    HELPER.assertCounter("namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001"
+      + "_metric_flushesQueuedCount", 6, agg);
+    HELPER.assertCounter("namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001"
+      + "_metric_maxCompactionQueueSize", 4, agg);
+    HELPER.assertCounter("namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001"
+      + "_metric_maxFlushQueueSize", 6, agg);
     mr.close();
   }
 }
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionMergeTransactionOnCluster.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionMergeTransactionOnCluster.java
index a5555ae09e1..a091d95fd61 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionMergeTransactionOnCluster.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionMergeTransactionOnCluster.java
@@ -393,17 +393,30 @@ public class TestRegionMergeTransactionOnCluster {
       for (Pair<RegionInfo, ServerName> p : currentRegionToServers) {
         currentRegions.add(p.getFirst());
       }
-      assertTrue(initialRegions.contains(mergedRegions.getFirst())); //this is the first region
-      assertTrue(initialRegions.contains(RegionReplicaUtil
-        .getRegionInfoForReplica(mergedRegions.getFirst(), 1))); //this is the replica of the first region
-      assertTrue(initialRegions.contains(mergedRegions.getSecond())); //this is the second region
-      assertTrue(initialRegions.contains(RegionReplicaUtil
-        .getRegionInfoForReplica(mergedRegions.getSecond(), 1))); //this is the replica of the second region
-      assertTrue(!initialRegions.contains(currentRegions.get(0))); //this is the new region
-      assertTrue(!initialRegions.contains(RegionReplicaUtil.getRegionInfoForReplica(currentRegions.get(0), 1))); //replica of the new region
-      assertTrue(currentRegions.contains(RegionReplicaUtil.getRegionInfoForReplica(currentRegions.get(0), 1))); //replica of the new region
-      assertTrue(!currentRegions.contains(RegionReplicaUtil.getRegionInfoForReplica(mergedRegions.getFirst(), 1))); //replica of the merged region
-      assertTrue(!currentRegions.contains(RegionReplicaUtil.getRegionInfoForReplica(mergedRegions.getSecond(), 1))); //replica of the merged region
+      // this is the first region
+      assertTrue(initialRegions.contains(mergedRegions.getFirst()));
+      // this is the replica of the first region
+      assertTrue(initialRegions
+        .contains(RegionReplicaUtil.getRegionInfoForReplica(mergedRegions.getFirst(), 1)));
+      // this is the second region
+      assertTrue(initialRegions.contains(mergedRegions.getSecond()));
+      // this is the replica of the second region
+      assertTrue(initialRegions
+        .contains(RegionReplicaUtil.getRegionInfoForReplica(mergedRegions.getSecond(), 1)));
+      // this is the new region
+      assertTrue(!initialRegions.contains(currentRegions.get(0)));
+      // replica of the new region
+      assertTrue(!initialRegions
+        .contains(RegionReplicaUtil.getRegionInfoForReplica(currentRegions.get(0), 1)));
+      // replica of the new region
+      assertTrue(currentRegions
+        .contains(RegionReplicaUtil.getRegionInfoForReplica(currentRegions.get(0), 1)));
+      // replica of the merged region
+      assertTrue(!currentRegions
+        .contains(RegionReplicaUtil.getRegionInfoForReplica(mergedRegions.getFirst(), 1)));
+      // replica of the merged region
+      assertTrue(!currentRegions
+        .contains(RegionReplicaUtil.getRegionInfoForReplica(mergedRegions.getSecond(), 1)));
       table.close();
     } finally {
       TEST_UTIL.deleteTable(tableName);
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreScanner.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreScanner.java
index c7ed4e858e2..e17ca515337 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreScanner.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreScanner.java
@@ -708,22 +708,30 @@ public class TestStoreScanner {
 
   @Test
   public void testWildCardScannerUnderDeletes() throws IOException {
-    KeyValue [] kvs = new KeyValue [] {
-        create("R1", "cf", "a", 2, KeyValue.Type.Put, "dont-care"), // inc
-        // orphaned delete column.
-        create("R1", "cf", "a", 1, KeyValue.Type.DeleteColumn, "dont-care"),
-        // column b
-        create("R1", "cf", "b", 2, KeyValue.Type.Put, "dont-care"), // inc
-        create("R1", "cf", "b", 1, KeyValue.Type.Put, "dont-care"), // inc
-        // column c
-        create("R1", "cf", "c", 10, KeyValue.Type.Delete, "dont-care"),
-        create("R1", "cf", "c", 10, KeyValue.Type.Put, "dont-care"), // no
-        create("R1", "cf", "c", 9, KeyValue.Type.Put, "dont-care"),  // inc
-        // column d
-        create("R1", "cf", "d", 11, KeyValue.Type.Put, "dont-care"), // inc
-        create("R1", "cf", "d", 10, KeyValue.Type.DeleteColumn, "dont-care"),
-        create("R1", "cf", "d", 9, KeyValue.Type.Put, "dont-care"),  // no
-        create("R1", "cf", "d", 8, KeyValue.Type.Put, "dont-care"),  // no
+    KeyValue[] kvs = new KeyValue[] {
+      // inc
+      create("R1", "cf", "a", 2, KeyValue.Type.Put, "dont-care"),
+      // orphaned delete column.
+      create("R1", "cf", "a", 1, KeyValue.Type.DeleteColumn, "dont-care"),
+      // column b
+      // inc
+      create("R1", "cf", "b", 2, KeyValue.Type.Put, "dont-care"),
+      // inc
+      create("R1", "cf", "b", 1, KeyValue.Type.Put, "dont-care"),
+      // column c
+      create("R1", "cf", "c", 10, KeyValue.Type.Delete, "dont-care"),
+      // no
+      create("R1", "cf", "c", 10, KeyValue.Type.Put, "dont-care"),
+      // inc
+      create("R1", "cf", "c", 9, KeyValue.Type.Put, "dont-care"),
+      // column d
+      // inc
+      create("R1", "cf", "d", 11, KeyValue.Type.Put, "dont-care"),
+      create("R1", "cf", "d", 10, KeyValue.Type.DeleteColumn, "dont-care"),
+      // no
+      create("R1", "cf", "d", 9, KeyValue.Type.Put, "dont-care"),
+      // no
+      create("R1", "cf", "d", 8, KeyValue.Type.Put, "dont-care"),
 
     };
     List<KeyValueScanner> scanners = scanFixture(kvs);
@@ -926,6 +934,7 @@ public class TestStoreScanner {
           return now;
         }
       });
+      // @formatter:off
       KeyValue[] kvs = new KeyValue[]{
         /*0*/ new KeyValue(Bytes.toBytes("R1"), Bytes.toBytes("cf"), null,
         now - 100, KeyValue.Type.DeleteFamily), // live
@@ -944,22 +953,23 @@ public class TestStoreScanner {
         /*7*/ create("R1", "cf", "a",
         now - 100, KeyValue.Type.DeleteColumn, "dont-care"), // max-version
         /*8*/ create("R1", "cf", "b", now - 600,
-        KeyValue.Type.DeleteColumn, "dont-care"), //expired
+        KeyValue.Type.DeleteColumn, "dont-care"), // expired
         /*9*/ create("R1", "cf", "b", now - 70,
-        KeyValue.Type.Put, "v2"), //live
+        KeyValue.Type.Put, "v2"), // live
         /*10*/ create("R1", "cf", "b", now - 750,
-        KeyValue.Type.Put, "v1"), //expired
+        KeyValue.Type.Put, "v1"), // expired
         /*11*/ create("R1", "cf", "c", now - 500,
-        KeyValue.Type.Delete, "dontcare"), //expired
+        KeyValue.Type.Delete, "dontcare"), // expired
         /*12*/ create("R1", "cf", "c", now - 600,
-        KeyValue.Type.Put, "v1"), //expired
+        KeyValue.Type.Put, "v1"), // expired
         /*13*/ create("R1", "cf", "c", now - 1000,
-        KeyValue.Type.Delete, "dontcare"), //expired
+        KeyValue.Type.Delete, "dontcare"), // expired
         /*14*/ create("R1", "cf", "d", now - 60,
-        KeyValue.Type.Put, "expired put"), //live
+        KeyValue.Type.Put, "expired put"), // live
         /*15*/ create("R1", "cf", "d", now - 100,
-        KeyValue.Type.Delete, "not-expired delete"), //live
+        KeyValue.Type.Delete, "not-expired delete"), // live
       };
+      // @formatter:on
       List<KeyValueScanner> scanners = scanFixture(kvs);
       ScanInfo scanInfo = new ScanInfo(CONF, Bytes.toBytes("cf"),
         0 /* minVersions */,
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/throttle/TestCompactionWithThroughputController.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/throttle/TestCompactionWithThroughputController.java
index 87a37697ec5..d40631fdbda 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/throttle/TestCompactionWithThroughputController.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/throttle/TestCompactionWithThroughputController.java
@@ -17,6 +17,9 @@
  */
 package org.apache.hadoop.hbase.regionserver.throttle;
 
+import static org.apache.hadoop.hbase.regionserver.throttle.PressureAwareCompactionThroughputController.HBASE_HSTORE_COMPACTION_MAX_THROUGHPUT_HIGHER_BOUND;
+import static org.apache.hadoop.hbase.regionserver.throttle.PressureAwareCompactionThroughputController.HBASE_HSTORE_COMPACTION_MAX_THROUGHPUT_LOWER_BOUND;
+import static org.apache.hadoop.hbase.regionserver.throttle.PressureAwareCompactionThroughputController.HBASE_HSTORE_COMPACTION_THROUGHPUT_TUNE_PERIOD;
 import static org.junit.Assert.assertEquals;
 import static org.junit.Assert.assertTrue;
 
@@ -111,14 +114,8 @@ public class TestCompactionWithThroughputController {
     conf.setInt(CompactionConfiguration.HBASE_HSTORE_COMPACTION_MIN_KEY, 100);
     conf.setInt(CompactionConfiguration.HBASE_HSTORE_COMPACTION_MAX_KEY, 200);
     conf.setInt(HStore.BLOCKING_STOREFILES_KEY, 10000);
-    conf.setLong(
-      PressureAwareCompactionThroughputController
-        .HBASE_HSTORE_COMPACTION_MAX_THROUGHPUT_HIGHER_BOUND,
-      throughputLimit);
-    conf.setLong(
-      PressureAwareCompactionThroughputController
-        .HBASE_HSTORE_COMPACTION_MAX_THROUGHPUT_LOWER_BOUND,
-      throughputLimit);
+    conf.setLong(HBASE_HSTORE_COMPACTION_MAX_THROUGHPUT_HIGHER_BOUND, throughputLimit);
+    conf.setLong(HBASE_HSTORE_COMPACTION_MAX_THROUGHPUT_LOWER_BOUND, throughputLimit);
     conf.set(CompactionThroughputControllerFactory.HBASE_THROUGHPUT_CONTROLLER_KEY,
       PressureAwareCompactionThroughputController.class.getName());
     TEST_UTIL.startMiniCluster(1);
@@ -183,21 +180,13 @@ public class TestCompactionWithThroughputController {
   public void testThroughputTuning() throws Exception {
     Configuration conf = TEST_UTIL.getConfiguration();
     conf.set(StoreEngine.STORE_ENGINE_CLASS_KEY, DefaultStoreEngine.class.getName());
-    conf.setLong(
-      PressureAwareCompactionThroughputController
-        .HBASE_HSTORE_COMPACTION_MAX_THROUGHPUT_HIGHER_BOUND,
-      20L * 1024 * 1024);
-    conf.setLong(
-      PressureAwareCompactionThroughputController
-        .HBASE_HSTORE_COMPACTION_MAX_THROUGHPUT_LOWER_BOUND,
-      10L * 1024 * 1024);
+    conf.setLong(HBASE_HSTORE_COMPACTION_MAX_THROUGHPUT_HIGHER_BOUND, 20L * 1024 * 1024);
+    conf.setLong(HBASE_HSTORE_COMPACTION_MAX_THROUGHPUT_LOWER_BOUND, 10L * 1024 * 1024);
     conf.setInt(CompactionConfiguration.HBASE_HSTORE_COMPACTION_MIN_KEY, 4);
     conf.setInt(HStore.BLOCKING_STOREFILES_KEY, 6);
     conf.set(CompactionThroughputControllerFactory.HBASE_THROUGHPUT_CONTROLLER_KEY,
       PressureAwareCompactionThroughputController.class.getName());
-    conf.setInt(
-      PressureAwareCompactionThroughputController.HBASE_HSTORE_COMPACTION_THROUGHPUT_TUNE_PERIOD,
-      1000);
+    conf.setInt(HBASE_HSTORE_COMPACTION_THROUGHPUT_TUNE_PERIOD, 1000);
     TEST_UTIL.startMiniCluster(1);
     Connection conn = ConnectionFactory.createConnection(conf);
     try {
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestRefreshPeerWhileRegionServerRestarts.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestRefreshPeerWhileRegionServerRestarts.java
index 4b4e6587aa3..73747c55deb 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestRefreshPeerWhileRegionServerRestarts.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestRefreshPeerWhileRegionServerRestarts.java
@@ -38,7 +38,7 @@ import org.junit.ClassRule;
 import org.junit.Test;
 import org.junit.experimental.categories.Category;
 
-import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.PeerModificationState;
 
 /**
  * This UT is used to make sure that we will not accidentally change the way to generate online
@@ -95,8 +95,8 @@ public class TestRefreshPeerWhileRegionServerRestarts extends TestReplicationBas
       UTIL1.waitFor(30000, () -> {
         for (Procedure<?> proc : UTIL1.getMiniHBaseCluster().getMaster().getProcedures()) {
           if (proc instanceof DisablePeerProcedure) {
-            return ((DisablePeerProcedure) proc).getCurrentStateId() ==
-              MasterProcedureProtos.PeerModificationState.POST_PEER_MODIFICATION_VALUE;
+            return ((DisablePeerProcedure) proc)
+              .getCurrentStateId() == PeerModificationState.POST_PEER_MODIFICATION_VALUE;
           }
         }
         return false;
diff --git a/hbase-zookeeper/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZKMulti.java b/hbase-zookeeper/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZKMulti.java
index 602eff87acd..b0eb375611e 100644
--- a/hbase-zookeeper/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZKMulti.java
+++ b/hbase-zookeeper/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZKMulti.java
@@ -288,11 +288,15 @@ public class TestZKMulti {
     // test that, even with operations that fail, the ones that would pass will pass
     // with runSequentialOnMultiFailure
     ops = new LinkedList<>();
-    ops.add(ZKUtilOp.setData(path1, Bytes.add(Bytes.toBytes(path1), Bytes.toBytes(path1)))); // pass
-    ops.add(ZKUtilOp.deleteNodeFailSilent(path2)); // pass
-    ops.add(ZKUtilOp.deleteNodeFailSilent(path3)); // fail -- node doesn't exist
-    ops.add(ZKUtilOp.createAndFailSilent(path4,
-      Bytes.add(Bytes.toBytes(path4), Bytes.toBytes(path4)))); // pass
+    // pass
+    ops.add(ZKUtilOp.setData(path1, Bytes.add(Bytes.toBytes(path1), Bytes.toBytes(path1))));
+    // pass
+    ops.add(ZKUtilOp.deleteNodeFailSilent(path2));
+    // fail -- node doesn't exist
+    ops.add(ZKUtilOp.deleteNodeFailSilent(path3));
+    // pass
+    ops.add(
+      ZKUtilOp.createAndFailSilent(path4, Bytes.add(Bytes.toBytes(path4), Bytes.toBytes(path4))));
     ZKUtil.multiOrSequential(zkw, ops, true);
     assertTrue(Bytes.equals(ZKUtil.getData(zkw, path1),
       Bytes.add(Bytes.toBytes(path1), Bytes.toBytes(path1))));