You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@geode.apache.org by ud...@apache.org on 2018/08/21 00:03:42 UTC

[geode] 01/02: Stupid spotless

This is an automated email from the ASF dual-hosted git repository.

udo pushed a commit to branch feature/Micrometer
in repository https://gitbox.apache.org/repos/asf/geode.git

commit 48bb8ae0ee0656d2def75703e61db22cf321da51
Author: Udo Kohlmeyer <uk...@pivotal.io>
AuthorDate: Mon Aug 20 09:58:41 2018 -0700

    Stupid spotless
---
 .../internal/common/AbstractSessionCache.java      |   5 +-
 .../session/catalina/AbstractSessionCache.java     |   5 +-
 .../catalina/internal/DeltaSessionStatistics.java  |   4 +-
 .../geode/modules/util/ModuleStatistics.java       |  24 +-
 .../management/MemoryThresholdsDUnitTest.java      |   3 +-
 .../geode/cache30/ClientMembershipDUnitTest.java   |   4 +-
 .../cache/DeltaPropagationStatsDUnitTest.java      |   1 -
 .../cache/PartitionedRegionStatsDUnitTest.java     |   2 +-
 .../execute/FunctionServiceStatsDUnitTest.java     |  33 +-
 .../compression/CompressionStatsDUnitTest.java     |   1 -
 .../statistics/StatisticsDistributedTest.java      |  25 +-
 .../client/internal/QueueManagerJUnitTest.java     |   6 +-
 .../pooling/ConnectionManagerJUnitTest.java        |   6 +-
 .../gms/fd/GMSHealthMonitorJUnitTest.java          |   4 +-
 .../gms/locator/GMSLocatorRecoveryJUnitTest.java   |   8 +-
 .../gms/messenger/JGroupsMessengerJUnitTest.java   |   4 +-
 .../internal/cache/ComplexDiskRegionJUnitTest.java |   2 +-
 .../internal/cache/DiskInitFileJUnitTest.java      |   2 +-
 .../geode/internal/cache/OplogJUnitTest.java       |   2 +-
 .../geode/internal/cache/OplogRVVJUnitTest.java    |   2 +-
 .../cache/PartitionedRegionStatsJUnitTest.java     |   2 +-
 .../cache/ha/HARegionQueueIntegrationTest.java     |   7 +-
 .../CacheClientNotifierIntegrationTest.java        |   7 +-
 .../CacheServerMaxConnectionsJUnitTest.java        |   4 +-
 .../ClientHealthMonitorIntegrationTest.java        |   7 +-
 .../sockets/ServerConnectionIntegrationTest.java   |   6 +-
 .../statistics/DiskSpaceLimitIntegrationTest.java  |   9 +-
 ...DistributedSystemStatisticsIntegrationTest.java |   2 +-
 ...ributedSystemStatisticsTypeIntegrationTest.java |   2 +-
 .../statistics/FileSizeLimitIntegrationTest.java   |   7 +-
 .../GemFireStatSamplerIntegrationTest.java         |  11 +-
 .../internal/statistics/LinuxSystemStatsTest.java  |   2 +-
 .../SimpleStatSamplerIntegrationTest.java          |   4 +-
 ...rchiveWithConsecutiveResourceInstGenerator.java |   6 +-
 ...chiveWithMissingResourceTypeRegressionTest.java |   3 +-
 .../StatArchiveWriterReaderIntegrationTest.java    |   4 +-
 .../statistics/StatSamplerIntegrationTest.java     |   4 +-
 .../StatTypesAreRolledOverRegressionTest.java      |   9 +-
 .../statistics/ValueMonitorIntegrationTest.java    |   6 +-
 .../internal/stats50/AtomicStatsJUnitTest.java     |   7 +-
 .../bean/stats/AsyncEventQueueStatsJUnitTest.java  |   2 +-
 .../bean/stats/CacheServerStatsJUnitTest.java      |   3 +-
 .../stats/DistributedSystemStatsJUnitTest.java     |   6 +-
 .../bean/stats/GatewayReceiverStatsJUnitTest.java  |   5 +-
 .../bean/stats/MemberLevelStatsJUnitTest.java      |  14 +-
 .../bean/stats/RegionStatsJUnitTest.java           |   8 +-
 .../internal/beans/ManagementAdapterTest.java      |   6 +-
 .../asyncqueue/internal/AsyncEventQueueStats.java  |  10 +-
 .../internal/ParallelAsyncEventQueueImpl.java      |   5 +-
 .../internal/SerialAsyncEventQueueImpl.java        |   5 +-
 .../cache/client/internal/ConnectionStatsImpl.java | 106 +++-
 .../cache/client/internal/EndpointManagerImpl.java |  21 +-
 .../cache/client/internal/ExecuteFunctionOp.java   |   1 -
 .../client/internal/ExecuteRegionFunctionOp.java   |   7 +-
 .../internal/ExecuteRegionFunctionSingleHopOp.java |   7 +-
 .../geode/cache/client/internal/PoolImpl.java      |  25 +-
 .../cache/query/internal/CqQueryVsdStatsImpl.java  |  10 +-
 .../query/internal/index/AbstractMapIndex.java     |   3 +-
 .../query/internal/index/CompactRangeIndex.java    |   3 +-
 .../cache/query/internal/index/HashIndex.java      |   3 +-
 .../cache/query/internal/index/IndexStats.java     |   6 +-
 .../query/internal/index/PartitionedIndex.java     |   3 +-
 .../cache/query/internal/index/RangeIndex.java     |   3 +-
 .../geode/cache/server/internal/LoadMonitor.java   |   7 +-
 .../geode/distributed/DistributedSystem.java       |   6 +-
 .../internal/ClusterDistributionManager.java       | 141 +++--
 .../internal/DistributionStatsImpl.java            |  96 ++--
 .../internal/InternalDistributedSystem.java        |  79 ++-
 .../distributed/internal/InternalLocator.java      |  62 +-
 .../distributed/internal/LocatorStatsImpl.java     |   6 +-
 .../distributed/internal/locks/DLockService.java   |   5 +-
 .../distributed/internal/locks/DLockStatsImpl.java |  64 ++-
 .../membership/gms/messenger/JGroupsMessenger.java |   1 -
 .../distributed/internal/tcpserver/TcpServer.java  |   1 -
 .../geode/internal/admin/ClientStatsManager.java   |  12 +-
 .../geode/internal/admin/StatAlertDefinition.java  |   2 +-
 .../geode/internal/admin/StatAlertsManager.java    |   9 +-
 .../remote/FetchResourceAttributesResponse.java    |   5 +-
 .../internal/admin/remote/FetchStatsResponse.java  |   2 -
 .../geode/internal/admin/remote/RemoteStat.java    |   4 +-
 .../internal/admin/remote/RemoteStatResource.java  |   2 +-
 .../admin/statalerts/BaseDecoratorImpl.java        |   2 +-
 .../admin/statalerts/DummyStatisticInfoImpl.java   |   2 +-
 .../admin/statalerts/FunctionDecoratorImpl.java    |   2 +-
 .../statalerts/GaugeThresholdDecoratorImpl.java    |   2 +-
 .../admin/statalerts/MultiAttrDefinitionImpl.java  |   4 +-
 .../statalerts/NumberThresholdDecoratorImpl.java   |   2 +-
 .../admin/statalerts/SingleAttrDefinitionImpl.java |   4 +-
 .../admin/statalerts/StatisticInfoImpl.java        |   2 +-
 .../geode/internal/cache/AbstractDiskRegion.java   |   2 +-
 .../geode/internal/cache/AbstractRegion.java       |   2 +-
 .../geode/internal/cache/CachePerfStatsImpl.java   | 173 +++---
 .../internal/cache/DiskDirectoryStatsImpl.java     |  14 +-
 .../geode/internal/cache/DiskRegionStatsImpl.java  |   8 +-
 .../apache/geode/internal/cache/DiskStoreImpl.java |   2 +-
 .../geode/internal/cache/DiskStoreStatsImpl.java   |  14 +-
 .../internal/cache/DistTXRollbackMessage.java      |   1 -
 .../geode/internal/cache/DummyCachePerfStats.java  |   2 +-
 .../geode/internal/cache/EvictableRegion.java      |   2 +-
 .../geode/internal/cache/GemFireCacheImpl.java     |   6 +-
 .../geode/internal/cache/InternalRegion.java       |   3 +-
 .../apache/geode/internal/cache/LocalRegion.java   | 628 ++++++++++++---------
 .../cache/OfflineCompactionDiskRegion.java         |   3 +-
 .../geode/internal/cache/PartitionedRegion.java    |   5 +-
 .../internal/cache/PartitionedRegionDataStore.java | 287 ++++++----
 .../internal/cache/PartitionedRegionHelper.java    |   3 +-
 .../internal/cache/PartitionedRegionStatsImpl.java |   6 +-
 .../apache/geode/internal/cache/PoolStatsImpl.java |   4 +-
 .../internal/cache/TXRemoteCommitMessage.java      |   1 -
 .../geode/internal/cache/ValidatingDiskRegion.java |  22 +-
 .../client/protocol/ClientProtocolService.java     |   2 +-
 .../internal/cache/control/HeapMemoryMonitor.java  |  12 +-
 .../cache/control/InternalResourceManager.java     |   3 +-
 .../cache/control/ResourceManagerStatsImpl.java    |  10 +-
 .../cache/eviction/AbstractEvictionController.java |  11 +-
 .../cache/eviction/CountLRUStatisticsImpl.java     |  19 +-
 .../cache/eviction/HeapLRUStatisticsImpl.java      |   7 +-
 .../cache/eviction/MemoryLRUStatisticsImpl.java    |   7 +-
 .../cache/execute/FunctionServiceStats.java        |   7 +-
 .../internal/cache/execute/FunctionStatsImpl.java  |  28 +-
 .../cache/execute/MemberFunctionResultSender.java  |  12 +-
 .../PartitionedRegionFunctionResultSender.java     |  18 +-
 .../execute/ServerRegionFunctionExecutor.java      |   6 +-
 .../geode/internal/cache/ha/HARegionQueue.java     |   5 +-
 .../internal/cache/ha/HARegionQueueStatsImpl.java  |  12 +-
 .../cache/partitioned/BucketSizeMessage.java       |   1 -
 .../cache/partitioned/CreateBucketMessage.java     |   1 -
 .../partitioned/CreateMissingBucketsTask.java      |   1 -
 .../internal/cache/partitioned/GetMessage.java     |  38 +-
 .../partitioned/ManageBackupBucketMessage.java     |   1 -
 .../internal/cache/tier/sockets/AcceptorImpl.java  |   3 +-
 .../internal/cache/tier/sockets/BaseCommand.java   |   1 -
 .../cache/tier/sockets/BaseCommandQuery.java       |   1 -
 .../cache/tier/sockets/CacheClientNotifier.java    | 177 +++---
 .../tier/sockets/CacheClientNotifierStatsImpl.java |  56 +-
 .../cache/tier/sockets/CacheClientProxy.java       |   2 +-
 .../cache/tier/sockets/CacheClientProxyStats.java  |   9 +-
 .../cache/tier/sockets/CacheClientUpdater.java     |  12 +-
 .../cache/tier/sockets/CacheServerStatsImpl.java   |  16 +-
 .../tier/sockets/OriginalServerConnection.java     |   6 +-
 .../tier/sockets/ProtobufServerConnection.java     |   6 +-
 .../cache/tier/sockets/ServerConnection.java       |   4 +-
 .../tier/sockets/ServerConnectionFactory.java      |  17 +-
 .../cache/tier/sockets/command/ClearRegion.java    |   2 -
 .../cache/tier/sockets/command/ClientReady.java    |   2 -
 .../tier/sockets/command/CloseConnection.java      |   2 -
 .../cache/tier/sockets/command/ContainsKey.java    |   2 -
 .../cache/tier/sockets/command/ContainsKey66.java  |   2 -
 .../cache/tier/sockets/command/Destroy.java        |   2 -
 .../cache/tier/sockets/command/Destroy65.java      |   2 -
 .../cache/tier/sockets/command/DestroyRegion.java  |   2 -
 .../sockets/command/GatewayReceiverCommand.java    |   1 -
 .../internal/cache/tier/sockets/command/Get70.java |   2 -
 .../cache/tier/sockets/command/Invalidate.java     |   2 -
 .../internal/cache/tier/sockets/command/Ping.java  |   1 -
 .../internal/cache/tier/sockets/command/Put.java   |   2 -
 .../internal/cache/tier/sockets/command/Put61.java |   2 -
 .../internal/cache/tier/sockets/command/Put65.java |   2 -
 .../cache/tier/sockets/command/PutAll.java         |   2 -
 .../cache/tier/sockets/command/PutAll70.java       |   2 -
 .../cache/tier/sockets/command/PutAll80.java       |   2 -
 .../cache/tier/sockets/command/RemoveAll.java      |   2 -
 .../cache/tier/sockets/command/Request.java        |   3 -
 .../internal/cache/tier/sockets/command/Size.java  |   2 -
 .../sockets/command/UpdateClientNotification.java  |   2 -
 .../internal/cache/tx/PartitionedTXRegionStub.java |   1 -
 .../internal/cache/tx/RemoteClearMessage.java      |   1 -
 .../cache/tx/RemoteContainsKeyValueMessage.java    |   1 -
 .../internal/cache/tx/RemoteFetchKeysMessage.java  |   1 -
 .../geode/internal/cache/tx/RemoteGetMessage.java  |   1 -
 .../geode/internal/cache/tx/RemoteSizeMessage.java |   1 -
 .../internal/cache/wan/AbstractGatewaySender.java  |   6 +-
 .../cache/wan/GatewayReceiverStatsImpl.java        |  13 +-
 .../internal/cache/wan/GatewaySenderStatsImpl.java |  50 +-
 .../apache/geode/internal/concurrent/Atomics.java  |   5 +-
 .../internal/offheap/MemoryAllocatorImpl.java      |   6 +-
 .../geode/internal/offheap/OffHeapStorage.java     |  31 +-
 .../internal/offheap/OffHeapStorageStatsImpl.java  |  17 +-
 .../statistics/AbstractStatisticsFactory.java      |  24 +-
 .../geode/internal/statistics/CallbackSampler.java |   2 +-
 .../internal/statistics/GemFireStatSampler.java    |  27 +-
 .../geode/internal/statistics/HostStatHelper.java  |  29 +-
 .../geode/internal/statistics/HostStatSampler.java |   2 +-
 .../statistics/InternalDistributedSystemStats.java |  43 +-
 .../statistics/LocalStatisticsFactory.java         |  11 +-
 .../internal/statistics/LocalStatisticsImpl.java   |   2 +-
 .../geode/internal/statistics/SampleCollector.java |   6 +-
 .../internal/statistics/StatArchiveWriter.java     |   2 +-
 .../internal/statistics/StatSamplerStats.java      |   5 +-
 .../statistics/StatisticDescriptorImpl.java        |   2 +-
 .../geode/internal/statistics/StatisticsImpl.java  |   6 +-
 .../statistics/StatisticsTypeFactoryImpl.java      |  14 +-
 .../internal/statistics/StatisticsTypeImpl.java    |   2 +-
 .../internal/statistics/StatisticsTypeXml.java     |   6 +-
 .../apache/geode/internal/statistics/VMStats.java  |   6 +-
 .../statistics/VMStatsContractFactory.java         |   2 +-
 .../geode/internal/statistics/ValueMonitor.java    |   2 +-
 .../statistics/platform/LinuxProcessStats.java     |   8 +-
 .../statistics/platform/LinuxSystemStats.java      |  15 +-
 .../statistics/platform/OSXProcessStats.java       |  11 +-
 .../statistics/platform/OSXSystemStats.java        |   6 +-
 .../statistics/platform/SolarisProcessStats.java   |  15 +-
 .../statistics/platform/SolarisSystemStats.java    |  22 +-
 .../statistics/platform/WindowsProcessStats.java   |  99 ++--
 .../statistics/platform/WindowsSystemStats.java    |   9 +-
 .../internal/stats50/Atomic50StatisticsImpl.java   |   4 +-
 .../apache/geode/internal/stats50/VMStats50.java   |  38 +-
 .../org/apache/geode/internal/tcp/Connection.java  |   1 -
 .../management/internal/FederatingManager.java     |  14 +-
 .../geode/management/internal/LocalManager.java    |   9 +-
 .../internal/beans/AsyncEventQueueMBeanBridge.java |   2 +-
 .../internal/beans/GatewaySenderMBeanBridge.java   |   2 +-
 .../internal/beans/MemberMBeanBridge.java          |   3 +-
 .../internal/beans/RegionMBeanBridge.java          |   2 +-
 .../management/internal/beans/ServerBridge.java    |   1 -
 .../beans/stats/AggregateRegionStatsMonitor.java   |   4 +-
 .../internal/beans/stats/GCStatsMonitor.java       |   2 +-
 .../beans/stats/GatewaySenderOverflowMonitor.java  |   4 +-
 .../internal/beans/stats/MBeanStatsMonitor.java    |   6 +-
 .../beans/stats/MemberLevelDiskMonitor.java        |   4 +-
 .../internal/beans/stats/VMStatsMonitor.java       |   2 +-
 .../tier/sockets/ServerConnectionFactoryTest.java  |   3 +-
 .../internal/offheap/OffHeapStorageJUnitTest.java  |   6 +-
 .../internal/statistics/SampleCollectorTest.java   |   6 +-
 .../internal/statistics/StatSamplerTestCase.java   |   4 +-
 .../internal/statistics/TestStatisticsManager.java |   4 +-
 .../management/bean/stats/MBeanStatsTestCase.java  |   6 +-
 .../management/bean/stats/StatsRateJUnitTest.java  |   2 +-
 .../beans/stats/MBeanStatsMonitorTest.java         |   6 +-
 .../query/cq/dunit/CqStatsUsingPoolDUnitTest.java  |   3 +-
 .../geode/cache/query/internal/cq/CqQueryImpl.java |   2 +-
 .../cache/query/internal/cq/CqServiceImpl.java     |   6 +-
 .../query/internal/cq/CqServiceVsdStatsImpl.java   |  26 +-
 .../cache/tier/sockets/command/CloseCQ.java        |   2 -
 .../cache/tier/sockets/command/ExecuteCQ.java      |   2 -
 .../cache/tier/sockets/command/ExecuteCQ61.java    |   2 -
 .../cache/tier/sockets/command/GetCQStats.java     |   2 -
 .../cache/tier/sockets/command/GetDurableCQs.java  |   1 -
 .../cache/tier/sockets/command/StopCQ.java         |   2 -
 .../dunit/internal/JUnit4DistributedTestCase.java  |   2 +-
 .../java/org/apache/geode/test/fake/Fakes.java     |   8 +-
 .../internal/LuceneIndexForPartitionedRegion.java  |   3 +-
 .../cache/lucene/internal/LuceneIndexImpl.java     |   3 +-
 .../cache/lucene/internal/LuceneIndexStats.java    |   9 +-
 .../internal/filesystem/FileSystemStats.java       |   2 -
 .../IndexRepositoryImplPerformanceTest.java        |   9 +-
 .../internal/LuceneServiceImplJUnitTest.java       |   4 +-
 .../v1/acceptance/LocatorConnectionDUnitTest.java  |   8 +-
 .../protobuf/v1/AuthenticationIntegrationTest.java |   5 +-
 .../v1/acceptance/CacheConnectionJUnitTest.java    |   5 +-
 .../statistics/ProtobufClientStatistics.java       |   2 +-
 .../protobuf/v1/ProtobufProtocolService.java       |   2 +-
 .../util/AutoBalancerIntegrationJUnitTest.java     |   3 +-
 .../cache/client/internal/ConnectionStats.java     |   2 -
 .../geode/cache/internal/cache/CachePerfStats.java |   1 +
 .../geode/statistics/GFSStatsImplementor.java      |  12 +-
 .../geode/statistics/StatisticsTypeFactory.java    |  21 +-
 .../org/apache/geode/statistics/StatsFactory.java  |  84 ++-
 .../apache/geode/statistics/StatsImplementor.java  |   2 +
 ...ueueOverflowMBeanAttributesDistributedTest.java |   7 +-
 ...nderOverflowMBeanAttributesDistributedTest.java |   7 +-
 .../geode/internal/cache/wan/WANTestBase.java      |   2 -
 .../KeepEventsOnGatewaySenderQueueDUnitTest.java   |   1 -
 263 files changed, 2079 insertions(+), 1600 deletions(-)

diff --git a/extensions/geode-modules-session-internal/src/main/java/org/apache/geode/modules/session/internal/common/AbstractSessionCache.java b/extensions/geode-modules-session-internal/src/main/java/org/apache/geode/modules/session/internal/common/AbstractSessionCache.java
index 602c6cb..afc73da 100644
--- a/extensions/geode-modules-session-internal/src/main/java/org/apache/geode/modules/session/internal/common/AbstractSessionCache.java
+++ b/extensions/geode-modules-session-internal/src/main/java/org/apache/geode/modules/session/internal/common/AbstractSessionCache.java
@@ -70,8 +70,9 @@ public abstract class AbstractSessionCache implements SessionCache {
   }
 
   protected void createStatistics() {
-    this.statistics = new DeltaSessionStatistics(getCache().getDistributedSystem().getStatisticsFactory(),
-        (String) properties.get(CacheProperty.STATISTICS_NAME));
+    this.statistics =
+        new DeltaSessionStatistics(getCache().getDistributedSystem().getStatisticsFactory(),
+            (String) properties.get(CacheProperty.STATISTICS_NAME));
   }
 
   /**
diff --git a/extensions/geode-modules/src/main/java/org/apache/geode/modules/session/catalina/AbstractSessionCache.java b/extensions/geode-modules/src/main/java/org/apache/geode/modules/session/catalina/AbstractSessionCache.java
index 8d2a172..10717bb 100644
--- a/extensions/geode-modules/src/main/java/org/apache/geode/modules/session/catalina/AbstractSessionCache.java
+++ b/extensions/geode-modules/src/main/java/org/apache/geode/modules/session/catalina/AbstractSessionCache.java
@@ -94,8 +94,9 @@ public abstract class AbstractSessionCache implements SessionCache {
   }
 
   protected void createStatistics() {
-    this.statistics = new DeltaSessionStatistics(getCache().getDistributedSystem().getStatisticsFactory(),
-        getSessionManager().getStatisticsName());
+    this.statistics =
+        new DeltaSessionStatistics(getCache().getDistributedSystem().getStatisticsFactory(),
+            getSessionManager().getStatisticsName());
   }
 
   protected RegionConfiguration createRegionConfiguration() {
diff --git a/extensions/geode-modules/src/main/java/org/apache/geode/modules/session/catalina/internal/DeltaSessionStatistics.java b/extensions/geode-modules/src/main/java/org/apache/geode/modules/session/catalina/internal/DeltaSessionStatistics.java
index 2c05869..e9e5c2c 100644
--- a/extensions/geode-modules/src/main/java/org/apache/geode/modules/session/catalina/internal/DeltaSessionStatistics.java
+++ b/extensions/geode-modules/src/main/java/org/apache/geode/modules/session/catalina/internal/DeltaSessionStatistics.java
@@ -18,8 +18,6 @@ import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.statistics.StatisticsTypeFactory;
-import org.apache.geode.internal.statistics.StatisticsTypeFactoryImpl;
 
 public class DeltaSessionStatistics {
 
@@ -38,7 +36,7 @@ public class DeltaSessionStatistics {
   private void initializeStats(StatisticsFactory factory) {
     // Initialize type
     type = factory.createType(typeName, typeName,
-        new StatisticDescriptor[]{
+        new StatisticDescriptor[] {
             factory.createIntCounter(SESSIONS_CREATED, "The number of sessions created",
                 "operations"),
             factory.createIntCounter(SESSIONS_INVALIDATED,
diff --git a/extensions/geode-modules/src/main/java/org/apache/geode/modules/util/ModuleStatistics.java b/extensions/geode-modules/src/main/java/org/apache/geode/modules/util/ModuleStatistics.java
index eb3901d..f05c610 100644
--- a/extensions/geode-modules/src/main/java/org/apache/geode/modules/util/ModuleStatistics.java
+++ b/extensions/geode-modules/src/main/java/org/apache/geode/modules/util/ModuleStatistics.java
@@ -14,13 +14,11 @@
  */
 package org.apache.geode.modules.util;
 
+import org.apache.geode.distributed.DistributedSystem;
 import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.statistics.StatisticsTypeFactory;
-import org.apache.geode.distributed.DistributedSystem;
-import org.apache.geode.internal.statistics.StatisticsTypeFactoryImpl;
 
 /**
  * Statistics for modules.
@@ -37,15 +35,17 @@ public class ModuleStatistics {
   private int hibernateEntityDestroyJobsScheduledId;
 
   private void initializeStats(StatisticsFactory factory) {
-    type = factory.createType("pluginStats", "statistics for hibernate plugin and hibernate L2 cache",
-        new StatisticDescriptor[] {
-            factory.createLongCounter("cacheHits", "number of times an entity was found in L2 cache",
-                "count"),
-            factory.createLongCounter("cacheMisses",
-                "number of times an entity was NOT found in l2 cache", "count"),
-            factory.createLongCounter("hibernateEntityDestroyJobsScheduled",
-                "number of entities scheduled for destroy because of version conflict with a remote member",
-                "jobs")});
+    type =
+        factory.createType("pluginStats", "statistics for hibernate plugin and hibernate L2 cache",
+            new StatisticDescriptor[] {
+                factory.createLongCounter("cacheHits",
+                    "number of times an entity was found in L2 cache",
+                    "count"),
+                factory.createLongCounter("cacheMisses",
+                    "number of times an entity was NOT found in l2 cache", "count"),
+                factory.createLongCounter("hibernateEntityDestroyJobsScheduled",
+                    "number of entities scheduled for destroy because of version conflict with a remote member",
+                    "jobs")});
 
     cacheHitsId = type.nameToId("cacheHits");
     cacheMissesId = type.nameToId("cacheMisses");
diff --git a/geode-core/src/distributedTest/java/org/apache/geode/cache/management/MemoryThresholdsDUnitTest.java b/geode-core/src/distributedTest/java/org/apache/geode/cache/management/MemoryThresholdsDUnitTest.java
index c8e2f72..fe381d4 100644
--- a/geode-core/src/distributedTest/java/org/apache/geode/cache/management/MemoryThresholdsDUnitTest.java
+++ b/geode-core/src/distributedTest/java/org/apache/geode/cache/management/MemoryThresholdsDUnitTest.java
@@ -1745,7 +1745,8 @@ public class MemoryThresholdsDUnitTest extends ClientServerTestCase {
     Cache cache = getCache();
     InternalDistributedSystem internalSystem =
         (InternalDistributedSystem) cache.getDistributedSystem();
-    final GemFireStatSampler sampler = internalSystem.getInternalDistributedSystemStats().getStatSampler();
+    final GemFireStatSampler sampler =
+        internalSystem.getInternalDistributedSystemStats().getStatSampler();
     sampler.waitForInitialization(10000); // fix: remove infinite wait
     final LocalStatListener l = new LocalStatListener() {
       public void statValueChanged(double value) {
diff --git a/geode-core/src/distributedTest/java/org/apache/geode/cache30/ClientMembershipDUnitTest.java b/geode-core/src/distributedTest/java/org/apache/geode/cache30/ClientMembershipDUnitTest.java
index b4ef284..6e7a237 100644
--- a/geode-core/src/distributedTest/java/org/apache/geode/cache30/ClientMembershipDUnitTest.java
+++ b/geode-core/src/distributedTest/java/org/apache/geode/cache30/ClientMembershipDUnitTest.java
@@ -44,8 +44,6 @@ import org.junit.Test;
 import org.junit.experimental.categories.Category;
 
 import org.apache.geode.InternalGemFireException;
-import org.apache.geode.statistics.Statistics;
-import org.apache.geode.statistics.StatisticsType;
 import org.apache.geode.cache.AttributesFactory;
 import org.apache.geode.cache.CacheException;
 import org.apache.geode.cache.Region;
@@ -64,6 +62,8 @@ import org.apache.geode.internal.cache.tier.sockets.ServerConnection;
 import org.apache.geode.management.membership.ClientMembership;
 import org.apache.geode.management.membership.ClientMembershipEvent;
 import org.apache.geode.management.membership.ClientMembershipListener;
+import org.apache.geode.statistics.Statistics;
+import org.apache.geode.statistics.StatisticsType;
 import org.apache.geode.test.dunit.Assert;
 import org.apache.geode.test.dunit.Host;
 import org.apache.geode.test.dunit.IgnoredException;
diff --git a/geode-core/src/distributedTest/java/org/apache/geode/internal/cache/DeltaPropagationStatsDUnitTest.java b/geode-core/src/distributedTest/java/org/apache/geode/internal/cache/DeltaPropagationStatsDUnitTest.java
index 1c605f4..8cbbb89 100755
--- a/geode-core/src/distributedTest/java/org/apache/geode/internal/cache/DeltaPropagationStatsDUnitTest.java
+++ b/geode-core/src/distributedTest/java/org/apache/geode/internal/cache/DeltaPropagationStatsDUnitTest.java
@@ -41,7 +41,6 @@ import org.apache.geode.cache.client.PoolManager;
 import org.apache.geode.cache.server.CacheServer;
 import org.apache.geode.cache.util.CacheListenerAdapter;
 import org.apache.geode.distributed.DistributedSystem;
-import org.apache.geode.distributed.internal.DistributionStats;
 import org.apache.geode.distributed.internal.DistributionStatsImpl;
 import org.apache.geode.internal.AvailablePort;
 import org.apache.geode.internal.cache.tier.sockets.CacheClientNotifier;
diff --git a/geode-core/src/distributedTest/java/org/apache/geode/internal/cache/PartitionedRegionStatsDUnitTest.java b/geode-core/src/distributedTest/java/org/apache/geode/internal/cache/PartitionedRegionStatsDUnitTest.java
index 9a92210..6cc9363 100644
--- a/geode-core/src/distributedTest/java/org/apache/geode/internal/cache/PartitionedRegionStatsDUnitTest.java
+++ b/geode-core/src/distributedTest/java/org/apache/geode/internal/cache/PartitionedRegionStatsDUnitTest.java
@@ -25,13 +25,13 @@ import java.util.concurrent.TimeoutException;
 import org.junit.Before;
 import org.junit.Test;
 
-import org.apache.geode.statistics.Statistics;
 import org.apache.geode.cache.Cache;
 import org.apache.geode.cache.PartitionAttributesFactory;
 import org.apache.geode.cache.Region;
 import org.apache.geode.cache.RegionFactory;
 import org.apache.geode.cache.RegionShortcut;
 import org.apache.geode.cache.control.RebalanceOperation;
+import org.apache.geode.statistics.Statistics;
 import org.apache.geode.test.dunit.VM;
 import org.apache.geode.test.dunit.cache.CacheTestCase;
 
diff --git a/geode-core/src/distributedTest/java/org/apache/geode/internal/cache/execute/FunctionServiceStatsDUnitTest.java b/geode-core/src/distributedTest/java/org/apache/geode/internal/cache/execute/FunctionServiceStatsDUnitTest.java
index 7a47362..176ccb9 100644
--- a/geode-core/src/distributedTest/java/org/apache/geode/internal/cache/execute/FunctionServiceStatsDUnitTest.java
+++ b/geode-core/src/distributedTest/java/org/apache/geode/internal/cache/execute/FunctionServiceStatsDUnitTest.java
@@ -218,7 +218,7 @@ public class FunctionServiceStatsDUnitTest extends PRClientServerTestBase {
             try {
               int j = 0;
               HashSet origVals = new HashSet();
-              for (Iterator i = testKeysSet.iterator(); i.hasNext(); ) {
+              for (Iterator i = testKeysSet.iterator(); i.hasNext();) {
                 Integer val = new Integer(j++);
                 origVals.add(val);
                 region.put(i.next(), val);
@@ -268,8 +268,7 @@ public class FunctionServiceStatsDUnitTest extends PRClientServerTestBase {
     SerializableCallable checkStatsOnClient = new SerializableCallable("checkStatsOnClient") {
       public Object call() throws Exception {
         // checks for the aggregate stats
-        InternalDistributedSystem
-            internalDistributedSystem =
+        InternalDistributedSystem internalDistributedSystem =
             (InternalDistributedSystem) cache.getDistributedSystem();
         FunctionServiceStats functionServiceStats =
             internalDistributedSystem.getInternalDistributedSystemStats().getFunctionServiceStats();
@@ -308,8 +307,7 @@ public class FunctionServiceStatsDUnitTest extends PRClientServerTestBase {
     SerializableCallable checkStatsOnServer = new SerializableCallable("checkStatsOnClient") {
       public Object call() throws Exception {
         // checks for the aggregate stats
-        InternalDistributedSystem
-            internalDistributedSystem =
+        InternalDistributedSystem internalDistributedSystem =
             (InternalDistributedSystem) cache.getDistributedSystem();
         FunctionServiceStats functionServiceStats =
             internalDistributedSystem.getInternalDistributedSystemStats().getFunctionServiceStats();
@@ -516,8 +514,7 @@ public class FunctionServiceStatsDUnitTest extends PRClientServerTestBase {
     SerializableCallable checkStatsOnClient = new SerializableCallable("checkStatsOnClient") {
       public Object call() throws Exception {
         // checks for the aggregate stats
-        InternalDistributedSystem
-            internalDistributedSystem =
+        InternalDistributedSystem internalDistributedSystem =
             (InternalDistributedSystem) cache.getDistributedSystem();
         FunctionServiceStats functionServiceStats =
             internalDistributedSystem.getInternalDistributedSystemStats().getFunctionServiceStats();
@@ -612,8 +609,7 @@ public class FunctionServiceStatsDUnitTest extends PRClientServerTestBase {
     SerializableCallable checkStatsOnClient = new SerializableCallable("checkStatsOnClient") {
       public Object call() throws Exception {
         // checks for the aggregate stats
-        InternalDistributedSystem
-            internalDistributedSystem =
+        InternalDistributedSystem internalDistributedSystem =
             (InternalDistributedSystem) cache.getDistributedSystem();
         FunctionServiceStats functionServiceStats =
             internalDistributedSystem.getInternalDistributedSystemStats().getFunctionServiceStats();
@@ -625,8 +621,7 @@ public class FunctionServiceStatsDUnitTest extends PRClientServerTestBase {
             functionServiceStats.getFunctionExecutionsCompleted());
         assertEquals(resultReceived_Aggregate, functionServiceStats.getResultsReceived());
 
-        FunctionStats
-            functionStats =
+        FunctionStats functionStats =
             StatsFactory.createFunctionStatsImpl(internalDistributedSystem.getStatisticsFactory(),
                 TestFunction.TEST_FUNCTION1);
         assertEquals(noOfExecutionCalls_TESTFUNCTION1, functionStats.getFunctionExecutionCalls());
@@ -651,8 +646,7 @@ public class FunctionServiceStatsDUnitTest extends PRClientServerTestBase {
     SerializableCallable checkStatsOnServer = new SerializableCallable("checkStatsOnClient") {
       public Object call() throws Exception {
         // checks for the aggregate stats
-        InternalDistributedSystem
-            internalDistributedSystem =
+        InternalDistributedSystem internalDistributedSystem =
             (InternalDistributedSystem) cache.getDistributedSystem();
         FunctionServiceStats functionServiceStats =
             internalDistributedSystem.getInternalDistributedSystemStats().getFunctionServiceStats();
@@ -800,7 +794,7 @@ public class FunctionServiceStatsDUnitTest extends PRClientServerTestBase {
           testKeys.add("execKey-" + i);
         }
         int j = 0;
-        for (Iterator i = testKeys.iterator(); i.hasNext(); ) {
+        for (Iterator i = testKeys.iterator(); i.hasNext();) {
           Integer val = new Integer(j++);
           pr.put(i.next(), val);
         }
@@ -1127,8 +1121,7 @@ public class FunctionServiceStatsDUnitTest extends PRClientServerTestBase {
 
     member1.invoke(new SerializableCallable("checkFunctionExecutionStatsForMember1") {
       public Object call() throws Exception {
-        FunctionServiceStats
-            functionServiceStats =
+        FunctionServiceStats functionServiceStats =
             distributedSystem.getInternalDistributedSystemStats().getFunctionServiceStats();
         waitNoFunctionsRunning(functionServiceStats);
 
@@ -1138,8 +1131,7 @@ public class FunctionServiceStatsDUnitTest extends PRClientServerTestBase {
             functionServiceStats.getFunctionExecutionsCompleted());
         assertEquals(resultReceived_Aggregate, functionServiceStats.getResultsReceived());
 
-        FunctionStats
-            functionStats =
+        FunctionStats functionStats =
             StatsFactory.createFunctionStatsImpl(distributedSystem.getStatisticsFactory(),
                 inlineFunction.getId());
         assertEquals(noOfExecutionCalls_Inline, functionStats.getFunctionExecutionCalls());
@@ -1153,8 +1145,7 @@ public class FunctionServiceStatsDUnitTest extends PRClientServerTestBase {
     SerializableCallable checkFunctionExecutionStatsForOtherMember =
         new SerializableCallable("checkFunctionExecutionStatsForOtherMember") {
           public Object call() throws Exception {
-            FunctionServiceStats
-                functionServiceStats =
+            FunctionServiceStats functionServiceStats =
                 distributedSystem.getInternalDistributedSystemStats().getFunctionServiceStats();
             waitNoFunctionsRunning(functionServiceStats);
 
@@ -1254,7 +1245,7 @@ public class FunctionServiceStatsDUnitTest extends PRClientServerTestBase {
           testKeys.add("execKey-" + i);
         }
         int j = 0;
-        for (Iterator i = testKeys.iterator(); i.hasNext(); ) {
+        for (Iterator i = testKeys.iterator(); i.hasNext();) {
           Integer key = new Integer(j++);
           pr.put(key, i.next());
         }
diff --git a/geode-core/src/distributedTest/java/org/apache/geode/internal/compression/CompressionStatsDUnitTest.java b/geode-core/src/distributedTest/java/org/apache/geode/internal/compression/CompressionStatsDUnitTest.java
index 360b5d1..8e7a81b 100644
--- a/geode-core/src/distributedTest/java/org/apache/geode/internal/compression/CompressionStatsDUnitTest.java
+++ b/geode-core/src/distributedTest/java/org/apache/geode/internal/compression/CompressionStatsDUnitTest.java
@@ -26,7 +26,6 @@ import org.junit.Test;
 import org.apache.geode.cache.DataPolicy;
 import org.apache.geode.cache.Region;
 import org.apache.geode.compression.Compressor;
-import org.apache.geode.internal.cache.CachePerfStats;
 import org.apache.geode.internal.cache.CachePerfStatsImpl;
 import org.apache.geode.internal.cache.GemFireCacheImpl;
 import org.apache.geode.internal.cache.LocalRegion;
diff --git a/geode-core/src/distributedTest/java/org/apache/geode/internal/statistics/StatisticsDistributedTest.java b/geode-core/src/distributedTest/java/org/apache/geode/internal/statistics/StatisticsDistributedTest.java
index 5bb5259..891900b 100755
--- a/geode-core/src/distributedTest/java/org/apache/geode/internal/statistics/StatisticsDistributedTest.java
+++ b/geode-core/src/distributedTest/java/org/apache/geode/internal/statistics/StatisticsDistributedTest.java
@@ -47,11 +47,6 @@ import org.junit.Rule;
 import org.junit.Test;
 import org.junit.experimental.categories.Category;
 
-import org.apache.geode.statistics.StatisticDescriptor;
-import org.apache.geode.statistics.Statistics;
-import org.apache.geode.statistics.StatisticsFactory;
-import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.statistics.StatisticsTypeFactory;
 import org.apache.geode.cache.Cache;
 import org.apache.geode.cache.CacheListener;
 import org.apache.geode.cache.EntryEvent;
@@ -68,6 +63,11 @@ import org.apache.geode.internal.NanoTimer;
 import org.apache.geode.internal.statistics.StatArchiveReader.ResourceInst;
 import org.apache.geode.internal.statistics.StatArchiveReader.StatSpec;
 import org.apache.geode.internal.statistics.StatArchiveReader.StatValue;
+import org.apache.geode.statistics.StatisticDescriptor;
+import org.apache.geode.statistics.Statistics;
+import org.apache.geode.statistics.StatisticsFactory;
+import org.apache.geode.statistics.StatisticsType;
+import org.apache.geode.statistics.StatisticsTypeFactory;
 import org.apache.geode.test.dunit.AsyncInvocation;
 import org.apache.geode.test.dunit.VM;
 import org.apache.geode.test.dunit.cache.internal.JUnit4CacheTestCase;
@@ -237,7 +237,8 @@ public class StatisticsDistributedTest extends JUnit4CacheTestCase {
         final int pubThread = j;
         publishers[pubThread] = pubs[pubVM]
             .invokeAsync("pub-connect-and-put-data-" + pubVM + "-thread-" + pubThread, () -> {
-              PubSubStats statistics = new PubSubStats(basicGetSystem().getStatisticsFactory(), "pub-" + pubThread, pubVM);
+              PubSubStats statistics = new PubSubStats(basicGetSystem().getStatisticsFactory(),
+                  "pub-" + pubThread, pubVM);
               pubStatsRef.set(pubThread, statistics);
 
               RegionMembershipListener rml = rmlRef.get();
@@ -279,8 +280,10 @@ public class StatisticsDistributedTest extends JUnit4CacheTestCase {
               assertEquals(MAX_PUTS, statistics.getPuts());
 
               // wait for 2 samples to ensure all stats have been archived
-              StatisticsType statSamplerType = getSystem().getInternalDistributedSystemStats().findType("StatSampler");
-              Statistics[] statsArray = getSystem().getInternalDistributedSystemStats().findStatisticsByType(statSamplerType);
+              StatisticsType statSamplerType =
+                  getSystem().getInternalDistributedSystemStats().findType("StatSampler");
+              Statistics[] statsArray = getSystem().getInternalDistributedSystemStats()
+                  .findStatisticsByType(statSamplerType);
               assertEquals(1, statsArray.length);
 
               Statistics statSamplerStats = statsArray[0];
@@ -301,8 +304,10 @@ public class StatisticsDistributedTest extends JUnit4CacheTestCase {
 
     sub.invoke("sub-wait-for-samples", () -> {
       // wait for 2 samples to ensure all stats have been archived
-      StatisticsType statSamplerType = getSystem().getInternalDistributedSystemStats().findType("StatSampler");
-      Statistics[] statsArray = getSystem().getInternalDistributedSystemStats().findStatisticsByType(statSamplerType);
+      StatisticsType statSamplerType =
+          getSystem().getInternalDistributedSystemStats().findType("StatSampler");
+      Statistics[] statsArray =
+          getSystem().getInternalDistributedSystemStats().findStatisticsByType(statSamplerType);
       assertEquals(1, statsArray.length);
 
       Statistics statSamplerStats = statsArray[0];
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/cache/client/internal/QueueManagerJUnitTest.java b/geode-core/src/integrationTest/java/org/apache/geode/cache/client/internal/QueueManagerJUnitTest.java
index fcde792..186e61d 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/cache/client/internal/QueueManagerJUnitTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/cache/client/internal/QueueManagerJUnitTest.java
@@ -84,9 +84,11 @@ public class QueueManagerJUnitTest {
     properties.put(MCAST_PORT, "0");
     properties.put(LOCATORS, "");
     distributedSystem = DistributedSystem.connect(properties);
-    stats = StatsFactory.createPoolStatsImpl(distributedSystem.getStatisticsFactory(), "QueueManagerJUnitTest");
+    stats = StatsFactory.createPoolStatsImpl(distributedSystem.getStatisticsFactory(),
+        "QueueManagerJUnitTest");
     pool = new DummyPool();
-    endpoints = new EndpointManagerImpl("pool", distributedSystem, distributedSystem.getCancelCriterion(), pool.getStats());
+    endpoints = new EndpointManagerImpl("pool", distributedSystem,
+        distributedSystem.getCancelCriterion(), pool.getStats());
     source = new DummySource();
     factory = new DummyFactory();
     background = Executors.newSingleThreadScheduledExecutor();
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/cache/client/internal/pooling/ConnectionManagerJUnitTest.java b/geode-core/src/integrationTest/java/org/apache/geode/cache/client/internal/pooling/ConnectionManagerJUnitTest.java
index 9605f3b..04305e1 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/cache/client/internal/pooling/ConnectionManagerJUnitTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/cache/client/internal/pooling/ConnectionManagerJUnitTest.java
@@ -93,8 +93,10 @@ public class ConnectionManagerJUnitTest {
     properties.put(LOCATORS, "");
     distributedSystem = DistributedSystem.connect(properties);
     background = Executors.newSingleThreadScheduledExecutor();
-    poolStats = StatsFactory.createPoolStatsImpl(distributedSystem.getStatisticsFactory(), "connectionManagerJUnitTest");
-    endpointManager = new EndpointManagerImpl("pool", distributedSystem, distributedSystem.getCancelCriterion(), poolStats);
+    poolStats = StatsFactory.createPoolStatsImpl(distributedSystem.getStatisticsFactory(),
+        "connectionManagerJUnitTest");
+    endpointManager = new EndpointManagerImpl("pool", distributedSystem,
+        distributedSystem.getCancelCriterion(), poolStats);
     cancelCriterion = new CancelCriterion() {
 
       public String cancelInProgress() {
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/distributed/internal/membership/gms/fd/GMSHealthMonitorJUnitTest.java b/geode-core/src/integrationTest/java/org/apache/geode/distributed/internal/membership/gms/fd/GMSHealthMonitorJUnitTest.java
index c2f7faf..9d41cf5 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/distributed/internal/membership/gms/fd/GMSHealthMonitorJUnitTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/distributed/internal/membership/gms/fd/GMSHealthMonitorJUnitTest.java
@@ -69,7 +69,6 @@ import org.apache.geode.distributed.internal.ClusterDistributionManager;
 import org.apache.geode.distributed.internal.DistributionConfig;
 import org.apache.geode.distributed.internal.DistributionConfigImpl;
 import org.apache.geode.distributed.internal.DistributionManager;
-import org.apache.geode.distributed.internal.DistributionStats;
 import org.apache.geode.distributed.internal.InternalDistributedSystem;
 import org.apache.geode.distributed.internal.membership.InternalDistributedMember;
 import org.apache.geode.distributed.internal.membership.NetView;
@@ -145,7 +144,8 @@ public class GMSHealthMonitorJUnitTest {
     when(services.getJoinLeave()).thenReturn(joinLeave);
     when(services.getCancelCriterion()).thenReturn(stopper);
     when(services.getManager()).thenReturn(manager);
-    when(services.getStatistics()).thenReturn(StatsFactory.createDistributionStatsImpl(system.getStatisticsFactory(), String.valueOf(statsId)));
+    when(services.getStatistics()).thenReturn(StatsFactory
+        .createDistributionStatsImpl(system.getStatisticsFactory(), String.valueOf(statsId)));
     when(stopper.isCancelInProgress()).thenReturn(false);
 
     if (mockMembers == null) {
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/distributed/internal/membership/gms/locator/GMSLocatorRecoveryJUnitTest.java b/geode-core/src/integrationTest/java/org/apache/geode/distributed/internal/membership/gms/locator/GMSLocatorRecoveryJUnitTest.java
index de3512a..1464d97 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/distributed/internal/membership/gms/locator/GMSLocatorRecoveryJUnitTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/distributed/internal/membership/gms/locator/GMSLocatorRecoveryJUnitTest.java
@@ -70,7 +70,7 @@ public class GMSLocatorRecoveryJUnitTest {
       this.tempStateFile.delete();
     }
     this.locator = new GMSLocator(null, null, false, false, StatsFactory.createLocatorStatsImpl(
-        (StatisticsFactory) new StatisticsTypeFactoryImpl(),"testLocator"), "");
+        (StatisticsFactory) new StatisticsTypeFactoryImpl(), "testLocator"), "");
     locator.setViewFile(tempStateFile);
     // System.out.println("temp state file: " + tempStateFile);
   }
@@ -180,8 +180,10 @@ public class GMSLocatorRecoveryJUnitTest {
       ((InternalLocator) l).getLocatorHandler().setMembershipManager(m1);
 
       GMSLocator l2 = new GMSLocator(SocketCreator.getLocalHost(),
-          m1.getLocalMember().getHost() + "[" + port + "]", true, true, StatsFactory.createLocatorStatsImpl(
-          (StatisticsFactory) new StatisticsTypeFactoryImpl(),"testLocator"), "");
+          m1.getLocalMember().getHost() + "[" + port + "]", true, true,
+          StatsFactory.createLocatorStatsImpl(
+              (StatisticsFactory) new StatisticsTypeFactoryImpl(), "testLocator"),
+          "");
       l2.setViewFile(new File("l2.dat"));
       l2.init(null);
 
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/distributed/internal/membership/gms/messenger/JGroupsMessengerJUnitTest.java b/geode-core/src/integrationTest/java/org/apache/geode/distributed/internal/membership/gms/messenger/JGroupsMessengerJUnitTest.java
index da8219a..68e4bf6 100755
--- a/geode-core/src/integrationTest/java/org/apache/geode/distributed/internal/membership/gms/messenger/JGroupsMessengerJUnitTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/distributed/internal/membership/gms/messenger/JGroupsMessengerJUnitTest.java
@@ -75,7 +75,6 @@ import org.apache.geode.distributed.internal.ClusterDistributionManager;
 import org.apache.geode.distributed.internal.DistributionConfigImpl;
 import org.apache.geode.distributed.internal.DistributionManager;
 import org.apache.geode.distributed.internal.DistributionMessage;
-import org.apache.geode.distributed.internal.DistributionStats;
 import org.apache.geode.distributed.internal.DistributionStatsImpl;
 import org.apache.geode.distributed.internal.InternalDistributedSystem;
 import org.apache.geode.distributed.internal.SerialAckedMessage;
@@ -169,7 +168,8 @@ public class JGroupsMessengerJUnitTest {
     DistributionManager dm = mock(DistributionManager.class);
     InternalDistributedSystem system =
         InternalDistributedSystem.newInstanceForTesting(dm, nonDefault);
-    when(services.getStatistics()).thenReturn(StatsFactory.createDistributionStatsImpl(system.getStatisticsFactory(), String.valueOf(statsId)));
+    when(services.getStatistics()).thenReturn(StatsFactory
+        .createDistributionStatsImpl(system.getStatisticsFactory(), String.valueOf(statsId)));
 
     messenger = new JGroupsMessenger();
     messenger.init(services);
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/ComplexDiskRegionJUnitTest.java b/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/ComplexDiskRegionJUnitTest.java
index 39cbb57..3d31602 100755
--- a/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/ComplexDiskRegionJUnitTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/ComplexDiskRegionJUnitTest.java
@@ -19,8 +19,8 @@ import static org.junit.Assert.fail;
 
 import org.junit.Test;
 
-import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.cache.Scope;
+import org.apache.geode.statistics.StatisticsFactory;
 
 /**
  * Unit testing for ComplexDiskRegion API's
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/DiskInitFileJUnitTest.java b/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/DiskInitFileJUnitTest.java
index 8d3ea63..8818318 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/DiskInitFileJUnitTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/DiskInitFileJUnitTest.java
@@ -30,8 +30,8 @@ import org.junit.Rule;
 import org.junit.Test;
 import org.junit.rules.TemporaryFolder;
 
-import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.internal.cache.persistence.DiskRegionView;
+import org.apache.geode.statistics.StatisticsFactory;
 
 public class DiskInitFileJUnitTest {
 
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/OplogJUnitTest.java b/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/OplogJUnitTest.java
index bc1ce9a..623a137 100755
--- a/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/OplogJUnitTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/OplogJUnitTest.java
@@ -41,7 +41,6 @@ import org.junit.Assert;
 import org.junit.Ignore;
 import org.junit.Test;
 
-import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.cache.CacheWriterException;
 import org.apache.geode.cache.DiskAccessException;
 import org.apache.geode.cache.DiskStore;
@@ -56,6 +55,7 @@ import org.apache.geode.internal.InternalDataSerializer;
 import org.apache.geode.internal.cache.Oplog.OPLOG_TYPE;
 import org.apache.geode.internal.cache.entries.AbstractDiskLRURegionEntry;
 import org.apache.geode.internal.cache.entries.DiskEntry;
+import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.test.dunit.ThreadUtils;
 
 /**
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/OplogRVVJUnitTest.java b/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/OplogRVVJUnitTest.java
index dd72767..372bc52 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/OplogRVVJUnitTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/OplogRVVJUnitTest.java
@@ -35,13 +35,13 @@ import org.junit.Rule;
 import org.junit.Test;
 import org.junit.rules.TemporaryFolder;
 
-import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.i18n.LogWriterI18n;
 import org.apache.geode.internal.cache.DiskInitFile.DiskRegionFlag;
 import org.apache.geode.internal.cache.DiskStoreImpl.OplogEntryIdSet;
 import org.apache.geode.internal.cache.persistence.DiskRecoveryStore;
 import org.apache.geode.internal.cache.persistence.DiskStoreID;
 import org.apache.geode.internal.cache.versions.DiskRegionVersionVector;
+import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatsFactory;
 
 public class OplogRVVJUnitTest {
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/PartitionedRegionStatsJUnitTest.java b/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/PartitionedRegionStatsJUnitTest.java
index ba060ea..4e9a948 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/PartitionedRegionStatsJUnitTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/PartitionedRegionStatsJUnitTest.java
@@ -34,7 +34,6 @@ import org.junit.Before;
 import org.junit.Test;
 
 import org.apache.geode.LogWriter;
-import org.apache.geode.statistics.Statistics;
 import org.apache.geode.cache.AttributesFactory;
 import org.apache.geode.cache.Cache;
 import org.apache.geode.cache.DataPolicy;
@@ -43,6 +42,7 @@ import org.apache.geode.cache.EvictionAttributes;
 import org.apache.geode.cache.PartitionAttributesFactory;
 import org.apache.geode.cache.PartitionedRegionStorageException;
 import org.apache.geode.cache.RegionExistsException;
+import org.apache.geode.statistics.Statistics;
 
 public class PartitionedRegionStatsJUnitTest {
   private static final File DISK_DIR = new File("PRStatsTest");
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/ha/HARegionQueueIntegrationTest.java b/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/ha/HARegionQueueIntegrationTest.java
index 4e55920..66e645d 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/ha/HARegionQueueIntegrationTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/ha/HARegionQueueIntegrationTest.java
@@ -55,8 +55,6 @@ import org.powermock.modules.junit4.PowerMockRunner;
 import util.TestException;
 
 import org.apache.geode.CancelCriterion;
-import org.apache.geode.statistics.Statistics;
-import org.apache.geode.statistics.StatisticsType;
 import org.apache.geode.cache.AttributesFactory;
 import org.apache.geode.cache.Cache;
 import org.apache.geode.cache.CacheFactory;
@@ -93,6 +91,8 @@ import org.apache.geode.internal.cache.tier.sockets.HAEventWrapper;
 import org.apache.geode.internal.concurrent.ConcurrentHashSet;
 import org.apache.geode.internal.util.BlobHelper;
 import org.apache.geode.internal.util.concurrent.StoppableReentrantReadWriteLock;
+import org.apache.geode.statistics.Statistics;
+import org.apache.geode.statistics.StatisticsType;
 
 @RunWith(PowerMockRunner.class)
 @PowerMockIgnore({"javax.script.*", "javax.management.*", "org.springframework.shell.event.*",
@@ -151,7 +151,8 @@ public class HARegionQueueIntegrationTest {
 
     doReturn(Mockito.mock(InternalDistributedMember.class)).when(mockInternalDistributedSystem)
         .getDistributedMember();
-    doReturn(Mockito.mock(Statistics.class)).when(mockInternalDistributedSystem.getInternalDistributedSystemStats())
+    doReturn(Mockito.mock(Statistics.class))
+        .when(mockInternalDistributedSystem.getInternalDistributedSystemStats())
         .createAtomicStatistics(any(StatisticsType.class), any(String.class));
     doReturn(Mockito.mock(DistributionConfig.class)).when(mockDistributionManager).getConfig();
     doReturn(mockDistributionManager).when(mockInternalDistributedSystem).getDistributionManager();
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/tier/sockets/CacheClientNotifierIntegrationTest.java b/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/tier/sockets/CacheClientNotifierIntegrationTest.java
index 1454a00..7c11493 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/tier/sockets/CacheClientNotifierIntegrationTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/tier/sockets/CacheClientNotifierIntegrationTest.java
@@ -53,8 +53,6 @@ import org.powermock.modules.junit4.PowerMockRunner;
 
 import org.apache.geode.CancelCriterion;
 import org.apache.geode.DataSerializer;
-import org.apache.geode.statistics.Statistics;
-import org.apache.geode.statistics.StatisticsType;
 import org.apache.geode.cache.AttributesMutator;
 import org.apache.geode.cache.Operation;
 import org.apache.geode.cache.RegionAttributes;
@@ -77,6 +75,8 @@ import org.apache.geode.internal.cache.InternalCache;
 import org.apache.geode.internal.cache.InternalRegionArguments;
 import org.apache.geode.internal.cache.LocalRegion;
 import org.apache.geode.internal.security.SecurityService;
+import org.apache.geode.statistics.Statistics;
+import org.apache.geode.statistics.StatisticsType;
 import org.apache.geode.test.junit.categories.ClientSubscriptionTest;
 
 @Category({ClientSubscriptionTest.class})
@@ -277,7 +277,8 @@ public class CacheClientNotifierIntegrationTest {
 
     doReturn(mock(InternalDistributedMember.class)).when(mockInternalDistributedSystem)
         .getDistributedMember();
-    doReturn(mock(Statistics.class)).when(mockInternalDistributedSystem.getInternalDistributedSystemStats())
+    doReturn(mock(Statistics.class))
+        .when(mockInternalDistributedSystem.getInternalDistributedSystemStats())
         .createAtomicStatistics(any(StatisticsType.class), any(String.class));
     doReturn(mock(DistributionConfig.class)).when(mockDistributionManager).getConfig();
     doReturn(mockDistributionManager).when(mockInternalDistributedSystem).getDistributionManager();
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/tier/sockets/CacheServerMaxConnectionsJUnitTest.java b/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/tier/sockets/CacheServerMaxConnectionsJUnitTest.java
index 8900108..16e5ecf 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/tier/sockets/CacheServerMaxConnectionsJUnitTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/tier/sockets/CacheServerMaxConnectionsJUnitTest.java
@@ -26,8 +26,6 @@ import org.junit.After;
 import org.junit.Test;
 import org.junit.experimental.categories.Category;
 
-import org.apache.geode.statistics.Statistics;
-import org.apache.geode.statistics.StatisticsType;
 import org.apache.geode.cache.AttributesFactory;
 import org.apache.geode.cache.Cache;
 import org.apache.geode.cache.CacheFactory;
@@ -41,6 +39,8 @@ import org.apache.geode.cache.client.internal.PoolImpl;
 import org.apache.geode.cache.server.CacheServer;
 import org.apache.geode.distributed.DistributedSystem;
 import org.apache.geode.internal.AvailablePort;
+import org.apache.geode.statistics.Statistics;
+import org.apache.geode.statistics.StatisticsType;
 import org.apache.geode.test.dunit.Wait;
 import org.apache.geode.test.dunit.WaitCriterion;
 import org.apache.geode.test.junit.categories.ClientServerTest;
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/tier/sockets/ClientHealthMonitorIntegrationTest.java b/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/tier/sockets/ClientHealthMonitorIntegrationTest.java
index a4264a9..bc3807c 100755
--- a/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/tier/sockets/ClientHealthMonitorIntegrationTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/tier/sockets/ClientHealthMonitorIntegrationTest.java
@@ -30,8 +30,6 @@ import org.junit.Test;
 import org.junit.contrib.java.lang.system.RestoreSystemProperties;
 import org.junit.experimental.categories.Category;
 
-import org.apache.geode.statistics.Statistics;
-import org.apache.geode.statistics.StatisticsType;
 import org.apache.geode.cache.AttributesFactory;
 import org.apache.geode.cache.Cache;
 import org.apache.geode.cache.CacheFactory;
@@ -47,6 +45,8 @@ import org.apache.geode.distributed.DistributedSystem;
 import org.apache.geode.internal.AvailablePort;
 import org.apache.geode.internal.cache.EventID;
 import org.apache.geode.internal.cache.InternalCache;
+import org.apache.geode.statistics.Statistics;
+import org.apache.geode.statistics.StatisticsType;
 import org.apache.geode.test.junit.categories.ClientServerTest;
 
 /**
@@ -180,7 +180,8 @@ public class ClientHealthMonitorIntegrationTest {
     PORT = createServer();
     createProxyAndRegionForClient();
     StatisticsType statisticsType = this.system.getStatisticsFactory().findType("CacheServerStats");
-    final Statistics statistics = this.system.getStatisticsFactory().findStatisticsByType(statisticsType)[0];
+    final Statistics statistics =
+        this.system.getStatisticsFactory().findStatisticsByType(statisticsType)[0];
     assertEquals(0, statistics.getInt("currentClients"));
     assertEquals(0, statistics.getInt("currentClientConnections"));
     this.system.getLogWriter()
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/tier/sockets/ServerConnectionIntegrationTest.java b/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/tier/sockets/ServerConnectionIntegrationTest.java
index ece5e09..4c5cede 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/tier/sockets/ServerConnectionIntegrationTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/internal/cache/tier/sockets/ServerConnectionIntegrationTest.java
@@ -124,9 +124,9 @@ public class ServerConnectionIntegrationTest {
      * client over a given <code>Socket</code>.
      */
     public TestServerConnection(Socket socket, InternalCache internalCache,
-                                CachedRegionHelper helper, CacheServerStatsImpl stats, int hsTimeout, int socketBufferSize,
-                                String communicationModeStr, byte communicationMode, Acceptor acceptor,
-                                SecurityService securityService) {
+        CachedRegionHelper helper, CacheServerStatsImpl stats, int hsTimeout, int socketBufferSize,
+        String communicationModeStr, byte communicationMode, Acceptor acceptor,
+        SecurityService securityService) {
       super(socket, internalCache, helper, stats, hsTimeout, socketBufferSize, communicationModeStr,
           communicationMode, acceptor, securityService);
 
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/DiskSpaceLimitIntegrationTest.java b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/DiskSpaceLimitIntegrationTest.java
index bf318de..cd422aa 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/DiskSpaceLimitIntegrationTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/DiskSpaceLimitIntegrationTest.java
@@ -39,13 +39,13 @@ import org.junit.experimental.categories.Category;
 import org.junit.rules.TemporaryFolder;
 import org.junit.rules.TestName;
 
-import org.apache.geode.statistics.StatisticDescriptor;
-import org.apache.geode.statistics.Statistics;
-import org.apache.geode.statistics.StatisticsType;
 import org.apache.geode.internal.NanoTimer;
 import org.apache.geode.internal.io.MainWithChildrenRollingFileHandler;
 import org.apache.geode.internal.io.RollingFileHandler;
 import org.apache.geode.internal.util.ArrayUtils;
+import org.apache.geode.statistics.StatisticDescriptor;
+import org.apache.geode.statistics.Statistics;
+import org.apache.geode.statistics.StatisticsType;
 import org.apache.geode.test.junit.categories.StatisticsTest;
 
 /**
@@ -129,8 +129,7 @@ public class DiskSpaceLimitIntegrationTest {
   }
 
   @After
-  public void tearDown() throws Exception {
-  }
+  public void tearDown() throws Exception {}
 
   @Test
   public void zeroKeepsAllFiles() throws Exception {
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/DistributedSystemStatisticsIntegrationTest.java b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/DistributedSystemStatisticsIntegrationTest.java
index bd4c56e..c2ae6b4 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/DistributedSystemStatisticsIntegrationTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/DistributedSystemStatisticsIntegrationTest.java
@@ -30,11 +30,11 @@ import org.junit.Test;
 import org.junit.experimental.categories.Category;
 import org.junit.rules.TestName;
 
+import org.apache.geode.distributed.DistributedSystem;
 import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.distributed.DistributedSystem;
 import org.apache.geode.test.junit.categories.StatisticsTest;
 
 /**
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/DistributedSystemStatisticsTypeIntegrationTest.java b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/DistributedSystemStatisticsTypeIntegrationTest.java
index d975d27..f87eb00 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/DistributedSystemStatisticsTypeIntegrationTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/DistributedSystemStatisticsTypeIntegrationTest.java
@@ -29,10 +29,10 @@ import org.junit.Test;
 import org.junit.experimental.categories.Category;
 import org.junit.rules.TestName;
 
+import org.apache.geode.distributed.DistributedSystem;
 import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.distributed.DistributedSystem;
 import org.apache.geode.test.junit.categories.StatisticsTest;
 
 /**
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/FileSizeLimitIntegrationTest.java b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/FileSizeLimitIntegrationTest.java
index b19edfd..44defe0 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/FileSizeLimitIntegrationTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/FileSizeLimitIntegrationTest.java
@@ -32,11 +32,11 @@ import org.junit.rules.TestName;
 import org.mockito.invocation.InvocationOnMock;
 import org.mockito.stubbing.Answer;
 
+import org.apache.geode.internal.NanoTimer;
+import org.apache.geode.internal.io.MainWithChildrenRollingFileHandler;
 import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.internal.NanoTimer;
-import org.apache.geode.internal.io.MainWithChildrenRollingFileHandler;
 import org.apache.geode.test.junit.categories.StatisticsTest;
 
 @Category({StatisticsTest.class})
@@ -110,8 +110,7 @@ public class FileSizeLimitIntegrationTest {
   }
 
   @After
-  public void tearDown() throws Exception {
-  }
+  public void tearDown() throws Exception {}
 
   @Test
   public void rollsWhenLimitIsReached() throws Exception { // TODO: add test to assert size is
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/GemFireStatSamplerIntegrationTest.java b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/GemFireStatSamplerIntegrationTest.java
index 8bbfddb..aba35c9 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/GemFireStatSamplerIntegrationTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/GemFireStatSamplerIntegrationTest.java
@@ -47,8 +47,6 @@ import org.junit.experimental.categories.Category;
 import org.junit.rules.TemporaryFolder;
 import org.junit.rules.TestName;
 
-import org.apache.geode.statistics.Statistics;
-import org.apache.geode.statistics.StatisticsType;
 import org.apache.geode.distributed.DistributedSystem;
 import org.apache.geode.distributed.internal.InternalDistributedSystem;
 import org.apache.geode.internal.GemFireVersion;
@@ -61,6 +59,8 @@ import org.apache.geode.internal.statistics.platform.OsStatisticsFactory;
 import org.apache.geode.internal.statistics.platform.ProcessStats;
 import org.apache.geode.internal.stats50.VMStats50;
 import org.apache.geode.internal.util.StopWatch;
+import org.apache.geode.statistics.Statistics;
+import org.apache.geode.statistics.StatisticsType;
 import org.apache.geode.test.junit.categories.StatisticsTest;
 
 /**
@@ -532,15 +532,16 @@ public class GemFireStatSamplerIntegrationTest extends StatSamplerTestCase {
 
   @Override
   protected StatisticsManager getStatisticsManager() {
-    return ((InternalDistributedSystem)this.distributedSystem).getInternalDistributedSystemStats();
+    return ((InternalDistributedSystem) this.distributedSystem).getInternalDistributedSystemStats();
   }
 
   protected OsStatisticsFactory getOsStatisticsFactory() {
-    return ((InternalDistributedSystem)this.distributedSystem).getInternalDistributedSystemStats();
+    return ((InternalDistributedSystem) this.distributedSystem).getInternalDistributedSystemStats();
   }
 
   private GemFireStatSampler getGemFireStatSampler() {
-    return ((InternalDistributedSystem) this.distributedSystem).getInternalDistributedSystemStats().getStatSampler();
+    return ((InternalDistributedSystem) this.distributedSystem).getInternalDistributedSystemStats()
+        .getStatSampler();
   }
 
   private SampleCollector getSampleCollector() {
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/LinuxSystemStatsTest.java b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/LinuxSystemStatsTest.java
index 714174d..cfd0966 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/LinuxSystemStatsTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/LinuxSystemStatsTest.java
@@ -1,5 +1,5 @@
 /*
-/*
+ * /*
  * Licensed to the Apache Software Foundation (ASF) under one or more contributor license
  * agreements. See the NOTICE file distributed with this work for additional information regarding
  * copyright ownership. The ASF licenses this file to You under the Apache License, Version 2.0 (the
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/SimpleStatSamplerIntegrationTest.java b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/SimpleStatSamplerIntegrationTest.java
index 5d61475..62e5a59 100755
--- a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/SimpleStatSamplerIntegrationTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/SimpleStatSamplerIntegrationTest.java
@@ -34,10 +34,10 @@ import org.junit.rules.TemporaryFolder;
 import org.junit.rules.TestName;
 
 import org.apache.geode.CancelCriterion;
-import org.apache.geode.statistics.Statistics;
-import org.apache.geode.statistics.StatisticsType;
 import org.apache.geode.internal.net.SocketCreator;
 import org.apache.geode.internal.stats50.VMStats50;
+import org.apache.geode.statistics.Statistics;
+import org.apache.geode.statistics.StatisticsType;
 import org.apache.geode.test.junit.categories.StatisticsTest;
 
 /**
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/StatArchiveWithConsecutiveResourceInstGenerator.java b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/StatArchiveWithConsecutiveResourceInstGenerator.java
index a420422..498fd9f 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/StatArchiveWithConsecutiveResourceInstGenerator.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/StatArchiveWithConsecutiveResourceInstGenerator.java
@@ -38,11 +38,11 @@ import org.junit.Test;
 import org.junit.rules.TemporaryFolder;
 import org.junit.rules.TestName;
 
+import org.apache.geode.internal.statistics.StatArchiveReader.ResourceInst;
+import org.apache.geode.internal.statistics.StatArchiveReader.StatValue;
 import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.internal.statistics.StatArchiveReader.ResourceInst;
-import org.apache.geode.internal.statistics.StatArchiveReader.StatValue;
 
 /**
  * Generates the stat archive file that is committed under src/test/resources for
@@ -101,7 +101,7 @@ public class StatArchiveWithConsecutiveResourceInstGenerator {
 
   @After
   public void tearDown() throws Exception {
-//    StatisticsTypeFactoryImpl.clear();
+    // StatisticsTypeFactoryImpl.clear();
   }
 
   @Test
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/StatArchiveWithMissingResourceTypeRegressionTest.java b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/StatArchiveWithMissingResourceTypeRegressionTest.java
index 016c87c..70756db 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/StatArchiveWithMissingResourceTypeRegressionTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/StatArchiveWithMissingResourceTypeRegressionTest.java
@@ -64,8 +64,7 @@ public class StatArchiveWithMissingResourceTypeRegressionTest {
   }
 
   @After
-  public void tearDown() throws Exception {
-  }
+  public void tearDown() throws Exception {}
 
   @Test
   public void throwsIllegalStateExceptionWithMessage() throws Exception {
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/StatArchiveWriterReaderIntegrationTest.java b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/StatArchiveWriterReaderIntegrationTest.java
index 762f4fe..6d09d9b 100755
--- a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/StatArchiveWriterReaderIntegrationTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/StatArchiveWriterReaderIntegrationTest.java
@@ -48,11 +48,11 @@ import org.junit.experimental.categories.Category;
 import org.junit.rules.TemporaryFolder;
 import org.junit.rules.TestName;
 
+import org.apache.geode.internal.NanoTimer;
+import org.apache.geode.internal.statistics.StatArchiveReader.StatValue;
 import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.internal.NanoTimer;
-import org.apache.geode.internal.statistics.StatArchiveReader.StatValue;
 import org.apache.geode.test.junit.categories.StatisticsTest;
 import org.apache.geode.util.test.TestUtil;
 
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/StatSamplerIntegrationTest.java b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/StatSamplerIntegrationTest.java
index c11b7d3..94ad23f 100755
--- a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/StatSamplerIntegrationTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/StatSamplerIntegrationTest.java
@@ -38,11 +38,11 @@ import org.junit.rules.TemporaryFolder;
 import org.junit.rules.TestName;
 
 import org.apache.geode.CancelCriterion;
+import org.apache.geode.internal.logging.LogService;
+import org.apache.geode.internal.statistics.StatArchiveReader.StatValue;
 import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.internal.logging.LogService;
-import org.apache.geode.internal.statistics.StatArchiveReader.StatValue;
 import org.apache.geode.test.junit.categories.StatisticsTest;
 
 /**
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/StatTypesAreRolledOverRegressionTest.java b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/StatTypesAreRolledOverRegressionTest.java
index 62c73f9..60531e2 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/StatTypesAreRolledOverRegressionTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/StatTypesAreRolledOverRegressionTest.java
@@ -33,12 +33,12 @@ import org.junit.rules.TemporaryFolder;
 import org.junit.rules.TestName;
 import org.mockito.stubbing.Answer;
 
-import org.apache.geode.statistics.StatisticDescriptor;
-import org.apache.geode.statistics.Statistics;
-import org.apache.geode.statistics.StatisticsType;
 import org.apache.geode.internal.NanoTimer;
 import org.apache.geode.internal.io.MainWithChildrenRollingFileHandler;
 import org.apache.geode.internal.statistics.StatArchiveReader.ResourceInst;
+import org.apache.geode.statistics.StatisticDescriptor;
+import org.apache.geode.statistics.Statistics;
+import org.apache.geode.statistics.StatisticsType;
 import org.apache.geode.test.junit.categories.StatisticsTest;
 
 @Category({StatisticsTest.class})
@@ -104,8 +104,7 @@ public class StatTypesAreRolledOverRegressionTest {
   }
 
   @After
-  public void tearDown() throws Exception {
-  }
+  public void tearDown() throws Exception {}
 
   @Test
   public void closedInstanceShouldHaveTypeInRolledArchives() throws Exception {
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/ValueMonitorIntegrationTest.java b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/ValueMonitorIntegrationTest.java
index 75fbb01..5f114b6 100755
--- a/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/ValueMonitorIntegrationTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/internal/statistics/ValueMonitorIntegrationTest.java
@@ -37,12 +37,12 @@ import org.junit.Test;
 import org.junit.experimental.categories.Category;
 import org.junit.rules.TestName;
 
-import org.apache.geode.statistics.StatisticDescriptor;
-import org.apache.geode.statistics.Statistics;
-import org.apache.geode.statistics.StatisticsType;
 import org.apache.geode.internal.NanoTimer;
 import org.apache.geode.internal.io.MainWithChildrenRollingFileHandler;
 import org.apache.geode.internal.statistics.StatisticsNotification.Type;
+import org.apache.geode.statistics.StatisticDescriptor;
+import org.apache.geode.statistics.Statistics;
+import org.apache.geode.statistics.StatisticsType;
 import org.apache.geode.test.dunit.WaitCriterion;
 import org.apache.geode.test.junit.categories.StatisticsTest;
 
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/internal/stats50/AtomicStatsJUnitTest.java b/geode-core/src/integrationTest/java/org/apache/geode/internal/stats50/AtomicStatsJUnitTest.java
index 9e2ce45..ef659b8 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/internal/stats50/AtomicStatsJUnitTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/internal/stats50/AtomicStatsJUnitTest.java
@@ -25,12 +25,12 @@ import java.util.concurrent.atomic.AtomicReference;
 
 import org.junit.Test;
 
+import org.apache.geode.distributed.DistributedSystem;
+import org.apache.geode.internal.statistics.StatisticsTypeFactoryImpl;
 import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsType;
 import org.apache.geode.statistics.StatisticsTypeFactory;
-import org.apache.geode.distributed.DistributedSystem;
-import org.apache.geode.internal.statistics.StatisticsTypeFactoryImpl;
 
 public class AtomicStatsJUnitTest {
 
@@ -102,7 +102,8 @@ public class AtomicStatsJUnitTest {
       thread1.start();
       thread3.start();
       for (int i = 0; i < 5000; i++) {
-        Statistics stats = distributedSystem.getStatisticsFactory().createAtomicStatistics(type, "stats");
+        Statistics stats =
+            distributedSystem.getStatisticsFactory().createAtomicStatistics(type, "stats");
         statsRef.set(stats);
         beforeIncrement.await();
         afterIncrement.await();
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/management/bean/stats/AsyncEventQueueStatsJUnitTest.java b/geode-core/src/integrationTest/java/org/apache/geode/management/bean/stats/AsyncEventQueueStatsJUnitTest.java
index 58cc3ca..90cfb2c 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/management/bean/stats/AsyncEventQueueStatsJUnitTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/management/bean/stats/AsyncEventQueueStatsJUnitTest.java
@@ -19,9 +19,9 @@ import static org.junit.Assert.assertEquals;
 import org.junit.Test;
 import org.junit.experimental.categories.Category;
 
-import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.cache.asyncqueue.internal.AsyncEventQueueStats;
 import org.apache.geode.management.internal.beans.AsyncEventQueueMBeanBridge;
+import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.test.junit.categories.JMXTest;
 
 @Category({JMXTest.class})
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/management/bean/stats/CacheServerStatsJUnitTest.java b/geode-core/src/integrationTest/java/org/apache/geode/management/bean/stats/CacheServerStatsJUnitTest.java
index ee18ae5..a55b513 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/management/bean/stats/CacheServerStatsJUnitTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/management/bean/stats/CacheServerStatsJUnitTest.java
@@ -58,7 +58,8 @@ public class CacheServerStatsJUnitTest extends MBeanStatsTestCase {
     cacheServerStats.incProcessPutTime(startTime);
 
     ServerLoad load = new ServerLoad(1, 1, 1, 1);
-    cacheServerStats.setLoad(load.getConnectionLoad(),load.getLoadPerConnection(),load.getSubscriptionConnectionLoad(),load.getLoadPerSubscriptionConnection());
+    cacheServerStats.setLoad(load.getConnectionLoad(), load.getLoadPerConnection(),
+        load.getSubscriptionConnectionLoad(), load.getLoadPerSubscriptionConnection());
 
     sample();
 
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/management/bean/stats/DistributedSystemStatsJUnitTest.java b/geode-core/src/integrationTest/java/org/apache/geode/management/bean/stats/DistributedSystemStatsJUnitTest.java
index 66f3e6f..5057045 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/management/bean/stats/DistributedSystemStatsJUnitTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/management/bean/stats/DistributedSystemStatsJUnitTest.java
@@ -76,7 +76,8 @@ public class DistributedSystemStatsJUnitTest {
 
     this.system = (InternalDistributedSystem) DistributedSystem.connect(props);
     assertNotNull(this.system.getInternalDistributedSystemStats().getStatSampler());
-    assertNotNull(this.system.getInternalDistributedSystemStats().getStatSampler().waitForSampleCollector(TIMEOUT));
+    assertNotNull(this.system.getInternalDistributedSystemStats().getStatSampler()
+        .waitForSampleCollector(TIMEOUT));
 
     this.cache = new CacheFactory().create();
 
@@ -120,7 +121,8 @@ public class DistributedSystemStatsJUnitTest {
   }
 
   protected void sample() throws InterruptedException {
-    this.system.getInternalDistributedSystemStats().getStatSampler().getSampleCollector().sample(NanoTimer.getTime());
+    this.system.getInternalDistributedSystemStats().getStatSampler().getSampleCollector()
+        .sample(NanoTimer.getTime());
     Thread.sleep(SLEEP);
   }
 
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/management/bean/stats/GatewayReceiverStatsJUnitTest.java b/geode-core/src/integrationTest/java/org/apache/geode/management/bean/stats/GatewayReceiverStatsJUnitTest.java
index 348d492..642f56d 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/management/bean/stats/GatewayReceiverStatsJUnitTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/management/bean/stats/GatewayReceiverStatsJUnitTest.java
@@ -34,7 +34,7 @@ public class GatewayReceiverStatsJUnitTest extends MBeanStatsTestCase {
   private GatewayReceiverStats receiverStats;
 
   public void init() {
-    receiverStats = StatsFactory.createGatewayReceiverStatsImpl(null,"Test Sock Name");
+    receiverStats = StatsFactory.createGatewayReceiverStatsImpl(null, "Test Sock Name");
 
     bridge = new GatewayReceiverMBeanBridge();
     bridge.addGatewayReceiverStats(receiverStats);
@@ -59,7 +59,8 @@ public class GatewayReceiverStatsJUnitTest extends MBeanStatsTestCase {
     receiverStats.incProcessPutTime(startTime);
 
     ServerLoad load = new ServerLoad(1, 1, 1, 1);
-    receiverStats.setLoad(load.getConnectionLoad(),load.getLoadPerConnection(),load.getSubscriptionConnectionLoad(),load.getLoadPerSubscriptionConnection());
+    receiverStats.setLoad(load.getConnectionLoad(), load.getLoadPerConnection(),
+        load.getSubscriptionConnectionLoad(), load.getLoadPerSubscriptionConnection());
 
     sample();
 
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/management/bean/stats/MemberLevelStatsJUnitTest.java b/geode-core/src/integrationTest/java/org/apache/geode/management/bean/stats/MemberLevelStatsJUnitTest.java
index 5441d35..4905a35 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/management/bean/stats/MemberLevelStatsJUnitTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/management/bean/stats/MemberLevelStatsJUnitTest.java
@@ -67,10 +67,11 @@ public class MemberLevelStatsJUnitTest extends MBeanStatsTestCase {
   private static long testStartTime = NanoTimer.getTime();
 
   public void init() {
-    cachePerfStats = StatsFactory.createCachePerfStatsImpl(system.getStatisticsFactory(),null);
+    cachePerfStats = StatsFactory.createCachePerfStatsImpl(system.getStatisticsFactory(), null);
     funcServiceStats = new FunctionServiceStats(system.getStatisticsFactory(), "FunctionExecution");
     long statId = OSProcess.getId();
-    distributionStats = StatsFactory.createDistributionStatsImpl(system.getStatisticsFactory(), String.valueOf(statId));
+    distributionStats = StatsFactory.createDistributionStatsImpl(system.getStatisticsFactory(),
+        String.valueOf(statId));
     DistributionStatsImpl.enableClockStats = true;
     dlockStats = StatsFactory.createDLockStatsImpl(system.getStatisticsFactory(), statId);
 
@@ -81,20 +82,23 @@ public class MemberLevelStatsJUnitTest extends MBeanStatsTestCase {
     bridge.addLockServiceStats(dlockStats);
 
 
-    VMStatsContract vmstats = system.getInternalDistributedSystemStats().getStatSampler().getVMStats();
+    VMStatsContract vmstats =
+        system.getInternalDistributedSystemStats().getStatSampler().getVMStats();
     assertTrue(vmstats instanceof VMStats50);
 
     bridge.addSystemStats();
     bridge.addVMStats();
 
     for (int i = 0; i < 4; i++) {
-      DiskStoreStats stats = StatsFactory.createDiskStoreStatsImpl(system.getStatisticsFactory(), name.getMethodName() + i);
+      DiskStoreStats stats = StatsFactory.createDiskStoreStatsImpl(system.getStatisticsFactory(),
+          name.getMethodName() + i);
       diskStatsList.add(stats);
       bridge.addDiskStoreStats(stats);
     }
 
     for (int i = 0; i < 4; i++) {
-      PartitionedRegionStats stats = StatsFactory.createPartitionedRegionStatsImpl(system.getStatisticsFactory(), name.getMethodName() + i);
+      PartitionedRegionStats stats = StatsFactory.createPartitionedRegionStatsImpl(
+          system.getStatisticsFactory(), name.getMethodName() + i);
       parRegionStatsList.add(stats);
       bridge.addPartionRegionStats(stats);
     }
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/management/bean/stats/RegionStatsJUnitTest.java b/geode-core/src/integrationTest/java/org/apache/geode/management/bean/stats/RegionStatsJUnitTest.java
index 1a52f8e..e1a46f7 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/management/bean/stats/RegionStatsJUnitTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/management/bean/stats/RegionStatsJUnitTest.java
@@ -49,9 +49,11 @@ public class RegionStatsJUnitTest extends MBeanStatsTestCase {
   private DiskRegionStats diskRegionStats;
 
   protected void init() {
-    cachePerfStats = StatsFactory.createCachePerfStatsImpl(system.getStatisticsFactory(),null);
-    partitionedRegionStats = StatsFactory.createPartitionedRegionStatsImpl(system.getStatisticsFactory(), "/tests");
-    diskRegionStats = StatsFactory.createDiskRegionStatsImpl(system.getStatisticsFactory(), "test-disk");
+    cachePerfStats = StatsFactory.createCachePerfStatsImpl(system.getStatisticsFactory(), null);
+    partitionedRegionStats =
+        StatsFactory.createPartitionedRegionStatsImpl(system.getStatisticsFactory(), "/tests");
+    diskRegionStats =
+        StatsFactory.createDiskRegionStatsImpl(system.getStatisticsFactory(), "test-disk");
 
     bridge = new RegionMBeanBridge(cachePerfStats);
     parBridge = new PartitionedRegionBridge(partitionedRegionStats);
diff --git a/geode-core/src/integrationTest/java/org/apache/geode/management/internal/beans/ManagementAdapterTest.java b/geode-core/src/integrationTest/java/org/apache/geode/management/internal/beans/ManagementAdapterTest.java
index 6779de6..10bdb23 100644
--- a/geode-core/src/integrationTest/java/org/apache/geode/management/internal/beans/ManagementAdapterTest.java
+++ b/geode-core/src/integrationTest/java/org/apache/geode/management/internal/beans/ManagementAdapterTest.java
@@ -31,7 +31,6 @@ import org.junit.Test;
 import org.apache.geode.distributed.internal.InternalDistributedSystem;
 import org.apache.geode.distributed.internal.ResourceEvent;
 import org.apache.geode.internal.cache.DiskStoreImpl;
-import org.apache.geode.internal.cache.DiskStoreStats;
 import org.apache.geode.internal.cache.InternalCache;
 import org.apache.geode.statistics.StatsFactory;
 import org.apache.geode.test.junit.rules.ServerStarterRule;
@@ -52,8 +51,9 @@ public class ManagementAdapterTest {
   @Before
   public void before() {
     cache = serverRule.getCache();
-    doReturn(StatsFactory.createDiskStoreStatsImpl(cache.getInternalDistributedSystem().getStatisticsFactory(), "disk-stats")).when(diskStore)
-        .getStats();
+    doReturn(StatsFactory.createDiskStoreStatsImpl(
+        cache.getInternalDistributedSystem().getStatisticsFactory(), "disk-stats")).when(diskStore)
+            .getStats();
     doReturn(new File[] {}).when(diskStore).getDiskDirs();
   }
 
diff --git a/geode-core/src/main/java/org/apache/geode/cache/asyncqueue/internal/AsyncEventQueueStats.java b/geode-core/src/main/java/org/apache/geode/cache/asyncqueue/internal/AsyncEventQueueStats.java
index b16012a..052b1c8 100644
--- a/geode-core/src/main/java/org/apache/geode/cache/asyncqueue/internal/AsyncEventQueueStats.java
+++ b/geode-core/src/main/java/org/apache/geode/cache/asyncqueue/internal/AsyncEventQueueStats.java
@@ -14,7 +14,6 @@
  */
 package org.apache.geode.cache.asyncqueue.internal;
 
-import org.apache.geode.internal.cache.wan.GatewaySenderStats;
 import org.apache.geode.internal.cache.wan.GatewaySenderStatsImpl;
 import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.StatisticsFactory;
@@ -47,7 +46,7 @@ public class AsyncEventQueueStats extends GatewaySenderStatsImpl {
   @Override
   public void initializeStats(StatisticsFactory factory) {
     type = factory.createType(typeName, "Stats for activity in the AsyncEventQueue",
-        new StatisticDescriptor[]{
+        new StatisticDescriptor[] {
             factory.createIntCounter(EVENTS_RECEIVED, "Number of events received by this queue.",
                 "operations"),
             factory.createIntCounter(EVENTS_QUEUED, "Number of events added to the event queue.",
@@ -161,11 +160,12 @@ public class AsyncEventQueueStats extends GatewaySenderStatsImpl {
 
   /**
    * Constructor.
+   *
    * @param factory The <code>StatisticsFactory</code> which creates the <code>Statistics</code>
-   * instance
+   *        instance
    * @param asyncQueueId The id of the <code>AsyncEventQueue</code> used to generate the name of
-   * the
-   * <code>Statistics</code>
+   *        the
+   *        <code>Statistics</code>
    */
   public AsyncEventQueueStats(StatisticsFactory factory, String asyncQueueId) {
     super();
diff --git a/geode-core/src/main/java/org/apache/geode/cache/asyncqueue/internal/ParallelAsyncEventQueueImpl.java b/geode-core/src/main/java/org/apache/geode/cache/asyncqueue/internal/ParallelAsyncEventQueueImpl.java
index 3c2cdba..42e3c5f 100644
--- a/geode-core/src/main/java/org/apache/geode/cache/asyncqueue/internal/ParallelAsyncEventQueueImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/cache/asyncqueue/internal/ParallelAsyncEventQueueImpl.java
@@ -50,8 +50,9 @@ public class ParallelAsyncEventQueueImpl extends AbstractGatewaySender {
     if (!(this.cache instanceof CacheCreation)) {
       // this sender lies underneath the AsyncEventQueue. Need to have
       // AsyncEventQueueStats
-      this.statistics = new AsyncEventQueueStats(cache.getDistributedSystem().getStatisticsFactory(),
-          AsyncEventQueueImpl.getAsyncEventQueueIdFromSenderId(id));
+      this.statistics =
+          new AsyncEventQueueStats(cache.getDistributedSystem().getStatisticsFactory(),
+              AsyncEventQueueImpl.getAsyncEventQueueIdFromSenderId(id));
     }
     this.isForInternalUse = true;
   }
diff --git a/geode-core/src/main/java/org/apache/geode/cache/asyncqueue/internal/SerialAsyncEventQueueImpl.java b/geode-core/src/main/java/org/apache/geode/cache/asyncqueue/internal/SerialAsyncEventQueueImpl.java
index 453e532..a076d78 100644
--- a/geode-core/src/main/java/org/apache/geode/cache/asyncqueue/internal/SerialAsyncEventQueueImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/cache/asyncqueue/internal/SerialAsyncEventQueueImpl.java
@@ -52,8 +52,9 @@ public class SerialAsyncEventQueueImpl extends AbstractGatewaySender {
     if (!(this.cache instanceof CacheCreation)) {
       // this sender lies underneath the AsyncEventQueue. Need to have
       // AsyncEventQueueStats
-      this.statistics = new AsyncEventQueueStats(cache.getDistributedSystem().getStatisticsFactory(),
-          AsyncEventQueueImpl.getAsyncEventQueueIdFromSenderId(id));
+      this.statistics =
+          new AsyncEventQueueStats(cache.getDistributedSystem().getStatisticsFactory(),
+              AsyncEventQueueImpl.getAsyncEventQueueIdFromSenderId(id));
     }
   }
 
diff --git a/geode-core/src/main/java/org/apache/geode/cache/client/internal/ConnectionStatsImpl.java b/geode-core/src/main/java/org/apache/geode/cache/client/internal/ConnectionStatsImpl.java
index 1755d19c2..34ef018 100644
--- a/geode-core/src/main/java/org/apache/geode/cache/client/internal/ConnectionStatsImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/cache/client/internal/ConnectionStatsImpl.java
@@ -14,15 +14,16 @@
  */
 package org.apache.geode.cache.client.internal;
 
+import org.apache.geode.internal.cache.PoolStatsImpl;
 import org.apache.geode.statistics.GFSStatsImplementor;
 import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.internal.cache.PoolStatsImpl;
 
 /**
  * Stats for a client to server {@link Connection}
+ *
  * @since GemFire 5.7
  */
 public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor {
@@ -440,7 +441,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
   public void initializeStats(StatisticsFactory factory) {
     try {
       type = factory.createType("ClientStats", "Statistics about client to server communication",
-          new StatisticDescriptor[]{
+          new StatisticDescriptor[] {
               factory.createIntGauge("getsInProgress", "Current number of gets being executed",
                   "gets"),
               factory.createIntCounter("gets", "Total number of gets completed successfully",
@@ -944,13 +945,15 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
       sendType =
           factory.createType("ClientSendStats", "Statistics about client to server communication",
-              new StatisticDescriptor[]{
+              new StatisticDescriptor[] {
                   ///////////////////////////////////////////////////////////////////////
                   /*
-                   * factory.createIntGauge("opSendsInProgress", "Current number of op sends being executed",
+                   * factory.createIntGauge("opSendsInProgress",
+                   * "Current number of op sends being executed",
                    * "sends"), factory.createIntCounter("opSends",
                    * "Total number of op sends that have completed successfully", "sends"),
-                   * factory.createIntCounter("opSendFailures", "Total number of op sends that have failed",
+                   * factory.createIntCounter("opSendFailures",
+                   * "Total number of op sends that have failed",
                    * "sends"), factory.createLongCounter("opSendTime",
                    * "Total amount of time, in nanoseconds spent doing op sends", "nanoseconds"),
                    */
@@ -1733,7 +1736,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
       addPdxTypeId = type.nameToId("addPdxTypeSuccessful");
       addPdxTypeDurationId = type.nameToId("addPdxTypeTime");
 
-      opIds = new int[]{getId, putId, destroyId, destroyRegionId, clearId, containsKeyId, keySetId,
+      opIds = new int[] {getId, putId, destroyId, destroyRegionId, clearId, containsKeyId, keySetId,
           registerInterestId, unregisterInterestId, queryId, createCQId, stopCQId, closeCQId,
           gatewayBatchId, readyForEventsId, makePrimaryId, closeConId, primaryAckId, pingId,
           putAllId, removeAllId, getAllId, registerInstantiatorsId, executeFunctionId,
@@ -1751,7 +1754,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
   private final PoolStatsImpl poolStats;
 
   public ConnectionStatsImpl(StatisticsFactory factory, String name,
-                             PoolStatsImpl poolStats/* , GatewayStats gatewayStats */) {
+      PoolStatsImpl poolStats/* , GatewayStats gatewayStats */) {
     initializeStats(factory);
     this.stats = factory.createAtomicStatistics(type, "ClientStats-" + name);
     this.sendStats = factory.createAtomicStatistics(sendType, "ClientSendStats-" + name);
@@ -1763,6 +1766,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link #endGetSend} and {@link
    * #endGet}.
+   *
    * @return the start time of this get
    */
   @Override
@@ -1775,6 +1779,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the get has completed
+   *
    * @param startTime the value returned by {@link #startGet}.
    * @param failed true if the send of the get failed
    */
@@ -1785,13 +1790,14 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
   }
 
   private void endOperation(long startTime, boolean failed, int operationInProgressId,
-                            int operationFailedId, int operationId, int operationDurationId) {
+      int operationFailedId, int operationId, int operationDurationId) {
     endOperationWithTimeout(startTime, false, failed, operationInProgressId, 0, operationFailedId,
         operationId, operationDurationId);
   }
 
   /**
    * Records that the specified get has ended
+   *
    * @param startTime the value returned by {@link #startGet}.
    * @param timedOut true if get timed out
    * @param failed true if get failed
@@ -1818,6 +1824,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link #endPutSend} and {@link
    * #endPut}.
+   *
    * @return the start time of this put
    */
   @Override
@@ -1830,6 +1837,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the put has completed
+   *
    * @param startTime the value returned by {@link #startPut}.
    * @param failed true if the send of the put failed
    */
@@ -1841,6 +1849,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified put has ended
+   *
    * @param startTime the value returned by {@link #startPut}.
    * @param timedOut true if put timed out
    * @param failed true if put failed
@@ -1853,9 +1862,9 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
   }
 
   public void endOperationWithTimeout(long startTime, boolean timedOut, boolean failed,
-                                      int operationInProgressId, int operationTimedOutId,
-                                      int operationFailedId, int operationId,
-                                      int operationDurationId) {
+      int operationInProgressId, int operationTimedOutId,
+      int operationFailedId, int operationId,
+      int operationDurationId) {
     long duration = System.nanoTime() - startTime;
     endClientOp(duration, timedOut, failed);
     this.stats.incInt(operationInProgressId, -1);
@@ -1886,6 +1895,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link #endDestroySend} and
    * {@link #endDestroy}.
+   *
    * @return the start time of this destroy
    */
   @Override
@@ -1898,6 +1908,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the destroy has completed
+   *
    * @param startTime the value returned by {@link #startDestroy}.
    * @param failed true if the send of the destroy failed
    */
@@ -1909,6 +1920,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified destroy has ended
+   *
    * @param startTime the value returned by {@link #startDestroy}.
    * @param timedOut true if destroy timed out
    * @param failed true if destroy failed
@@ -1924,6 +1936,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link #endDestroyRegionSend} and
    * {@link #endDestroyRegion}.
+   *
    * @return the start time of this destroyRegion
    */
   @Override
@@ -1936,6 +1949,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the destroyRegion has completed
+   *
    * @param startTime the value returned by {@link #startDestroyRegion}.
    * @param failed true if the send of the destroyRegion failed
    */
@@ -1947,6 +1961,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified destroyRegion has ended
+   *
    * @param startTime the value returned by {@link #startDestroyRegion}.
    * @param timedOut true if destroyRegion timed out
    * @param failed true if destroyRegion failed
@@ -1963,6 +1978,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link #endClearSend} and {@link
    * #endClear}.
+   *
    * @return the start time of this clear
    */
   @Override
@@ -1975,6 +1991,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the clear has completed
+   *
    * @param startTime the value returned by {@link #startClear}.
    * @param failed true if the send of the clear failed
    */
@@ -1986,6 +2003,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified clear has ended
+   *
    * @param startTime the value returned by {@link #startClear}.
    * @param timedOut true if clear timed out
    * @param failed true if clear failed
@@ -2002,6 +2020,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link #endContainsKeySend} and
    * {@link #endContainsKey}.
+   *
    * @return the start time of this containsKey
    */
   @Override
@@ -2014,6 +2033,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the containsKey has completed
+   *
    * @param startTime the value returned by {@link #startContainsKey}.
    * @param failed true if the send of the containsKey failed
    */
@@ -2025,6 +2045,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified containsKey has ended
+   *
    * @param startTime the value returned by {@link #startContainsKey}.
    * @param timedOut true if containsKey timed out
    * @param failed true if containsKey failed
@@ -2041,6 +2062,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link #endKeySetSend} and {@link
    * #endKeySet}.
+   *
    * @return the start time of this keySet
    */
   @Override
@@ -2053,6 +2075,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the keySet has completed
+   *
    * @param startTime the value returned by {@link #startKeySet}.
    * @param failed true if the send of the keySet failed
    */
@@ -2064,6 +2087,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified keySet has ended
+   *
    * @param startTime the value returned by {@link #startKeySet}.
    * @param timedOut true if keySet timed out
    * @param failed true if keySet failed
@@ -2080,6 +2104,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link #endRegisterInterestSend}
    * and {@link #endRegisterInterest}.
+   *
    * @return the start time of this registerInterest
    */
   @Override
@@ -2092,6 +2117,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the registerInterest has completed
+   *
    * @param startTime the value returned by {@link #startRegisterInterest}.
    * @param failed true if the send of the registerInterest failed
    */
@@ -2103,6 +2129,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified registerInterest has ended
+   *
    * @param startTime the value returned by {@link #startRegisterInterest}.
    * @param timedOut true if registerInterest timed out
    * @param failed true if registerInterest failed
@@ -2119,6 +2146,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link
    * #endUnregisterInterestSend} and {@link #endUnregisterInterest}.
+   *
    * @return the start time of this unregisterInterest
    */
   @Override
@@ -2131,6 +2159,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the unregisterInterest has completed
+   *
    * @param startTime the value returned by {@link #startUnregisterInterest}.
    * @param failed true if the send of the unregisterInterest failed
    */
@@ -2142,6 +2171,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified unregisterInterest has ended
+   *
    * @param startTime the value returned by {@link #startUnregisterInterest}.
    * @param timedOut true if unregisterInterest timed out
    * @param failed true if unregisterInterest failed
@@ -2158,6 +2188,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link #endQuerySend} and {@link
    * #endQuery}.
+   *
    * @return the start time of this query
    */
   @Override
@@ -2170,6 +2201,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the query has completed
+   *
    * @param startTime the value returned by {@link #startQuery}.
    * @param failed true if the send of the query failed
    */
@@ -2181,6 +2213,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified query has ended
+   *
    * @param startTime the value returned by {@link #startQuery}.
    * @param timedOut true if query timed out
    * @param failed true if query failed
@@ -2197,6 +2230,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link #endCreateCQSend} and
    * {@link #endCreateCQ}.
+   *
    * @return the start time of this createCQ
    */
   @Override
@@ -2209,6 +2243,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the createCQ has completed
+   *
    * @param startTime the value returned by {@link #startCreateCQ}.
    * @param failed true if the send of the createCQ failed
    */
@@ -2220,6 +2255,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified createCQ has ended
+   *
    * @param startTime the value returned by {@link #startCreateCQ}.
    * @param timedOut true if createCQ timed out
    * @param failed true if createCQ failed
@@ -2235,6 +2271,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link #endStopCQSend} and {@link
    * #endStopCQ}.
+   *
    * @return the start time of this stopCQ
    */
   @Override
@@ -2247,6 +2284,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the stopCQ has completed
+   *
    * @param startTime the value returned by {@link #startStopCQ}.
    * @param failed true if the send of the stopCQ failed
    */
@@ -2258,6 +2296,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified stopCQ has ended
+   *
    * @param startTime the value returned by {@link #startStopCQ}.
    * @param timedOut true if stopCQ timed out
    * @param failed true if stopCQ failed
@@ -2274,6 +2313,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link #endCloseCQSend} and
    * {@link #endCloseCQ}.
+   *
    * @return the start time of this closeCQ
    */
   @Override
@@ -2286,6 +2326,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the closeCQ has completed
+   *
    * @param startTime the value returned by {@link #startCloseCQ}.
    * @param failed true if the send of the closeCQ failed
    */
@@ -2297,6 +2338,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified closeCQ has ended
+   *
    * @param startTime the value returned by {@link #startCloseCQ}.
    * @param timedOut true if closeCQ timed out
    * @param failed true if closeCQ failed
@@ -2312,6 +2354,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link #endStopCQSend} and {@link
    * #endStopCQ}.
+   *
    * @return the start time of this stopCQ
    */
   @Override
@@ -2324,6 +2367,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the stopCQ has completed
+   *
    * @param startTime the value returned by {@link #startStopCQ}.
    * @param failed true if the send of the stopCQ failed
    */
@@ -2335,6 +2379,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified stopCQ has ended
+   *
    * @param startTime the value returned by {@link #startStopCQ}.
    * @param timedOut true if stopCQ timed out
    * @param failed true if stopCQ failed
@@ -2351,6 +2396,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link #endGatewayBatchSend} and
    * {@link #endGatewayBatch}.
+   *
    * @return the start time of this gatewayBatch
    */
   @Override
@@ -2363,6 +2409,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the gatewayBatch has completed
+   *
    * @param startTime the value returned by {@link #startGatewayBatch}.
    * @param failed true if the send of the gatewayBatch failed
    */
@@ -2374,6 +2421,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified gatewayBatch has ended
+   *
    * @param startTime the value returned by {@link #startGatewayBatch}.
    * @param timedOut true if gatewayBatch timed out
    * @param failed true if gatewayBatch failed
@@ -2390,6 +2438,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link #endReadyForEventsSend}
    * and {@link #endReadyForEvents}.
+   *
    * @return the start time of this readyForEvents
    */
   @Override
@@ -2402,6 +2451,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the readyForEvents has completed
+   *
    * @param startTime the value returned by {@link #startReadyForEvents}.
    * @param failed true if the send of the readyForEvents failed
    */
@@ -2413,6 +2463,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified readyForEvents has ended
+   *
    * @param startTime the value returned by {@link #startReadyForEvents}.
    * @param timedOut true if readyForEvents timed out
    * @param failed true if readyForEvents failed
@@ -2429,6 +2480,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link #endMakePrimarySend} and
    * {@link #endMakePrimary}.
+   *
    * @return the start time of this makePrimary
    */
   @Override
@@ -2441,6 +2493,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the makePrimary has completed
+   *
    * @param startTime the value returned by {@link #startMakePrimary}.
    * @param failed true if the send of the makePrimary failed
    */
@@ -2452,6 +2505,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified makePrimary has ended
+   *
    * @param startTime the value returned by {@link #startMakePrimary}.
    * @param timedOut true if makePrimary timed out
    * @param failed true if makePrimary failed
@@ -2468,6 +2522,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link #endCloseConSend} and
    * {@link #endCloseCon}.
+   *
    * @return the start time of this closeCon
    */
   @Override
@@ -2480,6 +2535,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the closeCon has completed
+   *
    * @param startTime the value returned by {@link #startCloseCon}.
    * @param failed true if the send of the closeCon failed
    */
@@ -2491,6 +2547,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified closeCon has ended
+   *
    * @param startTime the value returned by {@link #startCloseCon}.
    * @param timedOut true if closeCon timed out
    * @param failed true if closeCon failed
@@ -2506,6 +2563,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link #endPrimaryAckSend} and
    * {@link #endPrimaryAck}.
+   *
    * @return the start time of this primaryAck
    */
   @Override
@@ -2518,6 +2576,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the primaryAck has completed
+   *
    * @param startTime the value returned by {@link #startPrimaryAck}.
    * @param failed true if the send of the primaryAck failed
    */
@@ -2529,6 +2588,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified primaryAck has ended
+   *
    * @param startTime the value returned by {@link #startPrimaryAck}.
    * @param timedOut true if primaryAck timed out
    * @param failed true if primaryAck failed
@@ -2545,6 +2605,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link #endPingSend} and {@link
    * #endPing}.
+   *
    * @return the start time of this ping
    */
   @Override
@@ -2557,6 +2618,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the ping has completed
+   *
    * @param startTime the value returned by {@link #startPing}.
    * @param failed true if the send of the ping failed
    */
@@ -2568,6 +2630,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified ping has ended
+   *
    * @param startTime the value returned by {@link #startPing}.
    * @param timedOut true if ping timed out
    * @param failed true if ping failed
@@ -2584,6 +2647,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link
    * #endRegisterInstantiatorsSend} and {@link #endRegisterInstantiators}.
+   *
    * @return the start time of this registerInstantiators
    */
   @Override
@@ -2604,6 +2668,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the registerInstantiators has completed
+   *
    * @param startTime the value returned by {@link #startRegisterInstantiators}.
    * @param failed true if the send of the registerInstantiators failed
    */
@@ -2623,6 +2688,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified registerInstantiators has ended
+   *
    * @param startTime the value returned by {@link #startRegisterInstantiators}.
    * @param timedOut true if registerInstantiators timed out
    * @param failed true if registerInstantiators failed
@@ -2646,6 +2712,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link #endPutAllSend} and {@link
    * #endPutAll}.
+   *
    * @return the start time of this putAll
    */
   @Override
@@ -2658,6 +2725,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the putAll has completed
+   *
    * @param startTime the value returned by {@link #startPutAll}.
    * @param failed true if the send of the putAll failed
    */
@@ -2669,6 +2737,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified putAll has ended
+   *
    * @param startTime the value returned by {@link #startPutAll}.
    * @param timedOut true if putAll timed out
    * @param failed true if putAll failed
@@ -2685,6 +2754,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link #endRemoveAllSend} and
    * {@link #endRemoveAll}.
+   *
    * @return the start time of this removeAll
    */
   @Override
@@ -2697,6 +2767,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the removeAll has completed
+   *
    * @param startTime the value returned by {@link #startRemoveAll}.
    * @param failed true if the send of the removeAll failed
    */
@@ -2708,6 +2779,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified removeAll has ended
+   *
    * @param startTime the value returned by {@link #startRemoveAll}.
    * @param timedOut true if removeAll timed out
    * @param failed true if removeAll failed
@@ -2723,6 +2795,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link #endGetAllSend} and {@link
    * #endGetAll}.
+   *
    * @return the start time of this getAll
    */
   @Override
@@ -2735,6 +2808,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the getAll has completed
+   *
    * @param startTime the value returned by {@link #startGetAll}.
    * @param failed true if the send of the getAll failed
    */
@@ -2746,6 +2820,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified getAll has ended
+   *
    * @param startTime the value returned by {@link #startGetAll}.
    * @param timedOut true if getAll timed out
    * @param failed true if getAll failed
@@ -2834,6 +2909,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link #endExecuteFunctionSend}
    * and {@link #endExecuteFunction}.
+   *
    * @return the start time of this ExecuteFunction
    */
   @Override
@@ -2845,6 +2921,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the executeFunction has completed
+   *
    * @param startTime the value returned by {@link #startExecuteFunction}.
    * @param failed true if the send of the executeFunction failed
    */
@@ -2864,6 +2941,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified executeFunction has ended
+   *
    * @param startTime the value returned by {@link #startExecuteFunction}.
    * @param timedOut true if executeFunction timed out
    * @param failed true if executeFunction failed
@@ -2904,6 +2982,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link
    * #endGetClientPRMetadataSend} and {@link #endGetClientPRMetadata}.
+   *
    * @return the start time of this ExecuteFunction
    */
   @Override
@@ -2916,6 +2995,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the GetClientPRMetadata has completed
+   *
    * @param startTime the value returned by {@link #startGetClientPRMetadata}.
    * @param failed true if the send of the GetClientPRMetadata failed
    */
@@ -2928,6 +3008,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified GetClientPRMetadata has ended
+   *
    * @param startTime the value returned by {@link #startGetClientPRMetadata}.
    * @param timedOut true if GetClientPRMetadata timed out
    * @param failed true if GetClientPRMetadata failed
@@ -2944,6 +3025,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
    * <p>
    * Note: for every call of this method the caller must also call {@link
    * #endGetClientPartitionAttributesSend} and {@link #endGetClientPartitionAttributes}.
+   *
    * @return the start time of this GetClientPartitionAttributes
    */
   @Override
@@ -2956,6 +3038,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the send part of the GetClientPartitionAttributes operation has completed
+   *
    * @param startTime the value returned by {@link #startGetClientPartitionAttributes}.
    * @param failed true if the send of the GetClientPartitionAttributes failed
    */
@@ -2968,6 +3051,7 @@ public class ConnectionStatsImpl implements ConnectionStats, GFSStatsImplementor
 
   /**
    * Records that the specified GetClientPartitionAttributes has ended
+   *
    * @param startTime the value returned by {@link #startGetClientPartitionAttributes}.
    * @param timedOut true if GetClientPartitionAttributes timed out
    * @param failed true if GetClientPartitionAttributes failed
diff --git a/geode-core/src/main/java/org/apache/geode/cache/client/internal/EndpointManagerImpl.java b/geode-core/src/main/java/org/apache/geode/cache/client/internal/EndpointManagerImpl.java
index 6440b24..c864b00 100644
--- a/geode-core/src/main/java/org/apache/geode/cache/client/internal/EndpointManagerImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/cache/client/internal/EndpointManagerImpl.java
@@ -40,8 +40,7 @@ public class EndpointManagerImpl implements EndpointManager {
   private static final Logger logger = LogService.getLogger();
 
   private volatile Map<ServerLocation, Endpoint> endpointMap = Collections.emptyMap();
-  private final Map/* <ServerLocation, ConnectionStats> */<ServerLocation, ConnectionStats>
-      statMap =
+  private final Map/* <ServerLocation, ConnectionStats> */<ServerLocation, ConnectionStats> statMap =
       new HashMap<ServerLocation, ConnectionStats>();
   private final DistributedSystem distributedSystem;
   private final String poolName;
@@ -50,7 +49,7 @@ public class EndpointManagerImpl implements EndpointManager {
   private final PoolStats poolStats;
 
   public EndpointManagerImpl(String poolName, DistributedSystem ds, CancelCriterion cancelCriterion,
-                             PoolStats poolStats) {
+      PoolStats poolStats) {
     this.distributedSystem = ds;
     this.poolName = poolName;
     this.cancelCriterion = cancelCriterion;
@@ -184,7 +183,7 @@ public class EndpointManagerImpl implements EndpointManager {
    * @see org.apache.geode.cache.client.internal.EndpointManager#close()
    */
   public synchronized void close() {
-    for (Iterator<ConnectionStats> itr = statMap.values().iterator(); itr.hasNext(); ) {
+    for (Iterator<ConnectionStats> itr = statMap.values().iterator(); itr.hasNext();) {
       ConnectionStats stats = itr.next();
       stats.close();
     }
@@ -222,12 +221,14 @@ public class EndpointManagerImpl implements EndpointManager {
       PoolImpl pool = (PoolImpl) PoolManager.find(this.poolName);
       if (pool != null) {
         if (pool.getGatewaySender() != null) {
-          stats = StatsFactory.createConnectionStatsImpl(new DummyStatisticsFactory(distributedSystem.getStatisticsFactory()), statName,
+          stats = StatsFactory.createConnectionStatsImpl(
+              new DummyStatisticsFactory(distributedSystem.getStatisticsFactory()), statName,
               this.poolStats);
         }
       }
       if (stats == null) {
-        stats =StatsFactory.createConnectionStatsImpl(distributedSystem.getStatisticsFactory(), statName, this.poolStats);
+        stats = StatsFactory.createConnectionStatsImpl(distributedSystem.getStatisticsFactory(),
+            statName, this.poolStats);
       }
       statMap.put(location, stats);
     }
@@ -269,21 +270,21 @@ public class EndpointManagerImpl implements EndpointManager {
     }
 
     public void endpointCrashed(Endpoint endpoint) {
-      for (Iterator<EndpointListener> itr = endpointListeners.iterator(); itr.hasNext(); ) {
+      for (Iterator<EndpointListener> itr = endpointListeners.iterator(); itr.hasNext();) {
         EndpointManager.EndpointListener listener = itr.next();
         listener.endpointCrashed(endpoint);
       }
     }
 
     public void endpointNoLongerInUse(Endpoint endpoint) {
-      for (Iterator<EndpointListener> itr = endpointListeners.iterator(); itr.hasNext(); ) {
+      for (Iterator<EndpointListener> itr = endpointListeners.iterator(); itr.hasNext();) {
         EndpointManager.EndpointListener listener = itr.next();
         listener.endpointNoLongerInUse(endpoint);
       }
     }
 
     public void endpointNowInUse(Endpoint endpoint) {
-      for (Iterator<EndpointListener> itr = endpointListeners.iterator(); itr.hasNext(); ) {
+      for (Iterator<EndpointListener> itr = endpointListeners.iterator(); itr.hasNext();) {
         EndpointManager.EndpointListener listener = itr.next();
         if (!(listener instanceof PdxRegistryRecoveryListener)) {
           listener.endpointNowInUse(endpoint);
@@ -292,7 +293,7 @@ public class EndpointManagerImpl implements EndpointManager {
     }
 
     public void clearPdxRegistry(Endpoint endpoint) {
-      for (Iterator<EndpointListener> itr = endpointListeners.iterator(); itr.hasNext(); ) {
+      for (Iterator<EndpointListener> itr = endpointListeners.iterator(); itr.hasNext();) {
         EndpointManager.EndpointListener listener = itr.next();
         if (listener instanceof PdxRegistryRecoveryListener) {
           listener.endpointNowInUse(endpoint);
diff --git a/geode-core/src/main/java/org/apache/geode/cache/client/internal/ExecuteFunctionOp.java b/geode-core/src/main/java/org/apache/geode/cache/client/internal/ExecuteFunctionOp.java
index 6d69383..e0394bb 100755
--- a/geode-core/src/main/java/org/apache/geode/cache/client/internal/ExecuteFunctionOp.java
+++ b/geode-core/src/main/java/org/apache/geode/cache/client/internal/ExecuteFunctionOp.java
@@ -30,7 +30,6 @@ import org.apache.geode.distributed.DistributedMember;
 import org.apache.geode.distributed.internal.ServerLocation;
 import org.apache.geode.internal.Version;
 import org.apache.geode.internal.cache.execute.AbstractExecution;
-import org.apache.geode.internal.cache.execute.FunctionStats;
 import org.apache.geode.internal.cache.execute.FunctionStatsImpl;
 import org.apache.geode.internal.cache.execute.InternalFunctionException;
 import org.apache.geode.internal.cache.execute.InternalFunctionInvocationTargetException;
diff --git a/geode-core/src/main/java/org/apache/geode/cache/client/internal/ExecuteRegionFunctionOp.java b/geode-core/src/main/java/org/apache/geode/cache/client/internal/ExecuteRegionFunctionOp.java
index 2e31c01..7c0cb91 100644
--- a/geode-core/src/main/java/org/apache/geode/cache/client/internal/ExecuteRegionFunctionOp.java
+++ b/geode-core/src/main/java/org/apache/geode/cache/client/internal/ExecuteRegionFunctionOp.java
@@ -34,7 +34,6 @@ import org.apache.geode.distributed.DistributedMember;
 import org.apache.geode.internal.Version;
 import org.apache.geode.internal.cache.execute.AbstractExecution;
 import org.apache.geode.internal.cache.execute.BucketMovedException;
-import org.apache.geode.internal.cache.execute.FunctionStats;
 import org.apache.geode.internal.cache.execute.FunctionStatsImpl;
 import org.apache.geode.internal.cache.execute.InternalFunctionException;
 import org.apache.geode.internal.cache.execute.InternalFunctionInvocationTargetException;
@@ -519,7 +518,8 @@ public class ExecuteRegionFunctionOp {
                   DistributedMember memberID =
                       (DistributedMember) ((ArrayList) resultResponse).get(1);
                   this.resultCollector.addResult(memberID, cause);
-                  FunctionStatsImpl.getFunctionStats(this.functionId, this.executor.getRegion().getSystem())
+                  FunctionStatsImpl
+                      .getFunctionStats(this.functionId, this.executor.getRegion().getSystem())
                       .incResultsReceived();
                   continue;
                 } else if (((FunctionException) result)
@@ -573,7 +573,8 @@ public class ExecuteRegionFunctionOp {
                 DistributedMember memberID =
                     (DistributedMember) ((ArrayList) resultResponse).get(1);
                 this.resultCollector.addResult(memberID, result);
-                FunctionStatsImpl.getFunctionStats(this.functionId, this.executor.getRegion().getSystem())
+                FunctionStatsImpl
+                    .getFunctionStats(this.functionId, this.executor.getRegion().getSystem())
                     .incResultsReceived();
               }
             } while (!executeFunctionResponseMsg.isLastChunk());
diff --git a/geode-core/src/main/java/org/apache/geode/cache/client/internal/ExecuteRegionFunctionSingleHopOp.java b/geode-core/src/main/java/org/apache/geode/cache/client/internal/ExecuteRegionFunctionSingleHopOp.java
index 40a934f..0fcd685 100644
--- a/geode-core/src/main/java/org/apache/geode/cache/client/internal/ExecuteRegionFunctionSingleHopOp.java
+++ b/geode-core/src/main/java/org/apache/geode/cache/client/internal/ExecuteRegionFunctionSingleHopOp.java
@@ -37,7 +37,6 @@ import org.apache.geode.internal.cache.InternalCache;
 import org.apache.geode.internal.cache.LocalRegion;
 import org.apache.geode.internal.cache.execute.AbstractExecution;
 import org.apache.geode.internal.cache.execute.BucketMovedException;
-import org.apache.geode.internal.cache.execute.FunctionStats;
 import org.apache.geode.internal.cache.execute.FunctionStatsImpl;
 import org.apache.geode.internal.cache.execute.InternalFunctionException;
 import org.apache.geode.internal.cache.execute.InternalFunctionInvocationTargetException;
@@ -342,7 +341,8 @@ public class ExecuteRegionFunctionSingleHopOp {
                   Throwable cause = ex.getCause();
                   DistributedMember memberID = (DistributedMember) ((List) resultResponse).get(1);
                   this.resultCollector.addResult(memberID, cause);
-                  FunctionStatsImpl.getFunctionStats(this.functionId, this.executor.getRegion().getSystem())
+                  FunctionStatsImpl
+                      .getFunctionStats(this.functionId, this.executor.getRegion().getSystem())
                       .incResultsReceived();
                   continue;
                 } else if (((FunctionException) result)
@@ -373,7 +373,8 @@ public class ExecuteRegionFunctionSingleHopOp {
               } else {
                 DistributedMember memberID = (DistributedMember) ((List) resultResponse).get(1);
                 this.resultCollector.addResult(memberID, result);
-                FunctionStatsImpl.getFunctionStats(this.functionId, this.executor.getRegion().getSystem())
+                FunctionStatsImpl
+                    .getFunctionStats(this.functionId, this.executor.getRegion().getSystem())
                     .incResultsReceived();
               }
             } while (!executeFunctionResponseMsg.isLastChunk());
diff --git a/geode-core/src/main/java/org/apache/geode/cache/client/internal/PoolImpl.java b/geode-core/src/main/java/org/apache/geode/cache/client/internal/PoolImpl.java
index 9afcf47..f4d7e3e 100644
--- a/geode-core/src/main/java/org/apache/geode/cache/client/internal/PoolImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/cache/client/internal/PoolImpl.java
@@ -35,7 +35,6 @@ import org.apache.logging.log4j.Logger;
 
 import org.apache.geode.CancelCriterion;
 import org.apache.geode.CancelException;
-import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.SystemFailure;
 import org.apache.geode.cache.CacheClosedException;
 import org.apache.geode.cache.NoSubscriptionServersAvailableException;
@@ -68,6 +67,7 @@ import org.apache.geode.internal.logging.LogService;
 import org.apache.geode.internal.logging.log4j.LocalizedMessage;
 import org.apache.geode.internal.monitoring.ThreadsMonitoring;
 import org.apache.geode.internal.statistics.DummyStatisticsFactory;
+import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatsFactory;
 
 /**
@@ -229,7 +229,8 @@ public class PoolImpl implements InternalPool {
     }
     this.cache = cache;
     this.securityLogWriter = this.internalDistributedSystem.getSecurityInternalLogWriter();
-    if (!this.internalDistributedSystem.getConfig().getStatisticSamplingEnabled() && this.statisticInterval > 0) {
+    if (!this.internalDistributedSystem.getConfig().getStatisticSamplingEnabled()
+        && this.statisticInterval > 0) {
       logger.info(LocalizedMessage.create(
           LocalizedStrings.PoolImpl_STATISTIC_SAMPLING_MUST_BE_ENABLED_FOR_SAMPLING_RATE_OF_0_TO_TAKE_AFFECT,
           this.statisticInterval));
@@ -244,19 +245,22 @@ public class PoolImpl implements InternalPool {
     }
     StatisticsFactory statFactory = null;
     if (this.gatewaySender != null) {
-      statFactory = new DummyStatisticsFactory(this.internalDistributedSystem.getStatisticsFactory());
+      statFactory =
+          new DummyStatisticsFactory(this.internalDistributedSystem.getStatisticsFactory());
     } else {
       statFactory = this.internalDistributedSystem.getInternalDistributedSystemStats();
     }
-    this.stats = this.startDisabled ? null :
-        StatsFactory.createPoolStatsImpl(statFactory, getName() + "->"
+    this.stats =
+        this.startDisabled ? null : StatsFactory.createPoolStatsImpl(statFactory, getName() + "->"
             + (isEmpty(serverGroup) ? "[any servers]" : "[" + getServerGroup() + "]"));
 
     source = getSourceImpl(((PoolFactoryImpl.PoolAttributes) attributes).locatorCallback);
-    endpointManager = new EndpointManagerImpl(name, this.internalDistributedSystem, this.cancelCriterion, this.stats);
-    connectionFactory = new ConnectionFactoryImpl(source, endpointManager, this.internalDistributedSystem,
-        socketBufferSize, socketConnectTimeout, readTimeout, proxyId, this.cancelCriterion,
-        usedByGateway, gatewaySender, pingInterval, multiuserSecureModeEnabled, this);
+    endpointManager = new EndpointManagerImpl(name, this.internalDistributedSystem,
+        this.cancelCriterion, this.stats);
+    connectionFactory =
+        new ConnectionFactoryImpl(source, endpointManager, this.internalDistributedSystem,
+            socketBufferSize, socketConnectTimeout, readTimeout, proxyId, this.cancelCriterion,
+            usedByGateway, gatewaySender, pingInterval, multiuserSecureModeEnabled, this);
     if (subscriptionEnabled) {
       queueManager = new QueueManagerImpl(this, endpointManager, source, connectionFactory,
           subscriptionRedundancyLevel, pingInterval, securityLogWriter, proxyId);
@@ -357,7 +361,8 @@ public class PoolImpl implements InternalPool {
     }
 
 
-    if (this.statisticInterval > 0 && this.internalDistributedSystem.getConfig().getStatisticSamplingEnabled()) {
+    if (this.statisticInterval > 0
+        && this.internalDistributedSystem.getConfig().getStatisticSamplingEnabled()) {
       backgroundProcessor.scheduleWithFixedDelay(new PublishClientStatsTask(), statisticInterval,
           statisticInterval, TimeUnit.MILLISECONDS);
     }
diff --git a/geode-core/src/main/java/org/apache/geode/cache/query/internal/CqQueryVsdStatsImpl.java b/geode-core/src/main/java/org/apache/geode/cache/query/internal/CqQueryVsdStatsImpl.java
index 7a09ed7..61cb834 100644
--- a/geode-core/src/main/java/org/apache/geode/cache/query/internal/CqQueryVsdStatsImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/cache/query/internal/CqQueryVsdStatsImpl.java
@@ -14,13 +14,13 @@
  */
 package org.apache.geode.cache.query.internal;
 
+import org.apache.geode.cache.query.CqEvent;
+import org.apache.geode.internal.cache.tier.MessageType;
 import org.apache.geode.statistics.GFSStatsImplementor;
 import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.cache.query.CqEvent;
-import org.apache.geode.internal.cache.tier.MessageType;
 
 /**
  * This class tracks GemFire statistics related to a {@link org.apache.geode.cache.query.CqQuery}.
@@ -103,9 +103,11 @@ public abstract class CqQueryVsdStatsImpl implements CqQueryVsdStats, GFSStatsIm
             factory.createLongCounter(CQ_EVENTS,
                 "Total number of inserts, updates, and deletes done on this cq.", "operations"),
 
-            factory.createLongGauge(NUM_HA_QUEUED_CQ_EVENTS, "Number of events in this cq.", "events"),
+            factory.createLongGauge(NUM_HA_QUEUED_CQ_EVENTS, "Number of events in this cq.",
+                "events"),
 
-            factory.createLongCounter(CQ_LISTENER_INVOCATIONS, "Total number of CqListener invocations.",
+            factory.createLongCounter(CQ_LISTENER_INVOCATIONS,
+                "Total number of CqListener invocations.",
                 "operations"),
 
             factory.createLongGauge(QUEUED_CQ_LISTENER_EVENTS,
diff --git a/geode-core/src/main/java/org/apache/geode/cache/query/internal/index/AbstractMapIndex.java b/geode-core/src/main/java/org/apache/geode/cache/query/internal/index/AbstractMapIndex.java
index a1a5933..1cb9692 100644
--- a/geode-core/src/main/java/org/apache/geode/cache/query/internal/index/AbstractMapIndex.java
+++ b/geode-core/src/main/java/org/apache/geode/cache/query/internal/index/AbstractMapIndex.java
@@ -90,7 +90,8 @@ public abstract class AbstractMapIndex extends AbstractIndex {
     private IndexStats vsdStats;
 
     public MapIndexStatistics(String indexName) {
-      this.vsdStats = new IndexStats(getRegion().getCache().getDistributedSystem().getStatisticsFactory(), indexName);
+      this.vsdStats = new IndexStats(
+          getRegion().getCache().getDistributedSystem().getStatisticsFactory(), indexName);
     }
 
     /**
diff --git a/geode-core/src/main/java/org/apache/geode/cache/query/internal/index/CompactRangeIndex.java b/geode-core/src/main/java/org/apache/geode/cache/query/internal/index/CompactRangeIndex.java
index 00f6e73..edbc063 100644
--- a/geode-core/src/main/java/org/apache/geode/cache/query/internal/index/CompactRangeIndex.java
+++ b/geode-core/src/main/java/org/apache/geode/cache/query/internal/index/CompactRangeIndex.java
@@ -969,7 +969,8 @@ public class CompactRangeIndex extends AbstractIndex {
     private IndexStats vsdStats;
 
     public RangeIndexStatistics(String indexName) {
-      this.vsdStats = new IndexStats(getRegion().getCache().getDistributedSystem().getStatisticsFactory(), indexName);
+      this.vsdStats = new IndexStats(
+          getRegion().getCache().getDistributedSystem().getStatisticsFactory(), indexName);
     }
 
     /**
diff --git a/geode-core/src/main/java/org/apache/geode/cache/query/internal/index/HashIndex.java b/geode-core/src/main/java/org/apache/geode/cache/query/internal/index/HashIndex.java
index 2097f14..5bd41d7 100755
--- a/geode-core/src/main/java/org/apache/geode/cache/query/internal/index/HashIndex.java
+++ b/geode-core/src/main/java/org/apache/geode/cache/query/internal/index/HashIndex.java
@@ -842,7 +842,8 @@ public class HashIndex extends AbstractIndex {
     private IndexStats vsdStats;
 
     public RangeIndexStatistics(String indexName) {
-      this.vsdStats = new IndexStats(getRegion().getCache().getDistributedSystem().getStatisticsFactory(), indexName);
+      this.vsdStats = new IndexStats(
+          getRegion().getCache().getDistributedSystem().getStatisticsFactory(), indexName);
     }
 
     /**
diff --git a/geode-core/src/main/java/org/apache/geode/cache/query/internal/index/IndexStats.java b/geode-core/src/main/java/org/apache/geode/cache/query/internal/index/IndexStats.java
index 4030d34..9a0d30c 100644
--- a/geode-core/src/main/java/org/apache/geode/cache/query/internal/index/IndexStats.java
+++ b/geode-core/src/main/java/org/apache/geode/cache/query/internal/index/IndexStats.java
@@ -19,9 +19,6 @@ import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.statistics.StatisticsTypeFactory;
-import org.apache.geode.internal.cache.CachePerfStats;
-import org.apache.geode.internal.statistics.StatisticsTypeFactoryImpl;
 
 /**
  * IndexStats tracks statistics about query index use.
@@ -58,7 +55,8 @@ public class IndexStats {
             factory.createLongCounter("numUpdates", numUpdatesDesc, "operations"),
             factory.createLongCounter("numUses", numUsesDesc, "operations"),
             factory.createLongCounter("updateTime", updateTimeDesc, "nanoseconds"),
-            factory.createLongCounter("useTime", "Total time spent using this index", "nanoseconds"),
+            factory.createLongCounter("useTime", "Total time spent using this index",
+                "nanoseconds"),
             factory.createIntGauge("updatesInProgress", "Current number of updates in progress.",
                 "updates"),
             factory.createIntGauge("usesInProgress", "Current number of uses in progress.", "uses"),
diff --git a/geode-core/src/main/java/org/apache/geode/cache/query/internal/index/PartitionedIndex.java b/geode-core/src/main/java/org/apache/geode/cache/query/internal/index/PartitionedIndex.java
index f9f3977..aec5a39 100755
--- a/geode-core/src/main/java/org/apache/geode/cache/query/internal/index/PartitionedIndex.java
+++ b/geode-core/src/main/java/org/apache/geode/cache/query/internal/index/PartitionedIndex.java
@@ -425,7 +425,8 @@ public class PartitionedIndex extends AbstractIndex {
     private IndexStats vsdStats;
 
     public PartitionedIndexStatistics(String indexName) {
-      this.vsdStats = new IndexStats(getRegion().getCache().getDistributedSystem().getStatisticsFactory(), indexName);
+      this.vsdStats = new IndexStats(
+          getRegion().getCache().getDistributedSystem().getStatisticsFactory(), indexName);
     }
 
     /**
diff --git a/geode-core/src/main/java/org/apache/geode/cache/query/internal/index/RangeIndex.java b/geode-core/src/main/java/org/apache/geode/cache/query/internal/index/RangeIndex.java
index 045a0f4..24db5b3 100644
--- a/geode-core/src/main/java/org/apache/geode/cache/query/internal/index/RangeIndex.java
+++ b/geode-core/src/main/java/org/apache/geode/cache/query/internal/index/RangeIndex.java
@@ -1470,7 +1470,8 @@ public class RangeIndex extends AbstractIndex {
 
     public RangeIndexStatistics(String indexName) {
 
-      this.vsdStats = new IndexStats(getRegion().getCache().getDistributedSystem().getStatisticsFactory(), indexName);
+      this.vsdStats = new IndexStats(
+          getRegion().getCache().getDistributedSystem().getStatisticsFactory(), indexName);
     }
 
     /**
diff --git a/geode-core/src/main/java/org/apache/geode/cache/server/internal/LoadMonitor.java b/geode-core/src/main/java/org/apache/geode/cache/server/internal/LoadMonitor.java
index 192e471..ad46c55 100644
--- a/geode-core/src/main/java/org/apache/geode/cache/server/internal/LoadMonitor.java
+++ b/geode-core/src/main/java/org/apache/geode/cache/server/internal/LoadMonitor.java
@@ -29,7 +29,6 @@ import org.apache.geode.distributed.internal.membership.MembershipManager;
 import org.apache.geode.internal.cache.CacheServerAdvisor;
 import org.apache.geode.internal.cache.tier.CommunicationMode;
 import org.apache.geode.internal.cache.tier.sockets.CacheServerStats;
-import org.apache.geode.internal.cache.tier.sockets.CacheServerStatsImpl;
 import org.apache.geode.internal.cache.tier.sockets.ClientProxyMembershipID;
 import org.apache.geode.internal.cache.tier.sockets.ConnectionListener;
 import org.apache.geode.internal.i18n.LocalizedStrings;
@@ -71,7 +70,8 @@ public class LoadMonitor implements ConnectionListener {
     this.location = location;
     this.pollingThread.start();
     this.stats = cacheServerStats;
-    this.stats.setLoad(lastLoad.getConnectionLoad(),lastLoad.getLoadPerConnection(),lastLoad.getSubscriptionConnectionLoad(),lastLoad.getLoadPerSubscriptionConnection());
+    this.stats.setLoad(lastLoad.getConnectionLoad(), lastLoad.getLoadPerConnection(),
+        lastLoad.getSubscriptionConnectionLoad(), lastLoad.getLoadPerSubscriptionConnection());
   }
 
   /**
@@ -200,7 +200,8 @@ public class LoadMonitor implements ConnectionListener {
                   locators);
             }
 
-            stats.setLoad(load.getConnectionLoad(),load.getLoadPerConnection(),load.getSubscriptionConnectionLoad(),load.getLoadPerSubscriptionConnection());
+            stats.setLoad(load.getConnectionLoad(), load.getLoadPerConnection(),
+                load.getSubscriptionConnectionLoad(), load.getLoadPerSubscriptionConnection());
             if (locators != null) {
               CacheServerLoadMessage message =
                   new CacheServerLoadMessage(load, location, myClientIds);
diff --git a/geode-core/src/main/java/org/apache/geode/distributed/DistributedSystem.java b/geode-core/src/main/java/org/apache/geode/distributed/DistributedSystem.java
index c38a1c5..da755c5 100644
--- a/geode-core/src/main/java/org/apache/geode/distributed/DistributedSystem.java
+++ b/geode-core/src/main/java/org/apache/geode/distributed/DistributedSystem.java
@@ -31,7 +31,6 @@ import java.util.concurrent.TimeUnit;
 
 import org.apache.geode.CancelCriterion;
 import org.apache.geode.LogWriter;
-import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.cache.Cache;
 import org.apache.geode.cache.CacheFactory;
 import org.apache.geode.cache.client.ClientCache;
@@ -44,6 +43,7 @@ import org.apache.geode.internal.ClassPathLoader;
 import org.apache.geode.internal.i18n.LocalizedStrings;
 import org.apache.geode.internal.tcp.ConnectionTable;
 import org.apache.geode.internal.util.IOUtils;
+import org.apache.geode.statistics.StatisticsFactory;
 
 /**
  * A "connection" to a GemFire distributed system. A <code>DistributedSystem</code> is created by
@@ -77,8 +77,8 @@ import org.apache.geode.internal.util.IOUtils;
  *
  * @since GemFire 3.0
  */
-//public abstract class DistributedSystem implements StatisticsFactory {
-public abstract class DistributedSystem{
+// public abstract class DistributedSystem implements StatisticsFactory {
+public abstract class DistributedSystem {
 
   /**
    * The instances of <code>DistributedSystem</code> created in this VM. Presently only one connect
diff --git a/geode-core/src/main/java/org/apache/geode/distributed/internal/ClusterDistributionManager.java b/geode-core/src/main/java/org/apache/geode/distributed/internal/ClusterDistributionManager.java
index c6ae1a2..128fec3 100644
--- a/geode-core/src/main/java/org/apache/geode/distributed/internal/ClusterDistributionManager.java
+++ b/geode-core/src/main/java/org/apache/geode/distributed/internal/ClusterDistributionManager.java
@@ -105,6 +105,7 @@ import org.apache.geode.statistics.StatsFactory;
  * <code>DistributionManager</code> and invoke {@link #putOutgoing}.
  *
  * <P>
+ *
  * @see DistributionMessage#process
  * @see IgnoredByManager
  */
@@ -251,6 +252,7 @@ public class ClusterDistributionManager implements DistributionManager {
 
   /**
    * Executor for view related messages
+   *
    * @see org.apache.geode.distributed.internal.membership.gms.messages.ViewAckMessage
    */
   public static final int VIEW_EXECUTOR = 79;
@@ -286,6 +288,7 @@ public class ClusterDistributionManager implements DistributionManager {
 
   /**
    * Must be read/written while holding {@link #elderMonitor}
+   *
    * @see #elderChangeWait()
    */
   private boolean waitingForElderChange = false;
@@ -300,6 +303,7 @@ public class ClusterDistributionManager implements DistributionManager {
    * arbitration.
    *
    * Must hold {@link #elderMonitor} in order to change this.
+   *
    * @see #getElderId()
    */
   protected volatile InternalDistributedMember elder = null;
@@ -321,12 +325,14 @@ public class ClusterDistributionManager implements DistributionManager {
 
   /**
    * The <code>MembershipListener</code>s that are registered on this manager for ALL members.
+   *
    * @since GemFire 5.7
    */
   private volatile Set<MembershipListener> allMembershipListeners = Collections.emptySet();
 
   /**
    * A lock to hold while adding and removing all membership listeners.
+   *
    * @since GemFire 5.7
    */
   private final Object allMembershipListenersLock = new Object();
@@ -399,8 +405,12 @@ public class ClusterDistributionManager implements DistributionManager {
       Collections.emptyMap();
 
   /**
-   * The lock held while accessing the field references to the following:<br> 1) {@link
-   * #members}<br> 2) {@link #membersAndAdmin}<br> 3) {@link #hostedLocatorsAll}<br> 4) {@link
+   * The lock held while accessing the field references to the following:<br>
+   * 1) {@link
+   * #members}<br>
+   * 2) {@link #membersAndAdmin}<br>
+   * 3) {@link #hostedLocatorsAll}<br>
+   * 4) {@link
    * #hostedLocatorsWithSharedConfiguration}<br>
    */
   private final Object membersLock = new Object();
@@ -444,6 +454,7 @@ public class ClusterDistributionManager implements DistributionManager {
   /**
    * Thread used to decouple {@link org.apache.geode.internal.cache.partitioned.PartitionMessage}s
    * from {@link org.apache.geode.internal.cache.DistributedCacheOperation}s </b>
+   *
    * @see #SERIAL_EXECUTOR
    */
   private ThreadPoolExecutor partitionedRegionThread;
@@ -462,6 +473,7 @@ public class ClusterDistributionManager implements DistributionManager {
 
   /**
    * Message processing executor for view messages
+   *
    * @see org.apache.geode.distributed.internal.membership.gms.messages.ViewAckMessage
    */
   private ThreadPoolExecutor viewThread;
@@ -474,6 +486,7 @@ public class ClusterDistributionManager implements DistributionManager {
 
   /**
    * Thread Monitor mechanism to monitor system threads
+   *
    * @see org.apache.geode.internal.monitoring.ThreadsMonitoring
    */
   private final ThreadsMonitoring threadMonitor;
@@ -556,6 +569,7 @@ public class ClusterDistributionManager implements DistributionManager {
   /**
    * Creates a new distribution manager and discovers the other members of the distributed system.
    * Note that it does not check to see whether or not this VM already has a distribution manager.
+   *
    * @param system The distributed system to which this distribution manager will send messages.
    */
   static ClusterDistributionManager create(InternalDistributedSystem system) {
@@ -649,7 +663,7 @@ public class ClusterDistributionManager implements DistributionManager {
 
       if (logger.isInfoEnabled()) {
         long delta = System.currentTimeMillis() - start;
-        Object[] logArgs = new Object[]{distributionManager.getDistributionManagerId(), transport,
+        Object[] logArgs = new Object[] {distributionManager.getDistributionManagerId(), transport,
             Integer.valueOf(distributionManager.getOtherDistributionManagerIds().size()),
             distributionManager.getOtherDistributionManagerIds(),
             (logger.isInfoEnabled(LogMarker.DM_MARKER) ? " (VERBOSE, took " + delta + " ms)" : ""),
@@ -679,10 +693,11 @@ public class ClusterDistributionManager implements DistributionManager {
   /**
    * Creates a new <code>DistributionManager</code> by initializing itself, creating the membership
    * manager and executors
+   *
    * @param transport The configuration for the communications transport
    */
   private ClusterDistributionManager(RemoteTransportConfig transport,
-                                     InternalDistributedSystem system) {
+      InternalDistributedSystem system) {
 
     this.dmType = transport.getVmKind();
     this.system = system;
@@ -1074,7 +1089,7 @@ public class ClusterDistributionManager implements DistributionManager {
 
       logger.info(LocalizedMessage.create(
           LocalizedStrings.DistributionManager_STARTING_DISTRIBUTIONMANAGER_0_1,
-          new Object[]{this.localAddress,
+          new Object[] {this.localAddress,
               (logger.isInfoEnabled(LogMarker.DM_MARKER) ? sb.toString() : "")}));
 
       this.description = "Distribution manager on " + this.localAddress + " started at "
@@ -1090,10 +1105,11 @@ public class ClusterDistributionManager implements DistributionManager {
 
   /**
    * Creates a new distribution manager
+   *
    * @param system The distributed system to which this distribution manager will send messages.
    */
   private ClusterDistributionManager(InternalDistributedSystem system,
-                                     RemoteTransportConfig transport) {
+      RemoteTransportConfig transport) {
     this(transport, system);
 
     boolean finishedConstructor = false;
@@ -1105,7 +1121,7 @@ public class ClusterDistributionManager implements DistributionManager {
 
       // Allow events to start being processed.
       membershipManager.startEventProcessing();
-      for (; ; ) {
+      for (;;) {
         this.getCancelCriterion().checkCancelInProgress(null);
         boolean interrupted = Thread.interrupted();
         try {
@@ -1131,6 +1147,7 @@ public class ClusterDistributionManager implements DistributionManager {
   /**
    * Is this VM dedicated to administration (like a GUI console or a JMX agent)? If so, then it
    * creates {@link #ADMIN_ONLY_DM_TYPE} type distribution managers.
+   *
    * @since GemFire 4.0
    */
   public static boolean isDedicatedAdminVM() {
@@ -1177,12 +1194,13 @@ public class ClusterDistributionManager implements DistributionManager {
   /**
    * Returns true if the two members are on the same equivalent host based on overlapping IP
    * addresses collected for all NICs during exchange of startup messages.
+   *
    * @param member1 First member
    * @param member2 Second member
    */
   @Override
   public boolean areOnEquivalentHost(InternalDistributedMember member1,
-                                     InternalDistributedMember member2) {
+      InternalDistributedMember member2) {
     Set<InetAddress> equivalents1 = getEquivalents(member1.getInetAddress());
     return equivalents1.contains(member2.getInetAddress());
   }
@@ -1190,6 +1208,7 @@ public class ClusterDistributionManager implements DistributionManager {
   /**
    * Set the host equivalencies for a given host. This overrides any previous information in the
    * tables.
+   *
    * @param equivs list of InetAddress's that all point at same host
    */
   void setEquivalentHosts(Set<InetAddress> equivs) {
@@ -1204,6 +1223,7 @@ public class ClusterDistributionManager implements DistributionManager {
 
   /**
    * Return all of the InetAddress's that are equivalent to the given one (same host)
+   *
    * @param in host to match up
    * @return all the addresses thus equivalent
    */
@@ -1227,7 +1247,7 @@ public class ClusterDistributionManager implements DistributionManager {
     }
     if (member != getDistributionManagerId()) {
       String relationship = areInSameZone(getDistributionManagerId(), member) ? "" : "not ";
-      Object[] logArgs = new Object[]{member, relationship};
+      Object[] logArgs = new Object[] {member, relationship};
       logger.info(LocalizedMessage.create(
           LocalizedStrings.DistributionManager_DISTRIBUTIONMANAGER_MEMBER_0_IS_1_EQUIVALENT,
           logArgs));
@@ -1388,6 +1408,7 @@ public class ClusterDistributionManager implements DistributionManager {
 
   /**
    * Return when DM is ready to send out messages.
+   *
    * @param msg the messsage that is currently being sent
    */
   private void waitUntilReadyToSendMsgs(DistributionMessage msg) {
@@ -1503,11 +1524,12 @@ public class ClusterDistributionManager implements DistributionManager {
    * Adds the entry in {@link #hostedLocatorsAll} for a member with one or more hosted locators. The
    * value is a collection of host[port] strings. If a bind-address was used for a locator then the
    * form is bind-addr[port].
+   *
    * @since GemFire 6.6.3
    */
   @Override
   public void addHostedLocators(InternalDistributedMember member, Collection<String> locators,
-                                boolean isSharedConfigurationEnabled) {
+      boolean isSharedConfigurationEnabled) {
     synchronized (this.membersLock) {
       if (locators == null || locators.isEmpty()) {
         throw new IllegalArgumentException("Cannot use empty collection of locators");
@@ -1573,6 +1595,7 @@ public class ClusterDistributionManager implements DistributionManager {
    * Gets the value in {@link #hostedLocatorsAll} for a member with one or more hosted locators. The
    * value is a collection of host[port] strings. If a bind-address was used for a locator then the
    * form is bind-addr[port].
+   *
    * @since GemFire 6.6.3
    */
   @Override
@@ -1594,6 +1617,7 @@ public class ClusterDistributionManager implements DistributionManager {
    * The keyset of the map are the locator vms in this cluster.
    *
    * the value is a collection of strings in case one vm can have multiple locators ????
+   *
    * @since GemFire 6.6.3
    */
   @Override
@@ -1607,6 +1631,7 @@ public class ClusterDistributionManager implements DistributionManager {
    * Returns a copy of the map of all members hosting locators with shared configuration. The key is
    * the member, and the value is a collection of host[port] strings. If a bind-address was used for
    * a locator then the form is bind-addr[port].
+   *
    * @since GemFire 8.0
    */
   @Override
@@ -1756,7 +1781,7 @@ public class ClusterDistributionManager implements DistributionManager {
         : "");
     logger.info(LocalizedMessage.create(
         LocalizedStrings.DistributionManager_SHUTTING_DOWN_DISTRIBUTIONMANAGER_0_1,
-        new Object[]{this.localAddress, exceptionStatus}));
+        new Object[] {this.localAddress, exceptionStatus}));
 
     final long start = System.currentTimeMillis();
     try {
@@ -1880,7 +1905,7 @@ public class ClusterDistributionManager implements DistributionManager {
     long start = System.currentTimeMillis();
     long remaining = timeInMillis;
 
-    ExecutorService[] allExecutors = new ExecutorService[]{this.serialThread, this.viewThread,
+    ExecutorService[] allExecutors = new ExecutorService[] {this.serialThread, this.viewThread,
         this.functionExecutionThread, this.functionExecutionPool, this.partitionedRegionThread,
         this.partitionedRegionPool, this.highPriorityPool, this.waitingPool,
         this.prMetaDataCleanupThreadPool, this.threadPool};
@@ -1910,6 +1935,7 @@ public class ClusterDistributionManager implements DistributionManager {
 
   /**
    * Cheap tool to kill a referenced thread
+   *
    * @param t the thread to kill
    */
   private void clobberThread(Thread t) {
@@ -1945,6 +1971,7 @@ public class ClusterDistributionManager implements DistributionManager {
 
   /**
    * Cheap tool to examine an executor to see if it is still working
+   *
    * @return true if executor is still active
    */
   private boolean executorAlive(ThreadPoolExecutor tpe, String name) {
@@ -1970,7 +1997,7 @@ public class ClusterDistributionManager implements DistributionManager {
   private void forceThreadsToStop() {
     long endTime = System.currentTimeMillis() + MAX_STOP_TIME;
     String culprits = "";
-    for (; ; ) {
+    for (;;) {
       boolean stillAlive = false;
       culprits = "";
       if (executorAlive(this.serialThread, "serial thread")) {
@@ -2232,6 +2259,7 @@ public class ClusterDistributionManager implements DistributionManager {
 
   /**
    * This thread processes member events as they occur.
+   *
    * @see ClusterDistributionManager.MemberCrashedEvent
    * @see ClusterDistributionManager.MemberJoinedEvent
    * @see ClusterDistributionManager.MemberDepartedEvent
@@ -2242,7 +2270,7 @@ public class ClusterDistributionManager implements DistributionManager {
     @Override
     @SuppressWarnings("synthetic-access")
     public void run() {
-      for (; ; ) {
+      for (;;) {
         SystemFailure.checkFailure();
         // bug 41539 - member events need to be delivered during shutdown
         // or reply processors may hang waiting for replies from
@@ -2380,13 +2408,13 @@ public class ClusterDistributionManager implements DistributionManager {
 
   @Override
   public void retainMembersWithSameOrNewerVersion(Collection<InternalDistributedMember> members,
-                                                  Version version) {
+      Version version) {
     members.removeIf(id -> id.getVersionObject().compareTo(version) < 0);
   }
 
   @Override
   public void removeMembersWithSameOrNewerVersion(Collection<InternalDistributedMember> members,
-                                                  Version version) {
+      Version version) {
     members.removeIf(id -> id.getVersionObject().compareTo(version) >= 0);
   }
 
@@ -2516,7 +2544,7 @@ public class ClusterDistributionManager implements DistributionManager {
           throw new SystemConnectException(
               LocalizedStrings.DistributionManager_RECEIVED_NO_CONNECTION_ACKNOWLEDGMENTS_FROM_ANY_OF_THE_0_SENIOR_CACHE_MEMBERS_1
                   .toLocalizedString(
-                      new Object[]{Integer.toString(allOthers.size()), sb.toString()}));
+                      new Object[] {Integer.toString(allOthers.size()), sb.toString()}));
         } // and none responded
       } // there exist others
 
@@ -2599,13 +2627,14 @@ public class ClusterDistributionManager implements DistributionManager {
       if (numLeft != 0) {
         logger.info(LocalizedMessage.create(
             LocalizedStrings.DistributionManager_STILL_AWAITING_0_RESPONSES_FROM_1,
-            new Object[]{Integer.valueOf(numLeft), unfinishedStartups}));
+            new Object[] {Integer.valueOf(numLeft), unfinishedStartups}));
       }
     } // synchronized
   }
 
   /**
    * Processes the first startup response.
+   *
    * @see StartupResponseMessage#process
    */
   void processStartupResponse(InternalDistributedMember sender, String theRejectionMessage) {
@@ -2624,6 +2653,7 @@ public class ClusterDistributionManager implements DistributionManager {
 
   /**
    * Based on a recent JGroups view, return a member that might be the next elder.
+   *
    * @return the elder candidate, possibly this VM.
    */
   private InternalDistributedMember getElderCandidate() {
@@ -2821,6 +2851,7 @@ public class ClusterDistributionManager implements DistributionManager {
   /**
    * Makes note of a new distribution manager that has started up in the distributed cache. Invokes
    * the appropriately listeners.
+   *
    * @param theId The id of the distribution manager starting up
    */
   private void handleManagerStartup(InternalDistributedMember theId) {
@@ -2847,7 +2878,7 @@ public class ClusterDistributionManager implements DistributionManager {
     }
     logger.info(LocalizedMessage.create(
         LocalizedStrings.DistributionManager_ADMITTING_MEMBER_0_NOW_THERE_ARE_1_NONADMIN_MEMBERS,
-        new Object[]{theId, Integer.valueOf(tmp.size())}));
+        new Object[] {theId, Integer.valueOf(tmp.size())}));
     addMemberEvent(new MemberJoinedEvent(theId));
   }
 
@@ -2908,12 +2939,13 @@ public class ClusterDistributionManager implements DistributionManager {
 
   /**
    * Makes note of a console that has shut down.
+   *
    * @param theId The id of the console shutting down
    * @param crashed only true if we detect this id to be gone from a javagroup view
    * @see AdminConsoleDisconnectMessage#process
    */
   public void handleConsoleShutdown(InternalDistributedMember theId, boolean crashed,
-                                    String reason) {
+      String reason) {
     boolean removedConsole = false;
     boolean removedMember = false;
     synchronized (this.membersLock) {
@@ -2970,7 +3002,7 @@ public class ClusterDistributionManager implements DistributionManager {
       } else {
         msg = LocalizedStrings.DistributionManager_ADMINISTRATION_MEMBER_AT_0_CLOSED_1;
       }
-      logger.info(LocalizedMessage.create(msg, new Object[]{theId, reason}));
+      logger.info(LocalizedMessage.create(msg, new Object[] {theId, reason}));
     }
 
     redundancyZones.remove(theId);
@@ -2984,7 +3016,7 @@ public class ClusterDistributionManager implements DistributionManager {
 
   @Override
   public void handleManagerDeparture(InternalDistributedMember theId, boolean p_crashed,
-                                     String p_reason) {
+      String p_reason) {
 
     AlertAppender.getInstance().removeAlertListener(theId);
 
@@ -3024,7 +3056,7 @@ public class ClusterDistributionManager implements DistributionManager {
             LocalizedStrings.DistributionManager_MEMBER_AT_0_GRACEFULLY_LEFT_THE_DISTRIBUTED_CACHE_1;
         addMemberEvent(new MemberDepartedEvent(theId, p_reason));
       }
-      logger.info(LocalizedMessage.create(msg, new Object[]{theId, prettifyReason(p_reason)}));
+      logger.info(LocalizedMessage.create(msg, new Object[] {theId, prettifyReason(p_reason)}));
 
       // Remove this manager from the serialQueueExecutor.
       if (this.serialQueuedExecutorPool != null) {
@@ -3034,7 +3066,7 @@ public class ClusterDistributionManager implements DistributionManager {
   }
 
   private void handleManagerSuspect(InternalDistributedMember suspect,
-                                    InternalDistributedMember whoSuspected, String reason) {
+      InternalDistributedMember whoSuspected, String reason) {
     if (!isCurrentMember(suspect)) {
       return; // fault tolerance
     }
@@ -3052,7 +3084,7 @@ public class ClusterDistributionManager implements DistributionManager {
   }
 
   private void handleQuorumLost(Set<InternalDistributedMember> failures,
-                                List<InternalDistributedMember> remaining) {
+      List<InternalDistributedMember> remaining) {
     addMemberEvent(new QuorumLostEvent(failures, remaining));
   }
 
@@ -3128,9 +3160,10 @@ public class ClusterDistributionManager implements DistributionManager {
 
   /**
    * Actually does the work of sending a message out over the distribution channel.
+   *
    * @param message the message to send
    * @return list of recipients that did not receive the message because they left the view (null if
-   * all received it or it was sent to {@link DistributionMessage#ALL_RECIPIENTS}.
+   *         all received it or it was sent to {@link DistributionMessage#ALL_RECIPIENTS}.
    * @throws NotSerializableException If <code>message</code> cannot be serialized
    */
   Set sendOutgoing(DistributionMessage message) throws NotSerializableException {
@@ -3188,7 +3221,7 @@ public class ClusterDistributionManager implements DistributionManager {
 
       logger.fatal(
           LocalizedMessage.create(LocalizedStrings.DistributionManager_WHILE_PUSHING_MESSAGE_0_TO_1,
-              new Object[]{message, receiver}),
+              new Object[] {message, receiver}),
           ex);
       if (message == null || message.forAll()) {
         return null;
@@ -3204,7 +3237,7 @@ public class ClusterDistributionManager implements DistributionManager {
 
   /**
    * @return list of recipients who did not receive the message because they left the view (null if
-   * all received it or it was sent to {@link DistributionMessage#ALL_RECIPIENTS}).
+   *         all received it or it was sent to {@link DistributionMessage#ALL_RECIPIENTS}).
    * @throws NotSerializableException If content cannot be serialized
    */
   private Set<InternalDistributedMember> sendViaMembershipManager(
@@ -3357,6 +3390,7 @@ public class ClusterDistributionManager implements DistributionManager {
 
   /**
    * Waits until elder if newElder or newElder is no longer a member
+   *
    * @return true if newElder is the elder; false if he is no longer a member or we are the elder.
    */
   public boolean waitForElder(final InternalDistributedMember desiredElder) {
@@ -3387,13 +3421,13 @@ public class ClusterDistributionManager implements DistributionManager {
             l = new MembershipListener() {
               @Override
               public void memberJoined(DistributionManager distributionManager,
-                                       InternalDistributedMember theId) {
+                  InternalDistributedMember theId) {
                 // nothing needed
               }
 
               @Override
               public void memberDeparted(DistributionManager distributionManager,
-                                         InternalDistributedMember theId, boolean crashed) {
+                  InternalDistributedMember theId, boolean crashed) {
                 if (desiredElder.equals(theId)) {
                   notifyElderChangeWaiters();
                 }
@@ -3401,22 +3435,20 @@ public class ClusterDistributionManager implements DistributionManager {
 
               @Override
               public void memberSuspect(DistributionManager distributionManager,
-                                        InternalDistributedMember id,
-                                        InternalDistributedMember whoSuspected,
-                                        String reason) {
-              }
+                  InternalDistributedMember id,
+                  InternalDistributedMember whoSuspected,
+                  String reason) {}
 
               @Override
               public void quorumLost(DistributionManager distributionManager,
-                                     Set<InternalDistributedMember> failures,
-                                     List<InternalDistributedMember> remaining) {
-              }
+                  Set<InternalDistributedMember> failures,
+                  List<InternalDistributedMember> remaining) {}
             };
             addMembershipListener(l);
           }
           logger.info(LocalizedMessage.create(
               LocalizedStrings.DistributionManager_CHANGING_ELDER_FROM_0_TO_1,
-              new Object[]{currentElder, desiredElder}));
+              new Object[] {currentElder, desiredElder}));
           elderChangeWait();
         } // while true
       }
@@ -3581,8 +3613,9 @@ public class ClusterDistributionManager implements DistributionManager {
   /**
    * Returns a description of the distribution configuration used for this distribution manager. (in
    * ConsoleDistributionManager)
+   *
    * @return <code>null</code> if no admin {@linkplain #getAgent agent} is associated with this
-   * distribution manager
+   *         distribution manager
    */
   public String getDistributionConfigDescription() {
     if (this.agent == null) {
@@ -3601,9 +3634,10 @@ public class ClusterDistributionManager implements DistributionManager {
 
   /**
    * Returns the health monitor for this distribution manager and owner.
+   *
    * @param owner the agent that owns the returned monitor
    * @return the health monitor created by the owner; <code>null</code> if the owner has now created
-   * a monitor.
+   *         a monitor.
    * @since GemFire 3.5
    */
   @Override
@@ -3613,6 +3647,7 @@ public class ClusterDistributionManager implements DistributionManager {
 
   /**
    * Returns the health monitor for this distribution manager.
+   *
    * @param owner the agent that owns the created monitor
    * @param cfg the configuration to use when creating the monitor
    * @since GemFire 3.5
@@ -3638,6 +3673,7 @@ public class ClusterDistributionManager implements DistributionManager {
 
   /**
    * Remove a monitor that was previously created.
+   *
    * @param owner the agent that owns the monitor to remove
    */
   @Override
@@ -3768,10 +3804,11 @@ public class ClusterDistributionManager implements DistributionManager {
 
     /**
      * Constructor.
+     *
      * @param group thread group to which the threads will belog to.
      */
     SerialQueuedExecutorPool(ThreadGroup group, DistributionStats stats,
-                             boolean throttlingDisabled, ThreadsMonitoring tMonitoring) {
+        boolean throttlingDisabled, ThreadsMonitoring tMonitoring) {
       this.threadGroup = group;
       this.stats = stats;
       this.throttlingDisabled = throttlingDisabled;
@@ -3947,7 +3984,7 @@ public class ClusterDistributionManager implements DistributionManager {
         senderToSerialQueueIdMap.remove(member);
 
         // Check if any other members are using the same executor.
-        for (Iterator iter = senderToSerialQueueIdMap.values().iterator(); iter.hasNext(); ) {
+        for (Iterator iter = senderToSerialQueueIdMap.values().iterator(); iter.hasNext();) {
           Integer value = (Integer) iter.next();
           if (value.equals(queueId)) {
             isUsed = true;
@@ -3961,7 +3998,7 @@ public class ClusterDistributionManager implements DistributionManager {
             logger.info(LogMarker.DM_MARKER,
                 LocalizedMessage.create(
                     LocalizedStrings.DistributionManager_MARKING_THE_SERIALQUEUEDEXECUTOR_WITH_ID__0__USED_BY_THE_MEMBER__1__TO_BE_UNUSED,
-                    new Object[]{queueId, member}));
+                    new Object[] {queueId, member}));
           }
 
           threadMarkedForUse.add(queueId);
@@ -4049,7 +4086,7 @@ public class ClusterDistributionManager implements DistributionManager {
 
     @Override
     public void memberSuspect(InternalDistributedMember suspect,
-                              InternalDistributedMember whoSuspected, String reason) {
+        InternalDistributedMember whoSuspected, String reason) {
       dm.handleManagerSuspect(suspect, whoSuspected, reason);
     }
 
@@ -4061,7 +4098,7 @@ public class ClusterDistributionManager implements DistributionManager {
 
     @Override
     public void quorumLost(Set<InternalDistributedMember> failures,
-                           List<InternalDistributedMember> remaining) {
+        List<InternalDistributedMember> remaining) {
       dm.handleQuorumLost(failures, remaining);
     }
 
@@ -4100,10 +4137,10 @@ public class ClusterDistributionManager implements DistributionManager {
     }
 
     protected abstract void handleEvent(ClusterDistributionManager manager,
-                                        MembershipListener listener);
+        MembershipListener listener);
 
     private void handleEvent(ClusterDistributionManager manager,
-                             Set<MembershipListener> membershipListeners) {
+        Set<MembershipListener> membershipListeners) {
       for (MembershipListener listener : membershipListeners) {
         try {
           handleEvent(manager, listener);
@@ -4210,7 +4247,7 @@ public class ClusterDistributionManager implements DistributionManager {
     String reason;
 
     MemberSuspectEvent(InternalDistributedMember suspect, InternalDistributedMember whoSuspected,
-                       String reason) {
+        String reason) {
       super(suspect);
       this.whoSuspected = whoSuspected;
       this.reason = reason;
@@ -4268,7 +4305,7 @@ public class ClusterDistributionManager implements DistributionManager {
     List<InternalDistributedMember> remaining;
 
     QuorumLostEvent(Set<InternalDistributedMember> failures,
-                    List<InternalDistributedMember> remaining) {
+        List<InternalDistributedMember> remaining) {
       super(null);
       this.failures = failures;
       this.remaining = remaining;
@@ -4342,7 +4379,7 @@ public class ClusterDistributionManager implements DistributionManager {
     } else {
       buddyMembers.add(targetMember);
       Set<InetAddress> targetAddrs = getEquivalents(targetMember.getInetAddress());
-      for (Iterator i = getDistributionManagerIds().iterator(); i.hasNext(); ) {
+      for (Iterator i = getDistributionManagerIds().iterator(); i.hasNext();) {
         InternalDistributedMember o = (InternalDistributedMember) i.next();
         if (SetUtils.intersectsWith(targetAddrs, getEquivalents(o.getInetAddress()))) {
           buddyMembers.add(o);
@@ -4354,7 +4391,7 @@ public class ClusterDistributionManager implements DistributionManager {
 
   @Override
   public boolean areInSameZone(InternalDistributedMember member1,
-                               InternalDistributedMember member2) {
+      InternalDistributedMember member2) {
 
     if (!redundancyZones.isEmpty()) {
       String zone1 = redundancyZones.get(member1);
@@ -4402,7 +4439,7 @@ public class ClusterDistributionManager implements DistributionManager {
       requiresMessage.addAll(ids);
       ids.remove(localAddress);
     } else {
-      for (Iterator it = ids.iterator(); it.hasNext(); ) {
+      for (Iterator it = ids.iterator(); it.hasNext();) {
         InternalDistributedMember mbr = (InternalDistributedMember) it.next();
         if (mbr.getProcessId() > 0
             && mbr.getInetAddress().equals(this.localAddress.getInetAddress())) {
diff --git a/geode-core/src/main/java/org/apache/geode/distributed/internal/DistributionStatsImpl.java b/geode-core/src/main/java/org/apache/geode/distributed/internal/DistributionStatsImpl.java
index 8219448..326b85c 100644
--- a/geode-core/src/main/java/org/apache/geode/distributed/internal/DistributionStatsImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/distributed/internal/DistributionStatsImpl.java
@@ -16,14 +16,14 @@ package org.apache.geode.distributed.internal;
 
 import org.apache.logging.log4j.Logger;
 
+import org.apache.geode.internal.logging.LogService;
+import org.apache.geode.internal.tcp.Buffers;
+import org.apache.geode.internal.util.Breadcrumbs;
 import org.apache.geode.statistics.GFSStatsImplementor;
 import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.internal.logging.LogService;
-import org.apache.geode.internal.tcp.Buffers;
-import org.apache.geode.internal.util.Breadcrumbs;
 
 /**
  * This class maintains statistics in GemFire about the distribution manager and distribution in
@@ -441,7 +441,8 @@ public class DistributionStatsImpl implements DistributionStats, GFSStatsImpleme
         factory.createLongCounter("commitMessages", sentCommitMessagesDesc, "messages"),
         factory.createLongCounter("commitWaits", commitWaitsDesc, "messages"),
         factory.createLongCounter("sentMessagesTime", sentMessagesTimeDesc, "nanoseconds", false),
-        factory.createLongGauge("sentMessagesMaxTime", sentMessagesMaxTimeDesc, "milliseconds", false),
+        factory.createLongGauge("sentMessagesMaxTime", sentMessagesMaxTimeDesc, "milliseconds",
+            false),
         factory.createLongCounter("broadcastMessages", broadcastMessagesDesc, "messages"),
         factory.createLongCounter("broadcastMessagesTime", broadcastMessagesTimeDesc, "nanoseconds",
             false),
@@ -451,33 +452,41 @@ public class DistributionStatsImpl implements DistributionStats, GFSStatsImpleme
         factory.createLongCounter("processedMessages", processedMessagesDesc, "messages"),
         factory.createLongCounter("processedMessagesTime", processedMessagesTimeDesc, "nanoseconds",
             false),
-        factory.createLongCounter("messageProcessingScheduleTime", messageProcessingScheduleTimeDesc,
+        factory.createLongCounter("messageProcessingScheduleTime",
+            messageProcessingScheduleTimeDesc,
             "nanoseconds", false),
         factory.createIntGauge("overflowQueueSize", overflowQueueSizeDesc, "messages"),
         factory.createIntGauge("waitingQueueSize", waitingQueueSizeDesc, "messages"),
-        factory.createIntGauge("overflowQueueThrottleCount", overflowQueueThrottleCountDesc, "delays"),
+        factory.createIntGauge("overflowQueueThrottleCount", overflowQueueThrottleCountDesc,
+            "delays"),
         factory.createLongCounter("overflowQueueThrottleTime", overflowQueueThrottleTimeDesc,
             "nanoseconds", false),
         factory.createIntGauge("highPriorityQueueSize", highPriorityQueueSizeDesc, "messages"),
         factory.createIntGauge("highPriorityQueueThrottleCount", highPriorityQueueThrottleCountDesc,
             "delays"),
-        factory.createLongCounter("highPriorityQueueThrottleTime", highPriorityQueueThrottleTimeDesc,
+        factory.createLongCounter("highPriorityQueueThrottleTime",
+            highPriorityQueueThrottleTimeDesc,
             "nanoseconds", false),
         factory.createIntGauge("partitionedRegionQueueSize", highPriorityQueueSizeDesc, "messages"),
-        factory.createIntGauge("partitionedRegionQueueThrottleCount", highPriorityQueueThrottleCountDesc,
+        factory.createIntGauge("partitionedRegionQueueThrottleCount",
+            highPriorityQueueThrottleCountDesc,
             "delays"),
-        factory.createLongCounter("partitionedRegionQueueThrottleTime", highPriorityQueueThrottleTimeDesc,
+        factory.createLongCounter("partitionedRegionQueueThrottleTime",
+            highPriorityQueueThrottleTimeDesc,
             "nanoseconds", false),
         factory.createIntGauge("functionExecutionQueueSize", highPriorityQueueSizeDesc, "messages"),
-        factory.createIntGauge("functionExecutionQueueThrottleCount", highPriorityQueueThrottleCountDesc,
+        factory.createIntGauge("functionExecutionQueueThrottleCount",
+            highPriorityQueueThrottleCountDesc,
             "delays"),
-        factory.createLongCounter("functionExecutionQueueThrottleTime", highPriorityQueueThrottleTimeDesc,
+        factory.createLongCounter("functionExecutionQueueThrottleTime",
+            highPriorityQueueThrottleTimeDesc,
             "nanoseconds", false),
         factory.createIntGauge("serialQueueSize", serialQueueSizeDesc, "messages"),
         factory.createIntGauge("serialQueueBytes", serialQueueBytesDesc, "bytes"),
         factory.createIntCounter("serialPooledThread", serialPooledThreadDesc, "threads"),
         factory.createIntGauge("serialQueueThrottleCount", serialQueueThrottleCountDesc, "delays"),
-        factory.createLongCounter("serialQueueThrottleTime", serialQueueThrottleTimeDesc, "nanoseconds",
+        factory.createLongCounter("serialQueueThrottleTime", serialQueueThrottleTimeDesc,
+            "nanoseconds",
             false),
         factory.createIntGauge("serialThreads", serialThreadsDesc, "threads"),
         factory.createIntGauge("processingThreads", processingThreadsDesc, "threads"),
@@ -485,8 +494,10 @@ public class DistributionStatsImpl implements DistributionStats, GFSStatsImpleme
         factory.createIntGauge("partitionedRegionThreads", partitionedRegionThreadsDesc, "threads"),
         factory.createIntGauge("functionExecutionThreads", functionExecutionThreadsDesc, "threads"),
         factory.createIntGauge("waitingThreads", waitingThreadsDesc, "threads"),
-        factory.createLongCounter("messageChannelTime", messageChannelTimeDesc, "nanoseconds", false),
-        factory.createLongCounter("udpDispatchRequestTime", udpDispatchRequestTimeDesc, "nanoseconds",
+        factory.createLongCounter("messageChannelTime", messageChannelTimeDesc, "nanoseconds",
+            false),
+        factory.createLongCounter("udpDispatchRequestTime", udpDispatchRequestTimeDesc,
+            "nanoseconds",
             false),
         factory.createLongCounter("replyMessageTime", replyMessageTimeDesc, "nanoseconds", false),
         factory.createLongCounter("distributeMessageTime", distributeMessageTimeDesc, "nanoseconds",
@@ -517,10 +528,13 @@ public class DistributionStatsImpl implements DistributionStats, GFSStatsImpleme
         factory.createLongCounter("syncSocketWriteBytes",
             "Total number of bytes sent out in synchronous/blocking mode on sockets.", "bytes"),
 
-        factory.createIntCounter("ucastReads", "Total number of unicast datagrams received", "datagrams"),
-        factory.createLongCounter("ucastReadBytes", "Total number of bytes received in unicast datagrams",
+        factory.createIntCounter("ucastReads", "Total number of unicast datagrams received",
+            "datagrams"),
+        factory.createLongCounter("ucastReadBytes",
+            "Total number of bytes received in unicast datagrams",
             "bytes"),
-        factory.createIntCounter("ucastWrites", "Total number of unicast datagram socket write calls.",
+        factory.createIntCounter("ucastWrites",
+            "Total number of unicast datagram socket write calls.",
             "writes"),
         factory.createLongCounter("ucastWriteBytes",
             "Total number of bytes sent out on unicast datagram sockets.", "bytes"),
@@ -531,7 +545,8 @@ public class DistributionStatsImpl implements DistributionStats, GFSStatsImpleme
             "datagrams"),
         factory.createLongCounter("mcastReadBytes",
             "Total number of bytes received in multicast datagrams", "bytes"),
-        factory.createIntCounter("mcastWrites", "Total number of multicast datagram socket write calls.",
+        factory.createIntCounter("mcastWrites",
+            "Total number of multicast datagram socket write calls.",
             "writes"),
         factory.createLongCounter("mcastWriteBytes",
             "Total number of bytes sent out on multicast datagram sockets.", "bytes"),
@@ -561,7 +576,8 @@ public class DistributionStatsImpl implements DistributionStats, GFSStatsImpleme
         factory.createLongCounter("deserializedBytes",
             "Total number of bytes read by object deserialization. This includes deserialization that results in a PdxInstance.",
             "bytes"),
-        factory.createIntCounter("pdxDeserializations", "Total number of pdx deserializations.", "ops"),
+        factory.createIntCounter("pdxDeserializations", "Total number of pdx deserializations.",
+            "ops"),
         factory.createLongCounter("pdxDeserializedBytes",
             "Total number of bytes read by pdx deserialization.", "bytes"),
         factory.createLongCounter("msgSerializationTime",
@@ -616,10 +632,14 @@ public class DistributionStatsImpl implements DistributionStats, GFSStatsImpleme
             "operations"),
         factory.createIntCounter("asyncQueueFlushesCompleted", asyncQueueFlushesCompletedDesc,
             "operations"),
-        factory.createLongCounter("asyncQueueFlushTime", asyncQueueFlushTimeDesc, "nanoseconds", false),
-        factory.createIntCounter("asyncQueueTimeoutExceeded", asyncQueueTimeoutExceededDesc, "timeouts"),
-        factory.createIntCounter("asyncQueueSizeExceeded", asyncQueueSizeExceededDesc, "operations"),
-        factory.createIntCounter("asyncDistributionTimeoutExceeded", asyncDistributionTimeoutExceededDesc,
+        factory.createLongCounter("asyncQueueFlushTime", asyncQueueFlushTimeDesc, "nanoseconds",
+            false),
+        factory.createIntCounter("asyncQueueTimeoutExceeded", asyncQueueTimeoutExceededDesc,
+            "timeouts"),
+        factory.createIntCounter("asyncQueueSizeExceeded", asyncQueueSizeExceededDesc,
+            "operations"),
+        factory.createIntCounter("asyncDistributionTimeoutExceeded",
+            asyncDistributionTimeoutExceededDesc,
             "operations"),
         factory.createLongGauge("asyncQueueSize", asyncQueueSizeDesc, "bytes"),
         factory.createLongCounter("asyncQueuedMsgs", asyncQueuedMsgsDesc, "msgs"),
@@ -650,7 +670,8 @@ public class DistributionStatsImpl implements DistributionStats, GFSStatsImpleme
             "Total amount of time, in nanoseconds, spent locking a socket", "nanoseconds", false),
         factory.createIntGauge("bufferAcquiresInProgress",
             "Current number of threads waiting to acquire a buffer", "threads", false),
-        factory.createIntCounter("bufferAcquires", "Total number of times a buffer has been acquired.",
+        factory.createIntCounter("bufferAcquires",
+            "Total number of times a buffer has been acquired.",
             "operations"),
         factory.createLongCounter("bufferAcquireTime",
             "Total amount of time, in nanoseconds, spent acquiring a socket", "nanoseconds", false),
@@ -685,7 +706,8 @@ public class DistributionStatsImpl implements DistributionStats, GFSStatsImpleme
         factory.createLongCounter("serialPooledThreadStarts",
             "Total number of times a thread has been created for the serial pool(s).", "starts",
             false),
-        factory.createLongCounter("TOSentMsgs", "Total number of messages sent on thread owned senders",
+        factory.createLongCounter("TOSentMsgs",
+            "Total number of messages sent on thread owned senders",
             "messages", false),
         factory.createLongCounter("replyHandoffTime", replyHandoffTimeDesc, "nanoseconds"),
 
@@ -709,7 +731,8 @@ public class DistributionStatsImpl implements DistributionStats, GFSStatsImpleme
 
         // For GMSHealthMonitor
         factory.createLongCounter("heartbeatRequestsSent", heartbeatRequestsSentDesc, "messages"),
-        factory.createLongCounter("heartbeatRequestsReceived", heartbeatRequestsReceivedDesc, "messages"),
+        factory.createLongCounter("heartbeatRequestsReceived", heartbeatRequestsReceivedDesc,
+            "messages"),
         factory.createLongCounter("heartbeatsSent", heartbeatsSentDesc, "messages"),
         factory.createLongCounter("heartbeatsReceived", heartbeatsReceivedDesc, "messages"),
         factory.createLongCounter("suspectsSent", suspectsSentDesc, "messages"),
@@ -717,22 +740,29 @@ public class DistributionStatsImpl implements DistributionStats, GFSStatsImpleme
         factory.createLongCounter("finalCheckRequestsSent", finalCheckRequestsSentDesc, "messages"),
         factory.createLongCounter("finalCheckRequestsReceived", finalCheckRequestsReceivedDesc,
             "messages"),
-        factory.createLongCounter("finalCheckResponsesSent", finalCheckResponsesSentDesc, "messages"),
+        factory.createLongCounter("finalCheckResponsesSent", finalCheckResponsesSentDesc,
+            "messages"),
         factory.createLongCounter("finalCheckResponsesReceived", finalCheckResponsesReceivedDesc,
             "messages"),
-        factory.createLongCounter("tcpFinalCheckRequestsSent", tcpFinalCheckRequestsSentDesc, "messages"),
-        factory.createLongCounter("tcpFinalCheckRequestsReceived", tcpFinalCheckRequestsReceivedDesc,
+        factory.createLongCounter("tcpFinalCheckRequestsSent", tcpFinalCheckRequestsSentDesc,
+            "messages"),
+        factory.createLongCounter("tcpFinalCheckRequestsReceived",
+            tcpFinalCheckRequestsReceivedDesc,
             "messages"),
         factory.createLongCounter("tcpFinalCheckResponsesSent", tcpFinalCheckResponsesSentDesc,
             "messages"),
-        factory.createLongCounter("tcpFinalCheckResponsesReceived", tcpFinalCheckResponsesReceivedDesc,
+        factory.createLongCounter("tcpFinalCheckResponsesReceived",
+            tcpFinalCheckResponsesReceivedDesc,
+            "messages"),
+        factory.createLongCounter("udpFinalCheckRequestsSent", udpFinalCheckRequestsSentDesc,
             "messages"),
-        factory.createLongCounter("udpFinalCheckRequestsSent", udpFinalCheckRequestsSentDesc, "messages"),
-        factory.createLongCounter("udpFinalCheckRequestsReceived", udpFinalCheckRequestsReceivedDesc,
+        factory.createLongCounter("udpFinalCheckRequestsReceived",
+            udpFinalCheckRequestsReceivedDesc,
             "messages"),
         factory.createLongCounter("udpFinalCheckResponsesSent", udpFinalCheckResponsesSentDesc,
             "messages"),
-        factory.createLongCounter("udpFinalCheckResponsesReceived", udpFinalCheckResponsesReceivedDesc,
+        factory.createLongCounter("udpFinalCheckResponsesReceived",
+            udpFinalCheckResponsesReceivedDesc,
             "messages"),});
 
     // Initialize id fields
diff --git a/geode-core/src/main/java/org/apache/geode/distributed/internal/InternalDistributedSystem.java b/geode-core/src/main/java/org/apache/geode/distributed/internal/InternalDistributedSystem.java
index 23d81d7..9bce277 100644
--- a/geode-core/src/main/java/org/apache/geode/distributed/internal/InternalDistributedSystem.java
+++ b/geode-core/src/main/java/org/apache/geode/distributed/internal/InternalDistributedSystem.java
@@ -22,7 +22,6 @@ import java.io.File;
 import java.io.IOException;
 import java.lang.reflect.Array;
 import java.net.InetAddress;
-import java.time.Duration;
 import java.util.ArrayList;
 import java.util.Date;
 import java.util.HashMap;
@@ -40,6 +39,7 @@ import java.util.TreeSet;
 import java.util.concurrent.CopyOnWriteArrayList;
 import java.util.concurrent.TimeUnit;
 import java.util.concurrent.atomic.AtomicReference;
+
 import org.apache.logging.log4j.Logger;
 
 import org.apache.geode.CancelCriterion;
@@ -48,9 +48,6 @@ import org.apache.geode.ForcedDisconnectException;
 import org.apache.geode.GemFireConfigException;
 import org.apache.geode.GemFireIOException;
 import org.apache.geode.LogWriter;
-import org.apache.geode.internal.statistics.InternalDistributedSystemStats;
-import org.apache.geode.statistics.Statistics;
-import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.SystemConnectException;
 import org.apache.geode.SystemFailure;
 import org.apache.geode.admin.AlertLevel;
@@ -98,19 +95,21 @@ import org.apache.geode.internal.offheap.MemoryAllocator;
 import org.apache.geode.internal.offheap.OffHeapStorage;
 import org.apache.geode.internal.security.SecurityService;
 import org.apache.geode.internal.security.SecurityServiceFactory;
-import org.apache.geode.internal.statistics.GemFireStatSampler;
-import org.apache.geode.internal.statistics.StatisticsTypeFactoryImpl;
+import org.apache.geode.internal.statistics.InternalDistributedSystemStats;
 import org.apache.geode.internal.statistics.platform.LinuxProcFsStatistics;
 import org.apache.geode.internal.tcp.ConnectionTable;
 import org.apache.geode.management.ManagementException;
 import org.apache.geode.security.GemFireSecurityException;
 import org.apache.geode.security.PostProcessor;
 import org.apache.geode.security.SecurityManager;
+import org.apache.geode.statistics.Statistics;
+import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.micrometer.MicrometerStatisticsFactoryImpl;
 
 /**
  * The concrete implementation of {@link DistributedSystem} that provides internal-only
  * functionality.
+ *
  * @since GemFire 3.0
  */
 public class InternalDistributedSystem extends DistributedSystem {
@@ -326,7 +325,7 @@ public class InternalDistributedSystem extends DistributedSystem {
   }
 
   public static InternalDistributedSystem newInstance(Properties config,
-                                                      SecurityConfig securityConfig) {
+      SecurityConfig securityConfig) {
     if (securityConfig == null) {
       return newInstance(config, null, null);
     } else {
@@ -336,8 +335,8 @@ public class InternalDistributedSystem extends DistributedSystem {
   }
 
   public static InternalDistributedSystem newInstance(Properties config,
-                                                      SecurityManager securityManager,
-                                                      PostProcessor postProcessor) {
+      SecurityManager securityManager,
+      PostProcessor postProcessor) {
     boolean success = false;
     InternalDataSerializer.checkSerializationVersion();
     try {
@@ -359,10 +358,11 @@ public class InternalDistributedSystem extends DistributedSystem {
 
   /**
    * creates a non-functional instance for testing
+   *
    * @param nonDefault - non-default distributed system properties
    */
   public static InternalDistributedSystem newInstanceForTesting(DistributionManager dm,
-                                                                Properties nonDefault) {
+      Properties nonDefault) {
     InternalDistributedSystem sys = new InternalDistributedSystem(nonDefault);
     sys.config = new RuntimeDistributionConfigImpl(sys);
     sys.dm = dm;
@@ -377,6 +377,7 @@ public class InternalDistributedSystem extends DistributedSystem {
   /**
    * Returns a connection to the distributed system that is suitable for administration. For
    * administration, we are not as strict when it comes to existing connections.
+   *
    * @since GemFire 4.0
    */
   public static DistributedSystem connectForAdmin(Properties props) {
@@ -389,6 +390,7 @@ public class InternalDistributedSystem extends DistributedSystem {
    * <p>
    * <p>
    * author bruce
+   *
    * @since GemFire 5.0
    */
   public static InternalDistributedSystem getConnectedInstance() {
@@ -410,6 +412,7 @@ public class InternalDistributedSystem extends DistributedSystem {
    * size existingSystems uses copy-on-write.
    * <p>
    * author bruce
+   *
    * @since GemFire 5.0
    */
   public static InternalDistributedSystem unsafeGetConnectedInstance() {
@@ -485,6 +488,7 @@ public class InternalDistributedSystem extends DistributedSystem {
    * Creates a new <code>InternalDistributedSystem</code> with the given configuration properties.
    * Does all of the magic of finding the "default" values of properties. See {@link
    * DistributedSystem#connect} for a list of exceptions that may be thrown.
+   *
    * @param nonDefault The non-default configuration properties specified by the caller
    * @see DistributedSystem#connect
    */
@@ -541,6 +545,7 @@ public class InternalDistributedSystem extends DistributedSystem {
 
   /**
    * Registers a listener to the system
+   *
    * @param listener listener to be added
    */
   public void addResourceListener(ResourceEventsListener listener) {
@@ -549,6 +554,7 @@ public class InternalDistributedSystem extends DistributedSystem {
 
   /**
    * Un-Registers a listener to the system
+   *
    * @param listener listener to be removed
    */
   public void removeResourceListener(ResourceEventsListener listener) {
@@ -564,6 +570,7 @@ public class InternalDistributedSystem extends DistributedSystem {
 
   /**
    * Handles a particular event associated with a resource
+   *
    * @param event Resource event
    * @param resource resource on which event is generated
    */
@@ -604,9 +611,10 @@ public class InternalDistributedSystem extends DistributedSystem {
   }
 
   private void initializeStats() {
-    this.internalDistributedSystemStats = InternalDistributedSystemStats.createInstance(this.statsDisabled,this.getConfig(),
-//        this,new StatisticsTypeFactoryImpl());
-        this,new MicrometerStatisticsFactoryImpl());
+    this.internalDistributedSystemStats =
+        InternalDistributedSystemStats.createInstance(this.statsDisabled, this.getConfig(),
+            // this,new StatisticsTypeFactoryImpl());
+            this, new MicrometerStatisticsFactoryImpl());
   }
 
 
@@ -905,8 +913,9 @@ public class InternalDistributedSystem extends DistributedSystem {
 
   /**
    * Checks whether or not this connection to a distributed system is closed.
+   *
    * @throws DistributedSystemDisconnectedException This connection has been {@link
-   * #disconnect(boolean, String, boolean) disconnected}
+   *         #disconnect(boolean, String, boolean) disconnected}
    */
   private void checkConnected() {
     if (!isConnected()) {
@@ -1013,6 +1022,7 @@ public class InternalDistributedSystem extends DistributedSystem {
   /**
    * Disconnects this member from the distributed system when an internal error has caused
    * distribution to fail (e.g., this member was shunned)
+   *
    * @param reason a string describing why the disconnect is occurring
    * @param cause an optional exception showing the reason for abnormal disconnect
    * @param shunned whether this member was shunned by the membership coordinator
@@ -1041,6 +1051,7 @@ public class InternalDistributedSystem extends DistributedSystem {
   /**
    * Run a disconnect listener, checking for errors and honoring the timeout {@link
    * #MAX_DISCONNECT_WAIT}.
+   *
    * @param dc the listener to run
    */
   private void runDisconnect(final DisconnectListener dc, ThreadGroup tg) {
@@ -1102,6 +1113,7 @@ public class InternalDistributedSystem extends DistributedSystem {
 
   /**
    * Run a disconnect listener in the same thread sequence as the reconnect.
+   *
    * @param dc the listener to run
    * @param tg the thread group to run the listener in
    */
@@ -1125,6 +1137,7 @@ public class InternalDistributedSystem extends DistributedSystem {
 
   /**
    * Disconnect cache, run disconnect listeners.
+   *
    * @param doReconnect whether a reconnect will be done
    * @param reason the reason that the system is disconnecting
    * @return a collection of shutdownListeners
@@ -1133,7 +1146,7 @@ public class InternalDistributedSystem extends DistributedSystem {
     // Make a pass over the disconnect listeners, asking them _politely_
     // to clean up.
     HashSet shutdownListeners = new HashSet();
-    for (; ; ) {
+    for (;;) {
       DisconnectListener listener = null;
       synchronized (this.listeners) {
         Iterator itr = listeners.iterator();
@@ -1159,6 +1172,7 @@ public class InternalDistributedSystem extends DistributedSystem {
   /**
    * Process the shutdown listeners. It is essential that the DM has been shut down before calling
    * this step, to ensure that no new listeners are registering.
+   *
    * @param shutdownListeners shutdown listeners initially registered with us
    */
   private void doShutdownListeners(HashSet shutdownListeners) {
@@ -1194,7 +1208,7 @@ public class InternalDistributedSystem extends DistributedSystem {
     // disconnect listeners may have appeared. After messagingDisabled is
     // set, no new ones will be created. However, we must process any
     // that appeared in the interim.
-    for (; ; ) {
+    for (;;) {
       // Pluck next listener from the list
       DisconnectListener dcListener = null;
       ShutdownListener sdListener = null;
@@ -1244,6 +1258,7 @@ public class InternalDistributedSystem extends DistributedSystem {
 
   /**
    * Ensure that the MembershipManager class gets loaded.
+   *
    * @see SystemFailure#loadEmergencyClasses()
    */
   public static void loadEmergencyClasses() {
@@ -1256,6 +1271,7 @@ public class InternalDistributedSystem extends DistributedSystem {
 
   /**
    * Closes the membership manager
+   *
    * @see SystemFailure#emergencyClose()
    */
   public void emergencyClose() {
@@ -1314,6 +1330,7 @@ public class InternalDistributedSystem extends DistributedSystem {
 
   /**
    * Disconnects this VM from the distributed system. Shuts down the distribution manager.
+   *
    * @param preparingForReconnect true if called by a reconnect operation
    * @param reason the reason the disconnect is being performed
    * @param keepAlive true if user requested durable subscriptions are to be retained at server.
@@ -1488,6 +1505,7 @@ public class InternalDistributedSystem extends DistributedSystem {
    * If this DistributedSystem is attempting to reconnect to the distributed system this will return
    * the quorum checker created by the old MembershipManager for checking to see if a quorum of old
    * members can be reached.
+   *
    * @return the quorum checking service
    */
   public QuorumChecker getQuorumChecker() {
@@ -1504,6 +1522,7 @@ public class InternalDistributedSystem extends DistributedSystem {
   /**
    * Returns whether or not this distributed system has the same configuration as the given set of
    * properties.
+   *
    * @see DistributedSystem#connect
    */
   public boolean sameAs(Properties props, boolean isConnected) {
@@ -1523,6 +1542,7 @@ public class InternalDistributedSystem extends DistributedSystem {
   /**
    * Returns whether or not the given configuration properties refer to the same distributed system
    * as this <code>InternalDistributedSystem</code> connection.
+   *
    * @since GemFire 4.0
    */
   public boolean sameSystemAs(Properties props) {
@@ -1551,6 +1571,7 @@ public class InternalDistributedSystem extends DistributedSystem {
 
   /**
    * Canonicalizes a locators string so that they may be compared.
+   *
    * @since GemFire 4.0
    */
   private static String canonicalizeLocators(String locators) {
@@ -1573,7 +1594,7 @@ public class InternalDistributedSystem extends DistributedSystem {
     }
 
     StringBuilder sb = new StringBuilder();
-    for (Iterator iter = sorted.iterator(); iter.hasNext(); ) {
+    for (Iterator iter = sorted.iterator(); iter.hasNext();) {
       sb.append((String) iter.next());
       if (iter.hasNext()) {
         sb.append(",");
@@ -1670,6 +1691,7 @@ public class InternalDistributedSystem extends DistributedSystem {
    * <code>DistributedSystem</code> for each configuration, we can use the default implementation
    * of
    * <code>equals</code>.
+   *
    * @see #sameAs
    */
   @Override
@@ -1739,6 +1761,7 @@ public class InternalDistributedSystem extends DistributedSystem {
 
   /**
    * Used to "visit" each instance of Statistics registered with
+   *
    * @see #visitStatistics
    */
   public interface StatisticsVisitor {
@@ -1753,6 +1776,7 @@ public class InternalDistributedSystem extends DistributedSystem {
   /**
    * Makes note of a <code>ConnectListener</code> whose <code>onConnect</code> method will be
    * invoked when a connection is created to a distributed system.
+   *
    * @return set of currently existing system connections
    */
   public static List addConnectListener(ConnectListener listener) {
@@ -1784,6 +1808,7 @@ public class InternalDistributedSystem extends DistributedSystem {
   /**
    * Removes a <code>ConnectListener</code> from the list of listeners that will be notified when a
    * connection is created to a distributed system.
+   *
    * @return true if listener was in the list
    */
   public static boolean removeConnectListener(ConnectListener listener) {
@@ -1798,7 +1823,7 @@ public class InternalDistributedSystem extends DistributedSystem {
    */
   private static void notifyConnectListeners(InternalDistributedSystem sys) {
     synchronized (connectListeners) {
-      for (Iterator iter = connectListeners.iterator(); iter.hasNext(); ) {
+      for (Iterator iter = connectListeners.iterator(); iter.hasNext();) {
         try {
           ConnectListener listener = (ConnectListener) iter.next();
           listener.onConnect(sys);
@@ -1836,7 +1861,7 @@ public class InternalDistributedSystem extends DistributedSystem {
    * system has been recreated.
    */
   private static void notifyReconnectListeners(InternalDistributedSystem oldsys,
-                                               InternalDistributedSystem newsys, boolean starting) {
+      InternalDistributedSystem newsys, boolean starting) {
     List<ReconnectListener> listeners;
     synchronized (reconnectListeners) {
       listeners = new ArrayList<ReconnectListener>(reconnectListeners);
@@ -1872,6 +1897,7 @@ public class InternalDistributedSystem extends DistributedSystem {
   /**
    * Notifies all resource event listeners. All exceptions are caught here and only a warning
    * message is printed in the log
+   *
    * @param event Enumeration depicting particular resource event
    * @param resource the actual resource object.
    */
@@ -1937,6 +1963,7 @@ public class InternalDistributedSystem extends DistributedSystem {
   /**
    * Removes a <code>DisconnectListener</code> from the list of listeners that will be notified when
    * this connection to the distributed system is disconnected.
+   *
    * @return true if listener was in the list
    */
   public boolean removeDisconnectListener(DisconnectListener listener) {
@@ -1977,6 +2004,7 @@ public class InternalDistributedSystem extends DistributedSystem {
 
   /**
    * Fires an "informational" <code>SystemMembershipEvent</code> in admin VMs.
+   *
    * @since GemFire 4.0
    */
   public void fireInfoEvent(Object callback) {
@@ -2042,6 +2070,7 @@ public class InternalDistributedSystem extends DistributedSystem {
 
     /**
      * Invoked before a connection to the distributed system is disconnected.
+     *
      * @param sys the the system we are disconnecting from process should take before returning.
      */
     void onDisconnect(InternalDistributedSystem sys);
@@ -2055,13 +2084,15 @@ public class InternalDistributedSystem extends DistributedSystem {
 
     /**
      * Invoked when reconnect attempts are initiated
+     *
      * @param oldSystem the old DS, which is in a partially disconnected state and cannot be used
-     * for messaging
+     *        for messaging
      */
     void reconnecting(InternalDistributedSystem oldSystem);
 
     /**
      * Invoked after a reconnect to the distributed system
+     *
      * @param oldSystem the old DS
      * @param newSystem the new DS
      */
@@ -2197,6 +2228,7 @@ public class InternalDistributedSystem extends DistributedSystem {
 
   /**
    * Tries to reconnect to the distributed system on role loss if configure to reconnect.
+   *
    * @param oldCache cache that has apparently failed
    */
   public boolean tryReconnect(boolean forcedDisconnect, String reason, InternalCache oldCache) {
@@ -2380,7 +2412,7 @@ public class InternalDistributedSystem extends DistributedSystem {
         }
 
         logger.info(LocalizedMessage.create(LocalizedStrings.DISTRIBUTED_SYSTEM_RECONNECTING,
-            new Object[]{reconnectAttemptCounter}));
+            new Object[] {reconnectAttemptCounter}));
 
         int savNumOfTries = reconnectAttemptCounter;
         try {
@@ -2556,7 +2588,7 @@ public class InternalDistributedSystem extends DistributedSystem {
    * after an auto-reconnect we may need to recreate a cache server and start it
    */
   public void createAndStartCacheServers(List<CacheServerCreation> cacheServerCreation,
-                                         InternalCache cache) {
+      InternalCache cache) {
 
     List<CacheServer> servers = cache.getCacheServers();
 
@@ -2587,6 +2619,7 @@ public class InternalDistributedSystem extends DistributedSystem {
   /**
    * Validates that the configuration provided is the same as the configuration for this
    * InternalDistributedSystem
+   *
    * @param propsToCheck the Properties instance to compare with the existing Properties
    * @throws IllegalStateException when the configuration is not the same other returns
    */
@@ -2691,6 +2724,7 @@ public class InternalDistributedSystem extends DistributedSystem {
 
   /**
    * see {@link org.apache.geode.admin.AdminDistributedSystemFactory}
+   *
    * @since GemFire 5.7
    */
   public static void setEnableAdministrationOnly(boolean adminOnly) {
@@ -2710,6 +2744,7 @@ public class InternalDistributedSystem extends DistributedSystem {
    * distributed system. GemFire distributed system coordinator adjusts each member's time by an
    * offset. This offset for each member is calculated based on Berkeley Time Synchronization
    * algorithm.
+   *
    * @return time in milliseconds.
    */
   public long systemTimeMillis() {
diff --git a/geode-core/src/main/java/org/apache/geode/distributed/internal/InternalLocator.java b/geode-core/src/main/java/org/apache/geode/distributed/internal/InternalLocator.java
index 8ee1ba6..7655f4d 100644
--- a/geode-core/src/main/java/org/apache/geode/distributed/internal/InternalLocator.java
+++ b/geode-core/src/main/java/org/apache/geode/distributed/internal/InternalLocator.java
@@ -95,6 +95,7 @@ import org.apache.geode.statistics.StatsFactory;
  * The startLocator() methods provide a way to start all three services in one call. Otherwise, the
  * services can be started independently {@code  locator = createLocator();
  * locator.startPeerLocation(); locator.startDistributeSystem();}
+ *
  * @since GemFire 4.0
  */
 public class InternalLocator extends Locator implements ConnectListener {
@@ -241,19 +242,20 @@ public class InternalLocator extends Locator implements ConnectListener {
    * Create a locator that listens on a given port. This locator will not have peer or server
    * location services available until they are started by calling startServerLocation or
    * startPeerLocation on the locator object.
+   *
    * @param port the tcp/ip port to listen on
    * @param logFile the file that log messages should be written to
    * @param logger a log writer that should be used (logFile parameter is ignored)
    * @param securityLogger the logger to be used for security related log messages
    * @param distributedSystemProperties optional properties to configure the distributed system
-   * (e.g., mcast addr/port, other locators)
+   *        (e.g., mcast addr/port, other locators)
    * @param startDistributedSystem if true then this locator will also start its own ds
    */
   public static InternalLocator createLocator(int port, File logFile, InternalLogWriter logger,
-                                              InternalLogWriter securityLogger,
-                                              InetAddress bindAddress, String hostnameForClients,
-                                              Properties distributedSystemProperties,
-                                              boolean startDistributedSystem) {
+      InternalLogWriter securityLogger,
+      InetAddress bindAddress, String hostnameForClients,
+      Properties distributedSystemProperties,
+      boolean startDistributedSystem) {
     synchronized (locatorLock) {
       if (hasLocator()) {
         throw new IllegalStateException(
@@ -281,20 +283,21 @@ public class InternalLocator extends Locator implements ConnectListener {
    * Creates a distribution locator that runs in this VM on the given port and bind address.
    * <p>
    * This is for internal use only as it does not create a distributed system unless told to do so.
+   *
    * @param port the tcp/ip port to listen on
    * @param logFile the file that log messages should be written to
    * @param logger a log writer that should be used (logFile parameter is ignored)
    * @param securityLogger the logger to be used for security related log messages
    * @param startDistributedSystem if true, a distributed system is started
    * @param dsProperties optional properties to configure the distributed system (e.g., mcast
-   * addr/port, other locators)
+   *        addr/port, other locators)
    * @param hostnameForClients the name to give to clients for connecting to this locator
    */
   public static InternalLocator startLocator(int port, File logFile, InternalLogWriter logger,
-                                             InternalLogWriter securityLogger,
-                                             InetAddress bindAddress,
-                                             boolean startDistributedSystem,
-                                             Properties dsProperties, String hostnameForClients)
+      InternalLogWriter securityLogger,
+      InetAddress bindAddress,
+      boolean startDistributedSystem,
+      Properties dsProperties, String hostnameForClients)
       throws IOException {
 
     System.setProperty(FORCE_LOCATOR_DM_TYPE, "true");
@@ -306,9 +309,9 @@ public class InternalLocator extends Locator implements ConnectListener {
       newLocator = createLocator(port, logFile, logger, securityLogger, bindAddress,
           hostnameForClients, dsProperties, startDistributedSystem);
 
-//      newLocator.stats =
-//          StatsFactory.createLocatorStatsImpl(null,
-//              (newLocator.getHostnameForClients() + newLocator.getPort()));
+      // newLocator.stats =
+      // StatsFactory.createLocatorStatsImpl(null,
+      // (newLocator.getHostnameForClients() + newLocator.getPort()));
 
       try {
         newLocator.startPeerLocation();
@@ -381,21 +384,22 @@ public class InternalLocator extends Locator implements ConnectListener {
 
   /**
    * Creates a new {@code Locator} with the given port, log file, logger, and bind address.
+   *
    * @param port the tcp/ip port to listen on
    * @param logF the file that log messages should be written to
    * @param logWriter a log writer that should be used (logFile parameter is ignored)
    * @param securityLogWriter the log writer to be used for security related log messages
    * @param hostnameForClients the name to give to clients for connecting to this locator
    * @param distributedSystemProperties optional properties to configure the distributed system
-   * (e.g., mcast addr/port, other locators)
+   *        (e.g., mcast addr/port, other locators)
    * @param cfg the config if being called from a distributed system; otherwise null.
    * @param startDistributedSystem if true locator will start its own distributed system
    */
   private InternalLocator(int port, File logF, InternalLogWriter logWriter,
-                          InternalLogWriter securityLogWriter,
-                          InetAddress bindAddress, String hostnameForClients,
-                          Properties distributedSystemProperties,
-                          DistributionConfigImpl cfg, boolean startDistributedSystem) {
+      InternalLogWriter securityLogWriter,
+      InetAddress bindAddress, String hostnameForClients,
+      Properties distributedSystemProperties,
+      DistributionConfigImpl cfg, boolean startDistributedSystem) {
 
     // TODO: the following three assignments are already done in superclass
     this.logFile = logF;
@@ -509,6 +513,7 @@ public class InternalLocator extends Locator implements ConnectListener {
   /**
    * Start peer location in this locator. If you plan on starting a distributed system later, this
    * method should be called first so that the distributed system can use this locator.
+   *
    * @return returns the port that the locator to which the locator is bound
    * @since GemFire 5.7
    */
@@ -564,14 +569,15 @@ public class InternalLocator extends Locator implements ConnectListener {
    * For backward-compatibility we retain this method
    * <p>
    * TODO: parameters peerLocator and serverLocator and b1 are never used
+   *
    * @deprecated use a form of the method that does not have peerLocator/serverLocator parameters
    */
   @Deprecated
   public static InternalLocator startLocator(int locatorPort, File logFile,
-                                             InternalLogWriter logger, InternalLogWriter logger1,
-                                             InetAddress addr,
-                                             Properties dsProperties, boolean peerLocator,
-                                             boolean serverLocator, String s, boolean b1)
+      InternalLogWriter logger, InternalLogWriter logger1,
+      InetAddress addr,
+      Properties dsProperties, boolean peerLocator,
+      boolean serverLocator, String s, boolean b1)
       throws IOException {
     return startLocator(locatorPort, logFile, logger, logger1, addr, true, dsProperties, s);
   }
@@ -580,6 +586,7 @@ public class InternalLocator extends Locator implements ConnectListener {
    * Start a distributed system whose life cycle is managed by this locator. When the locator is
    * stopped, this distributed system will be disconnected. If a distributed system already exists,
    * this method will have no affect.
+   *
    * @since GemFire 5.7
    */
   private void startDistributedSystem() throws UnknownHostException {
@@ -675,6 +682,7 @@ public class InternalLocator extends Locator implements ConnectListener {
   /**
    * End the initialization of the locator. This method should be called once the location services
    * and distributed system are started.
+   *
    * @param distributedSystem The distributed system to use for the statistics.
    * @since GemFire 5.7
    */
@@ -699,8 +707,9 @@ public class InternalLocator extends Locator implements ConnectListener {
   /**
    * Start server location services in this locator. Server location can only be started once there
    * is a running distributed system.
+   *
    * @param distributedSystem The distributed system which the server location services should use.
-   * If null, the method will try to find an already connected distributed system.
+   *        If null, the method will try to find an already connected distributed system.
    * @since GemFire 5.7
    */
   void startServerLocation(InternalDistributedSystem distributedSystem) throws IOException {
@@ -749,6 +758,7 @@ public class InternalLocator extends Locator implements ConnectListener {
 
   /**
    * Stop this locator
+   *
    * @param stopForReconnect - stopping for distributed system reconnect
    * @param waitForDisconnect - wait up to 60 seconds for the locator to completely stop
    */
@@ -881,6 +891,7 @@ public class InternalLocator extends Locator implements ConnectListener {
 
   /**
    * Waits for a locator to be told to stop.
+   *
    * @throws InterruptedException thrown if the thread is interrupted
    */
   public void waitToStop() throws InterruptedException {
@@ -953,6 +964,7 @@ public class InternalLocator extends Locator implements ConnectListener {
    * succeeds. It will then wait for the system to finish reconnecting before returning. If quorum
    * checks are not being done this merely waits for the distributed system to reconnect and then
    * starts location services.
+   *
    * @return true if able to reconnect the locator to the new distributed system
    */
   private boolean attemptReconnect() throws InterruptedException, IOException {
@@ -1089,6 +1101,7 @@ public class InternalLocator extends Locator implements ConnectListener {
   /**
    * Return the port on which the locator is actually listening. If called before the locator has
    * actually started, this method will return null.
+   *
    * @return the port the locator is listening on or null if it has not yet been started
    */
   @Override
@@ -1165,7 +1178,7 @@ public class InternalLocator extends Locator implements ConnectListener {
 
     @Override
     public void restarting(DistributedSystem ds, GemFireCache cache,
-                           InternalConfigurationPersistenceService sharedConfig) {
+        InternalConfigurationPersistenceService sharedConfig) {
       if (ds != null) {
         for (TcpHandler handler : this.allHandlers) {
           handler.restarting(ds, cache, sharedConfig);
@@ -1271,6 +1284,7 @@ public class InternalLocator extends Locator implements ConnectListener {
   /**
    * Returns collection of locator strings representing every locator instance hosted by this
    * member.
+   *
    * @see #getLocators()
    */
   public static Collection<String> getLocatorStrings() {
diff --git a/geode-core/src/main/java/org/apache/geode/distributed/internal/LocatorStatsImpl.java b/geode-core/src/main/java/org/apache/geode/distributed/internal/LocatorStatsImpl.java
index 31776b9..7c4314c 100755
--- a/geode-core/src/main/java/org/apache/geode/distributed/internal/LocatorStatsImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/distributed/internal/LocatorStatsImpl.java
@@ -70,12 +70,14 @@ public class LocatorStatsImpl implements LocatorStats {
     String serverThreadsDesc =
         "The number of location requests currently being processed by the thread pool.";
     type = factory.createType(statName, statDescription, new StatisticDescriptor[] {
-        factory.createIntGauge(KNOWN_LOCATORS, "Number of locators known to this locator", LOCATORS),
+        factory.createIntGauge(KNOWN_LOCATORS, "Number of locators known to this locator",
+            LOCATORS),
         factory.createLongCounter(REQUESTS_TO_LOCATOR,
             "Number of requests this locator has received from clients", "requests"),
         factory.createLongCounter(RESPONSES_FROM_LOCATOR,
             "Number of responses this locator has sent to clients", "responses"),
-        factory.createIntGauge(ENDPOINTS_KNOWN, "Number of servers this locator knows about", "servers"),
+        factory.createIntGauge(ENDPOINTS_KNOWN, "Number of servers this locator knows about",
+            "servers"),
         factory.createIntGauge(REQUESTS_IN_PROGRESS, serverThreadsDesc, "requests"),
         factory.createLongCounter(REQUEST_TIME, "Time spent processing server location requests",
             "nanoseconds"),
diff --git a/geode-core/src/main/java/org/apache/geode/distributed/internal/locks/DLockService.java b/geode-core/src/main/java/org/apache/geode/distributed/internal/locks/DLockService.java
index f772ad1..cfc39ef 100644
--- a/geode-core/src/main/java/org/apache/geode/distributed/internal/locks/DLockService.java
+++ b/geode-core/src/main/java/org/apache/geode/distributed/internal/locks/DLockService.java
@@ -36,7 +36,6 @@ import org.apache.logging.log4j.Logger;
 import org.apache.geode.CancelCriterion;
 import org.apache.geode.CancelException;
 import org.apache.geode.InternalGemFireException;
-import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.SystemFailure;
 import org.apache.geode.distributed.DistributedLockService;
 import org.apache.geode.distributed.DistributedSystem;
@@ -62,6 +61,7 @@ import org.apache.geode.internal.logging.log4j.LocalizedMessage;
 import org.apache.geode.internal.logging.log4j.LogMarker;
 import org.apache.geode.internal.util.StopWatch;
 import org.apache.geode.internal.util.concurrent.FutureResult;
+import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatsFactory;
 
 /**
@@ -2904,7 +2904,8 @@ public class DLockService extends DistributedLockService {
   }
 
   /** Get or create static dlock stats */
-  protected static synchronized DistributedLockStats getOrCreateStats(DistributedSystem distributedSystem) {
+  protected static synchronized DistributedLockStats getOrCreateStats(
+      DistributedSystem distributedSystem) {
     if (stats == DUMMY_STATS) {
       Assert.assertTrue(distributedSystem != null, "Need an instance of InternalDistributedSystem");
       StatisticsFactory statFactory = distributedSystem.getStatisticsFactory();
diff --git a/geode-core/src/main/java/org/apache/geode/distributed/internal/locks/DLockStatsImpl.java b/geode-core/src/main/java/org/apache/geode/distributed/internal/locks/DLockStatsImpl.java
index 954c9ef..642e934 100755
--- a/geode-core/src/main/java/org/apache/geode/distributed/internal/locks/DLockStatsImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/distributed/internal/locks/DLockStatsImpl.java
@@ -16,12 +16,12 @@
 package org.apache.geode.distributed.internal.locks;
 
 import org.apache.geode.distributed.internal.DistributionStatsImpl;
+import org.apache.geode.distributed.internal.PoolStatHelper;
+import org.apache.geode.distributed.internal.QueueStatHelper;
 import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.distributed.internal.PoolStatHelper;
-import org.apache.geode.distributed.internal.QueueStatHelper;
 
 /**
  * This class maintains statistics in GemFire about the distributed lock service.
@@ -253,42 +253,57 @@ public class DLockStatsImpl implements DistributedLockStats {
         factory.createIntCounter("lockWaitsCompleted", lockWaitsCompletedDesc, "operations"),
         factory.createLongCounter("lockWaitTime", lockWaitTimeDesc, "nanoseconds", false),
         factory.createIntCounter("lockWaitsFailed", lockWaitsFailedDesc, "operations"),
-        factory.createLongCounter("lockWaitFailedTime", lockWaitFailedTimeDesc, "nanoseconds", false),
+        factory.createLongCounter("lockWaitFailedTime", lockWaitFailedTimeDesc, "nanoseconds",
+            false),
         factory.createIntGauge("grantWaitsInProgress", grantWaitsInProgressDesc, "operations"),
         factory.createIntCounter("grantWaitsCompleted", grantWaitsCompletedDesc, "operations"),
         factory.createLongCounter("grantWaitTime", grantWaitTimeDesc, "nanoseconds", false),
         factory.createIntCounter("grantWaitsNotGrantor", grantWaitsNotGrantorDesc, "operations"),
-        factory.createLongCounter("grantWaitNotGrantorTime", grantWaitNotGrantorTimeDesc, "nanoseconds",
+        factory.createLongCounter("grantWaitNotGrantorTime", grantWaitNotGrantorTimeDesc,
+            "nanoseconds",
             false),
         factory.createIntCounter("grantWaitsTimeout", grantWaitsTimeoutDesc, "operations"),
-        factory.createLongCounter("grantWaitTimeoutTime", grantWaitTimeoutTimeDesc, "nanoseconds", false),
+        factory.createLongCounter("grantWaitTimeoutTime", grantWaitTimeoutTimeDesc, "nanoseconds",
+            false),
         factory.createIntCounter("grantWaitsNotHolder", grantWaitsNotHolderDesc, "operations"),
-        factory.createLongCounter("grantWaitNotHolderTime", grantWaitNotHolderTimeDesc, "nanoseconds",
+        factory.createLongCounter("grantWaitNotHolderTime", grantWaitNotHolderTimeDesc,
+            "nanoseconds",
             false),
         factory.createIntCounter("grantWaitsFailed", grantWaitsFailedDesc, "operations"),
-        factory.createLongCounter("grantWaitFailedTime", grantWaitFailedTimeDesc, "nanoseconds", false),
+        factory.createLongCounter("grantWaitFailedTime", grantWaitFailedTimeDesc, "nanoseconds",
+            false),
         factory.createIntCounter("grantWaitsSuspended", grantWaitsSuspendedDesc, "operations"),
-        factory.createLongCounter("grantWaitSuspendedTime", grantWaitSuspendedTimeDesc, "nanoseconds",
+        factory.createLongCounter("grantWaitSuspendedTime", grantWaitSuspendedTimeDesc,
+            "nanoseconds",
             false),
         factory.createIntCounter("grantWaitsDestroyed", grantWaitsDestroyedDesc, "operations"),
-        factory.createLongCounter("grantWaitDestroyedTime", grantWaitDestroyedTimeDesc, "nanoseconds",
+        factory.createLongCounter("grantWaitDestroyedTime", grantWaitDestroyedTimeDesc,
+            "nanoseconds",
             false),
-        factory.createIntGauge("createGrantorsInProgress", createGrantorsInProgressDesc, "operations"),
-        factory.createIntCounter("createGrantorsCompleted", createGrantorsCompletedDesc, "operations"),
+        factory.createIntGauge("createGrantorsInProgress", createGrantorsInProgressDesc,
+            "operations"),
+        factory.createIntCounter("createGrantorsCompleted", createGrantorsCompletedDesc,
+            "operations"),
         factory.createLongCounter("createGrantorTime", createGrantorTimeDesc, "nanoseconds", false),
-        factory.createIntGauge("serviceCreatesInProgress", serviceCreatesInProgressDesc, "operations"),
-        factory.createIntCounter("serviceCreatesCompleted", serviceCreatesCompletedDesc, "operations"),
-        factory.createLongCounter("serviceCreateLatchTime", serviceCreateLatchTimeDesc, "nanoseconds",
+        factory.createIntGauge("serviceCreatesInProgress", serviceCreatesInProgressDesc,
+            "operations"),
+        factory.createIntCounter("serviceCreatesCompleted", serviceCreatesCompletedDesc,
+            "operations"),
+        factory.createLongCounter("serviceCreateLatchTime", serviceCreateLatchTimeDesc,
+            "nanoseconds",
+            false),
+        factory.createLongCounter("serviceInitLatchTime", serviceInitLatchTimeDesc, "nanoseconds",
             false),
-        factory.createLongCounter("serviceInitLatchTime", serviceInitLatchTimeDesc, "nanoseconds", false),
         factory.createIntGauge("grantorWaitsInProgress", grantorWaitsInProgressDesc, "operations"),
         factory.createIntCounter("grantorWaitsCompleted", grantorWaitsCompletedDesc, "operations"),
         factory.createLongCounter("grantorWaitTime", grantorWaitTimeDesc, "nanoseconds", false),
         factory.createIntCounter("grantorWaitsFailed", grantorWaitsFailedDesc, "operations"),
         factory.createLongCounter("grantorWaitFailedTime", grantorWaitFailedTimeDesc, "nanoseconds",
             false),
-        factory.createIntGauge("grantorThreadsInProgress", grantorThreadsInProgressDesc, "operations"),
-        factory.createIntCounter("grantorThreadsCompleted", grantorThreadsCompletedDesc, "operations"),
+        factory.createIntGauge("grantorThreadsInProgress", grantorThreadsInProgressDesc,
+            "operations"),
+        factory.createIntCounter("grantorThreadsCompleted", grantorThreadsCompletedDesc,
+            "operations"),
         factory.createLongCounter("grantorThreadExpireAndGrantLocksTime",
             grantorThreadExpireAndGrantLocksTimeDesc, "nanoseconds", false),
         factory.createLongCounter("grantorThreadHandleRequestTimeoutsTime",
@@ -301,16 +316,20 @@ public class DLockStatsImpl implements DistributedLockStats {
             "operations"),
         factory.createIntCounter("destroyReadWaitsCompleted", destroyReadWaitsCompletedDesc,
             "operations"),
-        factory.createLongCounter("destroyReadWaitTime", destroyReadWaitTimeDesc, "nanoseconds", false),
-        factory.createIntCounter("destroyReadWaitsFailed", destroyReadWaitsFailedDesc, "operations"),
+        factory.createLongCounter("destroyReadWaitTime", destroyReadWaitTimeDesc, "nanoseconds",
+            false),
+        factory.createIntCounter("destroyReadWaitsFailed", destroyReadWaitsFailedDesc,
+            "operations"),
         factory.createLongCounter("destroyReadWaitFailedTime", destroyReadWaitFailedTimeDesc,
             "nanoseconds", false),
         factory.createIntGauge("destroyWriteWaitsInProgress", destroyWriteWaitsInProgressDesc,
             "operations"),
         factory.createIntCounter("destroyWriteWaitsCompleted", destroyWriteWaitsCompletedDesc,
             "operations"),
-        factory.createLongCounter("destroyWriteWaitTime", destroyWriteWaitTimeDesc, "nanoseconds", false),
-        factory.createIntCounter("destroyWriteWaitsFailed", destroyWriteWaitsFailedDesc, "operations"),
+        factory.createLongCounter("destroyWriteWaitTime", destroyWriteWaitTimeDesc, "nanoseconds",
+            false),
+        factory.createIntCounter("destroyWriteWaitsFailed", destroyWriteWaitsFailedDesc,
+            "operations"),
         factory.createLongCounter("destroyWriteWaitFailedTime", destroyWriteWaitFailedTimeDesc,
             "nanoseconds", false),
         factory.createIntGauge("destroyReads", destroyReadsDesc, "destroyReads"),
@@ -319,7 +338,8 @@ public class DLockStatsImpl implements DistributedLockStats {
         factory.createIntCounter("lockReleasesCompleted", lockReleasesCompletedDesc, "operations"),
         factory.createLongCounter("lockReleaseTime", lockReleaseTimeDesc, "nanoseconds", false),
         factory.createIntCounter("becomeGrantorRequests", becomeGrantorRequestsDesc, "operations"),
-        factory.createIntCounter("freeResourcesCompleted", freeResourcesCompletedDesc, "operations"),
+        factory.createIntCounter("freeResourcesCompleted", freeResourcesCompletedDesc,
+            "operations"),
         factory.createIntCounter("freeResourcesFailed", freeResourcesFailedDesc, "operations"),});
 
     // Initialize id fields
diff --git a/geode-core/src/main/java/org/apache/geode/distributed/internal/membership/gms/messenger/JGroupsMessenger.java b/geode-core/src/main/java/org/apache/geode/distributed/internal/membership/gms/messenger/JGroupsMessenger.java
index c7b8942..c6bb653 100644
--- a/geode-core/src/main/java/org/apache/geode/distributed/internal/membership/gms/messenger/JGroupsMessenger.java
+++ b/geode-core/src/main/java/org/apache/geode/distributed/internal/membership/gms/messenger/JGroupsMessenger.java
@@ -77,7 +77,6 @@ import org.apache.geode.distributed.internal.ClusterDistributionManager;
 import org.apache.geode.distributed.internal.DMStats;
 import org.apache.geode.distributed.internal.DistributionConfig;
 import org.apache.geode.distributed.internal.DistributionMessage;
-import org.apache.geode.distributed.internal.DistributionStats;
 import org.apache.geode.distributed.internal.HighPriorityDistributionMessage;
 import org.apache.geode.distributed.internal.membership.InternalDistributedMember;
 import org.apache.geode.distributed.internal.membership.MemberAttributes;
diff --git a/geode-core/src/main/java/org/apache/geode/distributed/internal/tcpserver/TcpServer.java b/geode-core/src/main/java/org/apache/geode/distributed/internal/tcpserver/TcpServer.java
index 3f89174..de8e90c 100755
--- a/geode-core/src/main/java/org/apache/geode/distributed/internal/tcpserver/TcpServer.java
+++ b/geode-core/src/main/java/org/apache/geode/distributed/internal/tcpserver/TcpServer.java
@@ -46,7 +46,6 @@ import org.apache.geode.cache.IncompatibleVersionException;
 import org.apache.geode.cache.UnsupportedVersionException;
 import org.apache.geode.distributed.internal.DistributionConfig;
 import org.apache.geode.distributed.internal.DistributionConfigImpl;
-import org.apache.geode.distributed.internal.DistributionStats;
 import org.apache.geode.distributed.internal.InternalConfigurationPersistenceService;
 import org.apache.geode.distributed.internal.InternalDistributedSystem;
 import org.apache.geode.distributed.internal.InternalLocator;
diff --git a/geode-core/src/main/java/org/apache/geode/internal/admin/ClientStatsManager.java b/geode-core/src/main/java/org/apache/geode/internal/admin/ClientStatsManager.java
index 098d322..b91636c 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/admin/ClientStatsManager.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/admin/ClientStatsManager.java
@@ -20,9 +20,6 @@ import java.util.Map;
 import org.apache.commons.lang.exception.ExceptionUtils;
 import org.apache.logging.log4j.Logger;
 
-import org.apache.geode.internal.statistics.InternalDistributedSystemStats;
-import org.apache.geode.statistics.Statistics;
-import org.apache.geode.statistics.StatisticsType;
 import org.apache.geode.cache.CacheWriterException;
 import org.apache.geode.cache.client.internal.PoolImpl;
 import org.apache.geode.cache.client.internal.ServerRegionProxy;
@@ -36,6 +33,9 @@ import org.apache.geode.internal.cache.InternalCache;
 import org.apache.geode.internal.i18n.LocalizedStrings;
 import org.apache.geode.internal.logging.LogService;
 import org.apache.geode.internal.offheap.annotations.Released;
+import org.apache.geode.internal.statistics.InternalDistributedSystemStats;
+import org.apache.geode.statistics.Statistics;
+import org.apache.geode.statistics.StatisticsType;
 
 /**
  * This class publishes the client statistics using the admin region.
@@ -128,9 +128,9 @@ public class ClientStatsManager {
       return false;
     }
 
-    InternalDistributedSystem distributedSystem = (InternalDistributedSystem) currentCache.getDistributedSystem();
-    InternalDistributedSystemStats
-        internalDistributedSystemStats =
+    InternalDistributedSystem distributedSystem =
+        (InternalDistributedSystem) currentCache.getDistributedSystem();
+    InternalDistributedSystemStats internalDistributedSystemStats =
         distributedSystem.getInternalDistributedSystemStats();
     if (currentCache.isClosed()) {
       return false;
diff --git a/geode-core/src/main/java/org/apache/geode/internal/admin/StatAlertDefinition.java b/geode-core/src/main/java/org/apache/geode/internal/admin/StatAlertDefinition.java
index 34540a2..103802d 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/admin/StatAlertDefinition.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/admin/StatAlertDefinition.java
@@ -15,8 +15,8 @@
 package org.apache.geode.internal.admin;
 
 import org.apache.geode.DataSerializable;
-import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.internal.admin.statalerts.StatisticInfo;
+import org.apache.geode.statistics.StatisticsFactory;
 
 /**
  * Base interface that supports a StatAlertDefinition
diff --git a/geode-core/src/main/java/org/apache/geode/internal/admin/StatAlertsManager.java b/geode-core/src/main/java/org/apache/geode/internal/admin/StatAlertsManager.java
index deb623a..8a8c69c 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/admin/StatAlertsManager.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/admin/StatAlertsManager.java
@@ -26,9 +26,6 @@ import java.util.TimerTask;
 import org.apache.logging.log4j.Logger;
 
 import org.apache.geode.CancelException;
-import org.apache.geode.statistics.StatisticDescriptor;
-import org.apache.geode.statistics.Statistics;
-import org.apache.geode.statistics.StatisticsType;
 import org.apache.geode.admin.jmx.internal.StatAlertsAggregator;
 import org.apache.geode.distributed.internal.ClusterDistributionManager;
 import org.apache.geode.distributed.internal.InternalDistributedSystem;
@@ -42,6 +39,9 @@ import org.apache.geode.internal.admin.statalerts.StatisticInfoImpl;
 import org.apache.geode.internal.i18n.LocalizedStrings;
 import org.apache.geode.internal.logging.LogService;
 import org.apache.geode.internal.logging.log4j.LocalizedMessage;
+import org.apache.geode.statistics.StatisticDescriptor;
+import org.apache.geode.statistics.Statistics;
+import org.apache.geode.statistics.StatisticsType;
 
 /**
  * The alert manager maintains the list of alert definitions (added by client e.g GFMon 2.0).
@@ -276,7 +276,8 @@ public class StatAlertsManager {
         textId = statInfos[ii].getStatisticsTextId();
 
         // TODO If none by TextID, use StatType and getAll.
-        statistics = dm.getSystem().getInternalDistributedSystemStats().findStatisticsByTextId(textId);
+        statistics =
+            dm.getSystem().getInternalDistributedSystemStats().findStatisticsByTextId(textId);
         if (statistics.length == 0) {
           logger.error(LocalizedMessage.create(
               LocalizedStrings.StatAlertsManager_STATALERTSMANAGER_CREATEMEMBERSTATALERTDEFINITION_STATISTICS_WITH_GIVEN_TEXTID_0_NOT_FOUND,
diff --git a/geode-core/src/main/java/org/apache/geode/internal/admin/remote/FetchResourceAttributesResponse.java b/geode-core/src/main/java/org/apache/geode/internal/admin/remote/FetchResourceAttributesResponse.java
index 789dc03..3decdf2 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/admin/remote/FetchResourceAttributesResponse.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/admin/remote/FetchResourceAttributesResponse.java
@@ -19,13 +19,12 @@ import java.io.DataOutput;
 import java.io.IOException;
 
 import org.apache.geode.DataSerializer;
+import org.apache.geode.distributed.internal.DistributionManager;
+import org.apache.geode.distributed.internal.membership.InternalDistributedMember;
 import org.apache.geode.internal.statistics.InternalDistributedSystemStats;
 import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.distributed.internal.DistributionManager;
-import org.apache.geode.distributed.internal.InternalDistributedSystem;
-import org.apache.geode.distributed.internal.membership.InternalDistributedMember;
 
 public class FetchResourceAttributesResponse extends AdminResponse {
 
diff --git a/geode-core/src/main/java/org/apache/geode/internal/admin/remote/FetchStatsResponse.java b/geode-core/src/main/java/org/apache/geode/internal/admin/remote/FetchStatsResponse.java
index 07b7d1a..ac606d2 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/admin/remote/FetchStatsResponse.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/admin/remote/FetchStatsResponse.java
@@ -21,9 +21,7 @@ import java.util.ArrayList;
 import java.util.List;
 
 import org.apache.geode.DataSerializer;
-import org.apache.geode.statistics.Statistics;
 import org.apache.geode.distributed.internal.DistributionManager;
-import org.apache.geode.distributed.internal.InternalDistributedSystem.StatisticsVisitor;
 import org.apache.geode.distributed.internal.membership.InternalDistributedMember;
 
 /**
diff --git a/geode-core/src/main/java/org/apache/geode/internal/admin/remote/RemoteStat.java b/geode-core/src/main/java/org/apache/geode/internal/admin/remote/RemoteStat.java
index e2ac540..3e47cab 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/admin/remote/RemoteStat.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/admin/remote/RemoteStat.java
@@ -21,10 +21,10 @@ import java.io.IOException;
 
 import org.apache.geode.DataSerializable;
 import org.apache.geode.DataSerializer;
-import org.apache.geode.statistics.StatisticDescriptor;
-import org.apache.geode.statistics.Statistics;
 import org.apache.geode.internal.admin.Stat;
 import org.apache.geode.internal.statistics.StatisticDescriptorImpl;
+import org.apache.geode.statistics.StatisticDescriptor;
+import org.apache.geode.statistics.Statistics;
 
 public class RemoteStat implements Stat, DataSerializable {
   private static final long serialVersionUID = 8263951659282343027L;
diff --git a/geode-core/src/main/java/org/apache/geode/internal/admin/remote/RemoteStatResource.java b/geode-core/src/main/java/org/apache/geode/internal/admin/remote/RemoteStatResource.java
index 485fca0..95c903e 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/admin/remote/RemoteStatResource.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/admin/remote/RemoteStatResource.java
@@ -21,10 +21,10 @@ import java.io.IOException;
 
 import org.apache.geode.DataSerializable;
 import org.apache.geode.DataSerializer;
-import org.apache.geode.statistics.Statistics;
 import org.apache.geode.internal.admin.GemFireVM;
 import org.apache.geode.internal.admin.Stat;
 import org.apache.geode.internal.admin.StatResource;
+import org.apache.geode.statistics.Statistics;
 
 public class RemoteStatResource implements StatResource, DataSerializable {
   private static final long serialVersionUID = -3118720083415516133L;
diff --git a/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/BaseDecoratorImpl.java b/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/BaseDecoratorImpl.java
index 94c95bc..2c31f59 100755
--- a/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/BaseDecoratorImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/BaseDecoratorImpl.java
@@ -19,10 +19,10 @@ import java.io.DataOutput;
 import java.io.IOException;
 
 import org.apache.geode.DataSerializer;
-import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.SystemFailure;
 import org.apache.geode.internal.admin.StatAlert;
 import org.apache.geode.internal.admin.StatAlertDefinition;
+import org.apache.geode.statistics.StatisticsFactory;
 
 /**
  * Base class for all the decorators
diff --git a/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/DummyStatisticInfoImpl.java b/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/DummyStatisticInfoImpl.java
index 143aea5..8837da4 100755
--- a/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/DummyStatisticInfoImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/DummyStatisticInfoImpl.java
@@ -18,9 +18,9 @@ import java.io.DataInput;
 import java.io.DataOutput;
 import java.io.IOException;
 
+import org.apache.geode.internal.admin.StatAlertsManager;
 import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
-import org.apache.geode.internal.admin.StatAlertsManager;
 
 /**
  * Implementation {@link StatisticInfo} This does not contains associated statistics object
diff --git a/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/FunctionDecoratorImpl.java b/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/FunctionDecoratorImpl.java
index fa7e00f..5b89f53 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/FunctionDecoratorImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/FunctionDecoratorImpl.java
@@ -19,9 +19,9 @@ import java.io.DataOutput;
 import java.io.IOException;
 
 import org.apache.geode.DataSerializer;
-import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.internal.admin.StatAlert;
 import org.apache.geode.internal.admin.StatAlertDefinition;
+import org.apache.geode.statistics.StatisticsFactory;
 
 /**
  * Implementation of {@link StatAlertDefinition}, which defines function Function to be applied on
diff --git a/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/GaugeThresholdDecoratorImpl.java b/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/GaugeThresholdDecoratorImpl.java
index 2b54d54..5e200ba 100755
--- a/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/GaugeThresholdDecoratorImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/GaugeThresholdDecoratorImpl.java
@@ -19,11 +19,11 @@ import java.io.DataOutput;
 import java.io.IOException;
 
 import org.apache.geode.DataSerializer;
-import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.internal.DataSerializableFixedID;
 import org.apache.geode.internal.Version;
 import org.apache.geode.internal.admin.StatAlert;
 import org.apache.geode.internal.admin.StatAlertDefinition;
+import org.apache.geode.statistics.StatisticsFactory;
 
 /**
  * Implementation of {@link StatAlertDefinition}, represents threshold as data range
diff --git a/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/MultiAttrDefinitionImpl.java b/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/MultiAttrDefinitionImpl.java
index 786fafe..73652e8 100755
--- a/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/MultiAttrDefinitionImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/MultiAttrDefinitionImpl.java
@@ -19,11 +19,11 @@ import java.io.DataOutput;
 import java.io.IOException;
 
 import org.apache.geode.DataSerializer;
+import org.apache.geode.internal.admin.StatAlert;
+import org.apache.geode.internal.admin.StatAlertDefinition;
 import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsFactory;
-import org.apache.geode.internal.admin.StatAlert;
-import org.apache.geode.internal.admin.StatAlertDefinition;
 
 /**
  * Implementation of {@link StatAlertDefinition} This provides the definition for multiple statistic
diff --git a/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/NumberThresholdDecoratorImpl.java b/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/NumberThresholdDecoratorImpl.java
index e829462..3c7ccdf 100755
--- a/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/NumberThresholdDecoratorImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/NumberThresholdDecoratorImpl.java
@@ -19,11 +19,11 @@ import java.io.DataOutput;
 import java.io.IOException;
 
 import org.apache.geode.DataSerializer;
-import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.internal.DataSerializableFixedID;
 import org.apache.geode.internal.Version;
 import org.apache.geode.internal.admin.StatAlert;
 import org.apache.geode.internal.admin.StatAlertDefinition;
+import org.apache.geode.statistics.StatisticsFactory;
 
 /**
  * Implementation of {@link StatAlertDefinition}, represents threshold as number
diff --git a/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/SingleAttrDefinitionImpl.java b/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/SingleAttrDefinitionImpl.java
index d162ff0..6fde825 100755
--- a/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/SingleAttrDefinitionImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/SingleAttrDefinitionImpl.java
@@ -19,11 +19,11 @@ import java.io.DataOutput;
 import java.io.IOException;
 
 import org.apache.geode.DataSerializer;
+import org.apache.geode.internal.admin.StatAlert;
+import org.apache.geode.internal.admin.StatAlertDefinition;
 import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsFactory;
-import org.apache.geode.internal.admin.StatAlert;
-import org.apache.geode.internal.admin.StatAlertDefinition;
 
 /**
  * Implementation of {@link StatAlertDefinition} This provides the definition for single statistic
diff --git a/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/StatisticInfoImpl.java b/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/StatisticInfoImpl.java
index 6a8f9cd..48a5bc2 100755
--- a/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/StatisticInfoImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/admin/statalerts/StatisticInfoImpl.java
@@ -20,11 +20,11 @@ import java.io.IOException;
 
 import org.apache.commons.lang.StringUtils;
 
+import org.apache.geode.admin.Statistic;
 import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.admin.Statistic;
 
 /**
  *
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/AbstractDiskRegion.java b/geode-core/src/main/java/org/apache/geode/internal/cache/AbstractDiskRegion.java
index 5963127..c4d945d 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/AbstractDiskRegion.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/AbstractDiskRegion.java
@@ -24,7 +24,6 @@ import java.util.concurrent.atomic.AtomicLong;
 import joptsimple.internal.Strings;
 import org.apache.logging.log4j.Logger;
 
-import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.cache.EvictionAction;
 import org.apache.geode.cache.EvictionAlgorithm;
 import org.apache.geode.cache.EvictionAttributes;
@@ -45,6 +44,7 @@ import org.apache.geode.internal.i18n.LocalizedStrings;
 import org.apache.geode.internal.logging.LogService;
 import org.apache.geode.internal.logging.log4j.LogMarker;
 import org.apache.geode.internal.util.concurrent.ConcurrentMapWithReusableEntries;
+import org.apache.geode.statistics.StatisticsFactory;
 
 /**
  * Code shared by both DiskRegion and RecoveredDiskRegion.
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/AbstractRegion.java b/geode-core/src/main/java/org/apache/geode/internal/cache/AbstractRegion.java
index cd19917..d8b842c 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/AbstractRegion.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/AbstractRegion.java
@@ -32,7 +32,6 @@ import java.util.concurrent.atomic.AtomicLong;
 import org.apache.logging.log4j.Logger;
 
 import org.apache.geode.DataSerializer;
-import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.annotations.TestingOnly;
 import org.apache.geode.cache.AttributesMutator;
 import org.apache.geode.cache.CacheCallback;
@@ -93,6 +92,7 @@ import org.apache.geode.internal.logging.LogService;
 import org.apache.geode.internal.logging.log4j.LocalizedMessage;
 import org.apache.geode.internal.util.ArrayUtils;
 import org.apache.geode.pdx.internal.PeerTypeRegistration;
+import org.apache.geode.statistics.StatisticsFactory;
 
 /**
  * Takes care of RegionAttributes, AttributesMutator, and some no-brainer method implementations.
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/CachePerfStatsImpl.java b/geode-core/src/main/java/org/apache/geode/internal/cache/CachePerfStatsImpl.java
index 90e82ec..f7678a6 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/CachePerfStatsImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/CachePerfStatsImpl.java
@@ -15,14 +15,14 @@
 
 package org.apache.geode.internal.cache;
 
+import org.apache.geode.distributed.internal.PoolStatHelper;
+import org.apache.geode.distributed.internal.QueueStatHelper;
+import org.apache.geode.internal.NanoTimer;
 import org.apache.geode.statistics.GFSStatsImplementor;
 import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.distributed.internal.PoolStatHelper;
-import org.apache.geode.distributed.internal.QueueStatHelper;
-import org.apache.geode.internal.NanoTimer;
 
 /**
  * CachePerfStats tracks statistics about GemFire cache performance.
@@ -165,42 +165,36 @@ public class CachePerfStatsImpl implements CachePerfStats, GFSStatsImplementor {
 
   private static final String loadsInProgressDesc =
       "Current number of threads in this cache doing a cache load.";
-  private static final String
-      loadsCompletedDesc =
+  private static final String loadsCompletedDesc =
       "Total number of times a load on this cache has completed (as a result of either a local get() or a remote netload).";
   private static final String loadTimeDesc = "Total time spent invoking loaders on this cache.";
   private static final String netloadsInProgressDesc =
       "Current number of threads doing a network load initiated by a get() in this cache.";
   private static final String netloadsCompletedDesc =
       "Total number of times a network load initiated on this cache has completed.";
-  private static final String
-      netloadTimeDesc =
+  private static final String netloadTimeDesc =
       "Total time spent doing network loads on this cache.";
   private static final String netsearchesInProgressDesc =
       "Current number of threads doing a network search initiated by a get() in this cache.";
   private static final String netsearchesCompletedDesc =
       "Total number of times network searches initiated by this cache have completed.";
-  private static final String
-      netsearchTimeDesc =
+  private static final String netsearchTimeDesc =
       "Total time spent doing network searches for cache values.";
   private static final String cacheWriterCallsInProgressDesc =
       "Current number of threads doing a cache writer call.";
   private static final String cacheWriterCallsCompletedDesc =
       "Total number of times a cache writer call has completed.";
-  private static final String
-      cacheWriterCallTimeDesc =
+  private static final String cacheWriterCallTimeDesc =
       "Total time spent doing cache writer calls.";
   private static final String cacheListenerCallsInProgressDesc =
       "Current number of threads doing a cache listener call.";
   private static final String cacheListenerCallsCompletedDesc =
       "Total number of times a cache listener call has completed.";
-  private static final String
-      cacheListenerCallTimeDesc =
+  private static final String cacheListenerCallTimeDesc =
       "Total time spent doing cache listener calls.";
   private static final String getInitialImagesInProgressDesc =
       "Current number of getInitialImage operations currently in progress.";
-  private static final String
-      getInitialImagesCompletedDesc =
+  private static final String getInitialImagesCompletedDesc =
       "Total number of times getInitialImages (both delta and full GII) initiated by this cache have completed.";
   private static final String deltaGetInitialImagesCompletedDesc =
       "Total number of times delta getInitialImages initiated by this cache have completed.";
@@ -209,81 +203,61 @@ public class CachePerfStatsImpl implements CachePerfStats, GFSStatsImplementor {
   private static final String getInitialImageKeysReceivedDesc =
       "Total number of keys received while doing getInitialImage operations.";
   private static final String regionsDesc = "The current number of regions in the cache.";
-  private static final String
-      partitionedRegionsDesc =
+  private static final String partitionedRegionsDesc =
       "The current number of partitioned regions in the cache.";
   private static final String destroysDesc =
       "The total number of times a cache object entry has been destroyed in this cache.";
   private static final String updatesDesc =
       "The total number of updates originating remotely that have been applied to this cache.";
   private static final String updateTimeDesc = "Total time spent performing an update.";
-  private static final String
-      invalidatesDesc =
+  private static final String invalidatesDesc =
       "The total number of times an existing cache object entry value in this cache has been invalidated";
   private static final String getsDesc =
       "The total number of times a successful get has been done on this cache.";
-  private static final String
-      createsDesc =
+  private static final String createsDesc =
       "The total number of times an entry is added to this cache.";
-  private static final String
-      putsDesc =
+  private static final String putsDesc =
       "The total number of times an entry is added or replaced in this cache as a result of a local operation (put(), create(), or get() which results in load, netsearch, or netloading a value). Note that this only counts puts done explicitly on this cache. It does not count updates pushed from other caches.";
-  private static final String
-      putTimeDesc =
+  private static final String putTimeDesc =
       "Total time spent adding or replacing an entry in this cache as a result of a local operation.  This includes synchronizing on the map, invoking cache callbacks, sending messages to other caches and waiting for responses (if required).";
-  private static final String
-      putallsDesc =
+  private static final String putallsDesc =
       "The total number of times a map is added or replaced in this cache as a result of a local operation. Note that this only counts putAlls done explicitly on this cache. It does not count updates pushed from other caches.";
-  private static final String
-      putallTimeDesc =
+  private static final String putallTimeDesc =
       "Total time spent replacing a map in this cache as a result of a local operation.  This includes synchronizing on the map, invoking cache callbacks, sending messages to other caches and waiting for responses (if required).";
-  private static final String
-      removeAllsDesc =
+  private static final String removeAllsDesc =
       "The total number of removeAll operations that originated in this cache. Note that this only counts removeAlls done explicitly on this cache. It does not count removes pushed from other caches.";
-  private static final String
-      removeAllTimeDesc =
+  private static final String removeAllTimeDesc =
       "Total time spent performing removeAlls that originated in this cache. This includes time spent waiting for the removeAll to be done in remote caches (if required).";
   private static final String getTimeDesc =
       "Total time spent doing get operations from this cache (including netsearch and netload)";
-  private static final String
-      eventQueueSizeDesc =
+  private static final String eventQueueSizeDesc =
       "The number of cache events waiting to be processed.";
   private static final String eventQueueThrottleTimeDesc =
       "The total amount of time, in nanoseconds, spent delayed by the event queue throttle.";
   private static final String eventQueueThrottleCountDesc =
       "The total number of times a thread was delayed in adding an event to the event queue.";
-  private static final String
-      eventThreadsDesc =
+  private static final String eventThreadsDesc =
       "The number of threads currently processing events.";
-  private static final String
-      missesDesc =
+  private static final String missesDesc =
       "Total number of times a get on the cache did not find a value already in local memory. The number of hits (i.e. gets that did not miss) can be calculated by subtracting misses from gets.";
-  private static final String
-      queryExecutionsDesc =
+  private static final String queryExecutionsDesc =
       "Total number of times some query has been executed";
   private static final String queryExecutionTimeDesc = "Total time spent executing queries";
-  private static final String
-      queryResultsHashCollisionsDesc =
+  private static final String queryResultsHashCollisionsDesc =
       "Total number of times an hash code collision occurred when inserting an object into an OQL result set or rehashing it";
-  private static final String
-      queryResultsHashCollisionProbeTimeDesc =
+  private static final String queryResultsHashCollisionProbeTimeDesc =
       "Total time spent probing the hashtable in an OQL result set due to hash code collisions, includes reads, writes, and rehashes";
   private static final String partitionedRegionOQLQueryRetriesDesc =
       "Total number of times an OQL Query on a Partitioned Region had to be retried";
-  private static final String
-      txSuccessLifeTimeDesc =
+  private static final String txSuccessLifeTimeDesc =
       "The total amount of time, in nanoseconds, spent in a transaction before a successful commit. The time measured starts at transaction begin and ends when commit is called.";
-  private static final String
-      txFailedLifeTimeDesc =
+  private static final String txFailedLifeTimeDesc =
       "The total amount of time, in nanoseconds, spent in a transaction before a failed commit. The time measured starts at transaction begin and ends when commit is called.";
-  private static final String
-      txRollbackLifeTimeDesc =
+  private static final String txRollbackLifeTimeDesc =
       "The total amount of time, in nanoseconds, spent in a transaction before an explicit rollback. The time measured starts at transaction begin and ends when rollback is called.";
-  private static final String
-      txCommitsDesc =
+  private static final String txCommitsDesc =
       "Total number times a transaction commit has succeeded.";
-  private static final String
-      txFailuresDesc =
+  private static final String txFailuresDesc =
       "Total number times a transaction commit has failed.";
   private static final String txRollbacksDesc =
       "Total number times a transaction has been explicitly rolled back.";
@@ -293,49 +267,39 @@ public class CachePerfStatsImpl implements CachePerfStats, GFSStatsImplementor {
       "The total amount of time, in nanoseconds, spent doing failed transaction commits.";
   private static final String txRollbackTimeDesc =
       "The total amount of time, in nanoseconds, spent doing explicit transaction rollbacks.";
-  private static final String
-      txCommitChangesDesc =
+  private static final String txCommitChangesDesc =
       "Total number of changes made by committed transactions.";
-  private static final String
-      txFailureChangesDesc =
+  private static final String txFailureChangesDesc =
       "Total number of changes lost by failed transactions.";
   private static final String txRollbackChangesDesc =
       "Total number of changes lost by explicit transaction rollbacks.";
-  private static final String
-      txConflictCheckTimeDesc =
+  private static final String txConflictCheckTimeDesc =
       "The total amount of time, in nanoseconds, spent doing conflict checks during transaction commit";
   private static final String reliableQueuedOpsDesc =
       "Current number of cache operations queued for distribution to required roles.";
   private static final String reliableQueueSizeDesc =
       "Current size in megabytes of disk used to queue for distribution to required roles.";
-  private static final String
-      reliableQueueMaxDesc =
+  private static final String reliableQueueMaxDesc =
       "Maximum size in megabytes allotted for disk usage to queue for distribution to required roles.";
-  private static final String
-      reliableRegionsDesc =
+  private static final String reliableRegionsDesc =
       "Current number of regions configured for reliability.";
   private static final String reliableRegionsMissingDesc =
       "Current number regions configured for reliability that are missing required roles.";
   private static final String reliableRegionsQueuingDesc =
       "Current number regions configured for reliability that are queuing for required roles.";
-  private static final String
-      reliableRegionsMissingFullAccessDesc =
+  private static final String reliableRegionsMissingFullAccessDesc =
       "Current number of regions configured for reliablity that are missing require roles with full access";
-  private static final String
-      reliableRegionsMissingLimitedAccessDesc =
+  private static final String reliableRegionsMissingLimitedAccessDesc =
       "Current number of regions configured for reliablity that are missing required roles with Limited access";
-  private static final String
-      reliableRegionsMissingNoAccessDesc =
+  private static final String reliableRegionsMissingNoAccessDesc =
       "Current number of regions configured for reliablity that are missing required roles with No access";
-  private static final String
-      clearsDesc =
+  private static final String clearsDesc =
       "The total number of times a clear has been done on this cache.";
   private static final String nonSingleHopsDesc =
       "Total number of times client request observed more than one hop during operation.";
   private static final String metaDataRefreshCountDesc =
       "Total number of times the meta data is refreshed due to hopping observed.";
-  private static final String
-      conflatedEventsDesc =
+  private static final String conflatedEventsDesc =
       "Number of events not delivered due to conflation.  Typically this means that the event arrived after a later event was already applied to the cache.";
   private static final String tombstoneCountDesc =
       "Number of destroyed entries that are retained for concurrent modification detection";
@@ -345,8 +309,7 @@ public class CachePerfStatsImpl implements CachePerfStats, GFSStatsImplementor {
       "Amount of memory consumed by destroyed entries in replicated or partitioned regions";
   private static final String tombstoneOverhead2Desc =
       "Amount of memory consumed by destroyed entries in non-replicated regions";
-  private static final String
-      clearTimeoutsDesc =
+  private static final String clearTimeoutsDesc =
       "Number of timeouts waiting for events concurrent to a clear() operation to be received and applied before performing the clear()";
   private static final String deltaUpdatesDesc =
       "The total number of times entries in this cache are updated through delta bytes.";
@@ -354,14 +317,11 @@ public class CachePerfStatsImpl implements CachePerfStats, GFSStatsImplementor {
       "Total time spent applying the received delta bytes to entries in this cache.";
   private static final String deltaFailedUpdatesDesc =
       "The total number of times entries in this cache failed to be updated through delta bytes.";
-  private static final String
-      deltasPreparedDesc =
+  private static final String deltasPreparedDesc =
       "The total number of times delta was prepared in this cache.";
-  private static final String
-      deltasPreparedTimeDesc =
+  private static final String deltasPreparedTimeDesc =
       "Total time spent preparing delta bytes in this cache.";
-  private static final String
-      deltasSentDesc =
+  private static final String deltasSentDesc =
       "The total number of times delta was sent to remote caches. This excludes deltas sent from server to client.";
   private static final String deltaFullValuesSentDesc =
       "The total number of times a full value was sent to a remote cache.";
@@ -369,54 +329,43 @@ public class CachePerfStatsImpl implements CachePerfStats, GFSStatsImplementor {
       "The total number of times a full value was requested by this cache.";
   private static final String importedEntriesCountDesc =
       "The total number of entries imported from a snapshot file.";
-  private static final String
-      importTimeDesc =
+  private static final String importTimeDesc =
       "The total time spent importing entries from a snapshot file.";
   private static final String exportedEntriesCountDesc =
       "The total number of entries exported into a snapshot file.";
-  private static final String
-      exportTimeDesc =
+  private static final String exportTimeDesc =
       "The total time spent exporting entries into a snapshot file.";
-  private static final String
-      compressionCompressTimeDesc =
+  private static final String compressionCompressTimeDesc =
       "The total time spent compressing data.";
-  private static final String
-      compressionDecompressTimeDesc =
+  private static final String compressionDecompressTimeDesc =
       "The total time spent decompressing data.";
-  private static final String
-      compressionCompressionsDesc =
+  private static final String compressionCompressionsDesc =
       "The total number of compression operations.";
-  private static final String
-      compressionDecompressionsDesc =
+  private static final String compressionDecompressionsDesc =
       "The total number of decompression operations.";
   private static final String compressionPreCompresssedBytesDesc =
       "The total number of bytes before compressing.";
   private static final String compressionPostCompressedBytesDesc =
       "The total number of bytes after compressing.";
-  private static final String
-      evictByCriteria_evictionsDesc =
+  private static final String evictByCriteria_evictionsDesc =
       "The total number of entries evicted";
   // total actual evictions (entries evicted)
-  private static final String
-      evictByCriteria_evictionTimeDesc =
+  private static final String evictByCriteria_evictionTimeDesc =
       "Time taken for eviction process";
   // total eviction time including product +user expr.
-  private static final String
-      evictByCriteria_evictionsInProgressDesc =
+  private static final String evictByCriteria_evictionsInProgressDesc =
       "Total number of evictions in progress";
-  private static final String
-      evictByCriteria_evaluationsDesc =
+  private static final String evictByCriteria_evaluationsDesc =
       "Total number of evaluations for eviction";
   // total eviction attempts
-  private static final String
-      evictByCriteria_evaluationTimeDesc =
+  private static final String evictByCriteria_evaluationTimeDesc =
       "Total time taken for evaluation of user expression during eviction";
-// time taken to evaluate user expression.
+  // time taken to evaluate user expression.
 
   void initializeStats(StatisticsFactory factory, String name) {
     type =
         factory.createType("CachePerfStats-" + name, "Statistics about GemFire cache performance",
-            new StatisticDescriptor[]{
+            new StatisticDescriptor[] {
                 factory.createIntGauge("loadsInProgress", loadsInProgressDesc, "operations"),
                 factory.createIntCounter("loadsCompleted", loadsCompletedDesc, "operations"),
                 factory.createLongCounter("loadTime", loadTimeDesc, "nanoseconds", false),
@@ -1516,6 +1465,7 @@ public class CachePerfStatsImpl implements CachePerfStats, GFSStatsImplementor {
   /**
    * Closes these stats so that they can not longer be used. The stats are closed when the cache is
    * closed.
+   *
    * @since GemFire 3.5
    */
   @Override
@@ -1525,6 +1475,7 @@ public class CachePerfStatsImpl implements CachePerfStats, GFSStatsImplementor {
 
   /**
    * Returns whether or not these stats have been closed
+   *
    * @since GemFire 3.5
    */
   @Override
@@ -1614,6 +1565,7 @@ public class CachePerfStatsImpl implements CachePerfStats, GFSStatsImplementor {
 
   /**
    * Returns the Statistics instance that stores the cache perf stats.
+   *
    * @since GemFire 3.5
    */
   @Override
@@ -1649,6 +1601,7 @@ public class CachePerfStatsImpl implements CachePerfStats, GFSStatsImplementor {
   /**
    * Returns a helper object so that the event pool can record its stats to the proper cache perf
    * stats.
+   *
    * @since GemFire 3.5
    */
   @Override
@@ -1829,11 +1782,11 @@ public class CachePerfStatsImpl implements CachePerfStats, GFSStatsImplementor {
 
   @Override
   public void incDecompressedEntryCount() {
-    stats.setLong(compressionDecompressionsId,1);
+    stats.setLong(compressionDecompressionsId, 1);
   }
 
   @Override
   public void incDecompressionTime(long startTime) {
-    stats.setLong(compressionDecompressTimeId,startTime);
+    stats.setLong(compressionDecompressTimeId, startTime);
   }
-}
\ No newline at end of file
+}
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/DiskDirectoryStatsImpl.java b/geode-core/src/main/java/org/apache/geode/internal/cache/DiskDirectoryStatsImpl.java
index 32615f4..f3d2e9c 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/DiskDirectoryStatsImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/DiskDirectoryStatsImpl.java
@@ -19,8 +19,6 @@ import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.statistics.StatisticsTypeFactory;
-import org.apache.geode.internal.statistics.StatisticsTypeFactoryImpl;
 
 /**
  * GemFire statistics about Disk Directories
@@ -53,12 +51,16 @@ public class DiskDirectoryStatsImpl implements DiskDirectoryStats, GFSStatsImple
     type = factory.createType(statName, statDescription,
         new StatisticDescriptor[] {factory.createLongGauge("diskSpace", diskSpaceDesc, "bytes"),
             factory.createLongGauge("maximumSpace", maxSpaceDesc, "bytes"),
-            factory.createLongGauge("volumeSize", "The total size in bytes of the disk volume", "bytes"),
-            factory.createLongGauge("volumeFreeSpace", "The total free space in bytes on the disk volume",
+            factory.createLongGauge("volumeSize", "The total size in bytes of the disk volume",
                 "bytes"),
-            factory.createLongCounter("volumeFreeSpaceChecks", "The total number of disk space checks",
+            factory.createLongGauge("volumeFreeSpace",
+                "The total free space in bytes on the disk volume",
+                "bytes"),
+            factory.createLongCounter("volumeFreeSpaceChecks",
+                "The total number of disk space checks",
                 "checks"),
-            factory.createLongCounter("volumeFreeSpaceTime", "The total time spent checking disk usage",
+            factory.createLongCounter("volumeFreeSpaceTime",
+                "The total time spent checking disk usage",
                 "nanoseconds")});
 
     // Initialize id fields
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/DiskRegionStatsImpl.java b/geode-core/src/main/java/org/apache/geode/internal/cache/DiskRegionStatsImpl.java
index a3f2a47..3176ef5 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/DiskRegionStatsImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/DiskRegionStatsImpl.java
@@ -19,8 +19,6 @@ import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.statistics.StatisticsTypeFactory;
-import org.apache.geode.internal.statistics.StatisticsTypeFactoryImpl;
 
 /**
  * GemFire statistics about a {@link DiskRegion}.
@@ -90,10 +88,12 @@ public class DiskRegionStatsImpl implements DiskRegionStats, GFSStatsImplementor
         factory.createLongGauge("entriesOnlyOnDisk", numOverflowOnDiskDesc, "entries"),
         factory.createLongGauge("bytesOnlyOnDisk", numOverflowBytesOnDiskDesc, "bytes"),
         factory.createLongGauge("entriesInVM", numEntriesInVMDesc, "entries"),
-        factory.createIntGauge("writesInProgress", "current number of oplog writes that are in progress",
+        factory.createIntGauge("writesInProgress",
+            "current number of oplog writes that are in progress",
             "writes"),
         factory.createIntGauge("localInitializations", localInitializationsDesc, "initializations"),
-        factory.createIntGauge("remoteInitializations", remoteInitializationsDesc, "initializations"),});
+        factory.createIntGauge("remoteInitializations", remoteInitializationsDesc,
+            "initializations"),});
 
     // Initialize id fields
     writesId = type.nameToId("writes");
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/DiskStoreImpl.java b/geode-core/src/main/java/org/apache/geode/internal/cache/DiskStoreImpl.java
index d5e3f79..70391f4 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/DiskStoreImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/DiskStoreImpl.java
@@ -72,7 +72,6 @@ import org.apache.logging.log4j.Logger;
 
 import org.apache.geode.CancelCriterion;
 import org.apache.geode.CancelException;
-import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.SystemFailure;
 import org.apache.geode.cache.Cache;
 import org.apache.geode.cache.CacheClosedException;
@@ -122,6 +121,7 @@ import org.apache.geode.pdx.internal.EnumInfo;
 import org.apache.geode.pdx.internal.PdxField;
 import org.apache.geode.pdx.internal.PdxType;
 import org.apache.geode.pdx.internal.PeerTypeRegistration;
+import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatsFactory;
 
 /**
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/DiskStoreStatsImpl.java b/geode-core/src/main/java/org/apache/geode/internal/cache/DiskStoreStatsImpl.java
index fcd2bf2..b67a6d5 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/DiskStoreStatsImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/DiskStoreStatsImpl.java
@@ -18,9 +18,6 @@ import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.statistics.StatisticsTypeFactory;
-import org.apache.geode.distributed.internal.DistributionStats;
-import org.apache.geode.internal.statistics.StatisticsTypeFactoryImpl;
 
 /**
  * GemFire statistics about a {@link DiskStoreImpl}.
@@ -28,7 +25,7 @@ import org.apache.geode.internal.statistics.StatisticsTypeFactoryImpl;
  *
  * @since GemFire prPersistSprint2
  */
-public class DiskStoreStatsImpl implements DiskStoreStats{
+public class DiskStoreStatsImpl implements DiskStoreStats {
 
   private StatisticsType type;
 
@@ -173,10 +170,13 @@ public class DiskStoreStatsImpl implements DiskStoreStats{
                 "current number of oplog flushes that are in progress", "flushes"),
             factory.createLongCounter("compactTime",
                 "Total amount of time, in nanoseconds, spent compacting oplogs", "nanoseconds"),
-            factory.createIntCounter("compacts", "Total number of completed oplog compacts", "compacts"),
-            factory.createIntGauge("openOplogs", "Current number of oplogs this disk store has open",
+            factory.createIntCounter("compacts", "Total number of completed oplog compacts",
+                "compacts"),
+            factory.createIntGauge("openOplogs",
+                "Current number of oplogs this disk store has open",
                 "oplogs"),
-            factory.createIntGauge("compactableOplogs", "Current number of oplogs ready to be compacted",
+            factory.createIntGauge("compactableOplogs",
+                "Current number of oplogs ready to be compacted",
                 "oplogs"),
             factory.createIntGauge("inactiveOplogs",
                 "Current number of oplogs that are no longer being written but are not ready ready to compact",
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/DistTXRollbackMessage.java b/geode-core/src/main/java/org/apache/geode/internal/cache/DistTXRollbackMessage.java
index 42148eb..d812cba 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/DistTXRollbackMessage.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/DistTXRollbackMessage.java
@@ -36,7 +36,6 @@ import org.apache.geode.distributed.DistributedMember;
 import org.apache.geode.distributed.internal.ClusterDistributionManager;
 import org.apache.geode.distributed.internal.DistributionManager;
 import org.apache.geode.distributed.internal.DistributionMessage;
-import org.apache.geode.distributed.internal.DistributionStats;
 import org.apache.geode.distributed.internal.DistributionStatsImpl;
 import org.apache.geode.distributed.internal.InternalDistributedSystem;
 import org.apache.geode.distributed.internal.ReplyException;
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/DummyCachePerfStats.java b/geode-core/src/main/java/org/apache/geode/internal/cache/DummyCachePerfStats.java
index 29e1663..b50928e 100755
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/DummyCachePerfStats.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/DummyCachePerfStats.java
@@ -14,8 +14,8 @@
  */
 package org.apache.geode.internal.cache;
 
-import org.apache.geode.statistics.Statistics;
 import org.apache.geode.distributed.internal.PoolStatHelper;
+import org.apache.geode.statistics.Statistics;
 
 /**
  * @since GemFire 5.7
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/EvictableRegion.java b/geode-core/src/main/java/org/apache/geode/internal/cache/EvictableRegion.java
index 7d588e0..c7c3ff1 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/EvictableRegion.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/EvictableRegion.java
@@ -14,9 +14,9 @@
  */
 package org.apache.geode.internal.cache;
 
-import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.cache.EvictionAttributes;
 import org.apache.geode.internal.cache.eviction.EvictionController;
+import org.apache.geode.statistics.StatisticsFactory;
 
 public interface EvictableRegion extends RegionMapOwner {
 
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/GemFireCacheImpl.java b/geode-core/src/main/java/org/apache/geode/internal/cache/GemFireCacheImpl.java
index 1d4b331..d6cfcdd 100755
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/GemFireCacheImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/GemFireCacheImpl.java
@@ -866,7 +866,8 @@ public class GemFireCacheImpl implements InternalCache, InternalClientCache, Has
         this.securityService = SecurityServiceFactory.create();
       }
 
-      if (!this.isClient && PoolManager.getAll().isEmpty() && this.getSystem().getConfig().getJmxManager()) {
+      if (!this.isClient && PoolManager.getAll().isEmpty()
+          && this.getSystem().getConfig().getJmxManager()) {
         // We only support management on members of a distributed system
         // Should do this: if (!getSystem().isLoner()) {
         // but it causes quickstart.CqClientTest to hang
@@ -893,7 +894,8 @@ public class GemFireCacheImpl implements InternalCache, InternalClientCache, Has
       this.cqService = CqServiceProvider.create(this);
 
       // Create the CacheStatistics
-      this.cachePerfStats = StatsFactory.createCachePerfStatsImpl(system.getStatisticsFactory(),null);
+      this.cachePerfStats =
+          StatsFactory.createCachePerfStatsImpl(system.getStatisticsFactory(), null);
       CachePerfStatsImpl.enableClockStats = this.system.getConfig().getEnableTimeStatistics();
 
       this.transactionManager = new TXManagerImpl(this.cachePerfStats, this);
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/InternalRegion.java b/geode-core/src/main/java/org/apache/geode/internal/cache/InternalRegion.java
index 898c0d7..07ff6bb 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/InternalRegion.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/InternalRegion.java
@@ -21,8 +21,6 @@ import java.util.Set;
 import java.util.concurrent.locks.Lock;
 
 import org.apache.geode.CancelCriterion;
-import org.apache.geode.internal.cache.eviction.EvictionStats;
-import org.apache.geode.statistics.Statistics;
 import org.apache.geode.cache.CacheWriter;
 import org.apache.geode.cache.CacheWriterException;
 import org.apache.geode.cache.EntryNotFoundException;
@@ -48,6 +46,7 @@ import org.apache.geode.internal.cache.versions.RegionVersionVector;
 import org.apache.geode.internal.cache.versions.VersionSource;
 import org.apache.geode.internal.cache.versions.VersionTag;
 import org.apache.geode.internal.util.concurrent.StoppableCountDownLatch;
+import org.apache.geode.statistics.Statistics;
 
 /**
  * Interface to be used instead of type-casting to LocalRegion.
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/LocalRegion.java b/geode-core/src/main/java/org/apache/geode/internal/cache/LocalRegion.java
index 9c0f2a9..b83fa65 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/LocalRegion.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/LocalRegion.java
@@ -66,7 +66,6 @@ import org.apache.geode.DeltaSerializationException;
 import org.apache.geode.InternalGemFireError;
 import org.apache.geode.InternalGemFireException;
 import org.apache.geode.LogWriter;
-import org.apache.geode.statistics.Statistics;
 import org.apache.geode.SystemFailure;
 import org.apache.geode.admin.internal.SystemMemberCacheEventProcessor;
 import org.apache.geode.cache.AttributesMutator;
@@ -224,6 +223,7 @@ import org.apache.geode.internal.util.concurrent.StoppableCountDownLatch;
 import org.apache.geode.internal.util.concurrent.StoppableReadWriteLock;
 import org.apache.geode.pdx.JSONFormatter;
 import org.apache.geode.pdx.PdxInstance;
+import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatsFactory;
 
 /**
@@ -269,6 +269,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Internal interface used to simulate failures when performing entry operations
+   *
    * @since GemFire 5.7
    */
   public interface TestCallable {
@@ -369,7 +370,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    */
   private final RegionVersionVector versionVector;
 
-  private static final Pattern[] QUERY_PATTERNS = new Pattern[]{
+  private static final Pattern[] QUERY_PATTERNS = new Pattern[] {
       Pattern.compile("^\\(*select .*",
           Pattern.CASE_INSENSITIVE | Pattern.UNICODE_CASE | Pattern.DOTALL),
       Pattern.compile("^import .*",
@@ -381,6 +382,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   /**
    * Used by unit tests to set expiry to milliseconds instead of the default seconds. Used in
    * ExpiryTask.
+   *
    * @since GemFire 5.0
    */
   final boolean EXPIRY_UNITS_MS;
@@ -405,6 +407,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Used to hold off cache listener events until the afterRegionCreate is called
+   *
    * @since GemFire 5.0
    */
   private final StoppableCountDownLatch afterRegionCreateEventLatch;
@@ -604,7 +607,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * Creates new region
    */
   protected LocalRegion(String regionName, RegionAttributes attrs, LocalRegion parentRegion,
-                        InternalCache cache, InternalRegionArguments internalRegionArgs)
+      InternalCache cache, InternalRegionArguments internalRegionArgs)
       throws DiskAccessException {
     super(cache, attrs, regionName, internalRegionArgs);
 
@@ -818,6 +821,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Test hook - returns the version stamp for an entry in the form of a version tag
+   *
    * @return the entry version information
    */
   public VersionTag getVersionTag(Object key) {
@@ -908,6 +912,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Returns the member id of my distributed system
+   *
    * @since GemFire 5.0
    */
   @Override
@@ -927,7 +932,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   // TODO: createSubregion method is too complex for IDE to analyze
   @Override
   public Region createSubregion(String subregionName, RegionAttributes attrs,
-                                InternalRegionArguments internalRegionArgs)
+      InternalRegionArguments internalRegionArgs)
       throws RegionExistsException, TimeoutException, IOException, ClassNotFoundException {
 
     checkReadiness();
@@ -983,7 +988,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
               boolean local = regionAttributes.getScope().isLocal();
               newRegion = local
                   ? new LocalRegion(subregionName, regionAttributes, this, this.cache,
-                  internalRegionArgs)
+                      internalRegionArgs)
                   : new DistributedRegion(subregionName, regionAttributes, this, this.cache,
                       internalRegionArgs);
             }
@@ -1045,7 +1050,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
               newRegion.initialCriticalMembers(
                   this.cache.getInternalResourceManager().getHeapMonitor().getState().isCritical()
                       || this.cache.getInternalResourceManager().getOffHeapMonitor().getState()
-                      .isCritical(),
+                          .isCritical(),
                   this.cache.getResourceAdvisor().adviseCritialMembers());
             }
 
@@ -1122,7 +1127,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
         false, // ifOld
         null, // expectedOldValue
         true // requireOldValue TODO txMerge why is oldValue required for
-        // create? I think so that the EntryExistsException will have it.
+    // create? I think so that the EntryExistsException will have it.
     )) {
       throw new EntryExistsException(event.getKey().toString(), event.getOldValue());
     } else {
@@ -1146,8 +1151,10 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * The default Region implementation will generate EvenTID in the EntryEvent object. This method
-   * is overridden in special Region objects like HARegion or SingleWriteSingleReadRegionQueue.SingleReadWriteMetaRegion
+   * is overridden in special Region objects like HARegion or
+   * SingleWriteSingleReadRegionQueue.SingleReadWriteMetaRegion
    * to return false as the event propagation from those regions do not need EventID objects
+   *
    * @return boolean indicating whether to generate eventID or not
    */
   @Override
@@ -1218,6 +1225,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Fetch the de-serialized value from non-transactional state.
+   *
    * @param keyInfo to which the value is associated
    * @param updateStats true if the entry stats should be updated.
    * @param disableCopyOnRead if true then disable copy on read
@@ -1228,10 +1236,10 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * @return the value for the given key
    */
   public Object getDeserializedValue(RegionEntry regionEntry, final KeyInfo keyInfo,
-                                     final boolean updateStats, boolean disableCopyOnRead,
-                                     boolean preferCachedDeserializable,
-                                     EntryEventImpl clientEvent, boolean returnTombstones,
-                                     boolean retainResult) {
+      final boolean updateStats, boolean disableCopyOnRead,
+      boolean preferCachedDeserializable,
+      EntryEventImpl clientEvent, boolean returnTombstones,
+      boolean retainResult) {
     if (this.diskRegion != null) {
       this.diskRegion.setClearCountReference();
     }
@@ -1290,15 +1298,15 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * @param preferCachedDeserializable true if the preferred result form is CachedDeserializable
    * @param retainResult if true then the result may be a retained off-heap reference
    * @return the value found, which can be
-   * <ul>
-   * <li>null if the value was removed from the region entry
-   * <li>Token.INVALID if the value of the region entry is invalid
-   * <li>Token.LOCAL_INVALID if the value of the region entry is local invalid
-   * </ul>
+   *         <ul>
+   *         <li>null if the value was removed from the region entry
+   *         <li>Token.INVALID if the value of the region entry is invalid
+   *         <li>Token.LOCAL_INVALID if the value of the region entry is local invalid
+   *         </ul>
    */
   @Retained
   Object getDeserialized(RegionEntry regionEntry, boolean updateStats, boolean disableCopyOnRead,
-                         boolean preferCachedDeserializable, boolean retainResult) {
+      boolean preferCachedDeserializable, boolean retainResult) {
     assert !retainResult || preferCachedDeserializable;
     boolean disabledLRUCallback = this.entries.disableLruUpdateCallback();
     try {
@@ -1352,7 +1360,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   @Override
   public Object get(Object key, Object aCallbackArgument, boolean generateCallbacks,
-                    EntryEventImpl clientEvent) throws TimeoutException, CacheLoaderException {
+      EntryEventImpl clientEvent) throws TimeoutException, CacheLoaderException {
     Object result =
         get(key, aCallbackArgument, generateCallbacks, false, false, null, clientEvent, false);
     if (Token.isInvalid(result)) {
@@ -1363,12 +1371,12 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * @see BucketRegion#getSerialized(KeyInfo, boolean, boolean, ClientProxyMembershipID,
-   * EntryEventImpl, boolean)
+   *      EntryEventImpl, boolean)
    */
   public Object get(Object key, Object aCallbackArgument, boolean generateCallbacks,
-                    boolean disableCopyOnRead, boolean preferCD,
-                    ClientProxyMembershipID requestingClient,
-                    EntryEventImpl clientEvent, boolean returnTombstones)
+      boolean disableCopyOnRead, boolean preferCD,
+      ClientProxyMembershipID requestingClient,
+      EntryEventImpl clientEvent, boolean returnTombstones)
       throws TimeoutException, CacheLoaderException {
     return get(key, aCallbackArgument, generateCallbacks, disableCopyOnRead, preferCD,
         requestingClient, clientEvent, returnTombstones, false, false);
@@ -1379,8 +1387,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    */
   @Retained
   public Object getRetained(Object key, Object aCallbackArgument, boolean generateCallbacks,
-                            boolean disableCopyOnRead, ClientProxyMembershipID requestingClient,
-                            EntryEventImpl clientEvent, boolean returnTombstones)
+      boolean disableCopyOnRead, ClientProxyMembershipID requestingClient,
+      EntryEventImpl clientEvent, boolean returnTombstones)
       throws TimeoutException, CacheLoaderException {
     return getRetained(key, aCallbackArgument, generateCallbacks, disableCopyOnRead,
         requestingClient, clientEvent, returnTombstones, false);
@@ -1388,14 +1396,15 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * The result of this operation may be an off-heap reference that the caller must release.
+   *
    * @param opScopeIsLocal if true then just check local storage for a value; if false then try to
-   * find the value if it is not local
+   *        find the value if it is not local
    */
   @Retained
   private Object getRetained(Object key, Object aCallbackArgument, boolean generateCallbacks,
-                             boolean disableCopyOnRead, ClientProxyMembershipID requestingClient,
-                             EntryEventImpl clientEvent, boolean returnTombstones,
-                             boolean opScopeIsLocal)
+      boolean disableCopyOnRead, ClientProxyMembershipID requestingClient,
+      EntryEventImpl clientEvent, boolean returnTombstones,
+      boolean opScopeIsLocal)
       throws TimeoutException, CacheLoaderException {
     return get(key, aCallbackArgument, generateCallbacks, disableCopyOnRead, true, requestingClient,
         clientEvent, returnTombstones, opScopeIsLocal, false /* see GEODE-1291 */);
@@ -1403,14 +1412,14 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * @param opScopeIsLocal if true then just check local storage for a value; if false then try to
-   * find the value if it is not local
+   *        find the value if it is not local
    * @param retainResult if true then the result may be a retained off-heap reference.
    */
   public Object get(Object key, Object aCallbackArgument, boolean generateCallbacks,
-                    boolean disableCopyOnRead, boolean preferCD,
-                    ClientProxyMembershipID requestingClient,
-                    EntryEventImpl clientEvent, boolean returnTombstones, boolean opScopeIsLocal,
-                    boolean retainResult) throws TimeoutException, CacheLoaderException {
+      boolean disableCopyOnRead, boolean preferCD,
+      ClientProxyMembershipID requestingClient,
+      EntryEventImpl clientEvent, boolean returnTombstones, boolean opScopeIsLocal,
+      boolean retainResult) throws TimeoutException, CacheLoaderException {
     assert !retainResult || preferCD;
     validateKey(key);
     checkReadiness();
@@ -1457,6 +1466,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Update region and potentially entry stats for the miss case
+   *
    * @param re optional region entry, fetched if null
    * @param key the key used to fetch the region entry
    */
@@ -1475,6 +1485,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * optimized to only allow one thread to do a search/load, other threads wait on a future
+   *
    * @param isCreate true if call found no entry; false if updating an existing entry
    * @param localValue the value retrieved from the region for this object.
    * @param disableCopyOnRead if true then do not make a copy
@@ -1484,9 +1495,9 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    */
   @Retained
   Object nonTxnFindObject(KeyInfo keyInfo, boolean isCreate, boolean generateCallbacks,
-                          Object localValue, boolean disableCopyOnRead, boolean preferCD,
-                          ClientProxyMembershipID requestingClient, EntryEventImpl clientEvent,
-                          boolean returnTombstones) throws TimeoutException, CacheLoaderException {
+      Object localValue, boolean disableCopyOnRead, boolean preferCD,
+      ClientProxyMembershipID requestingClient, EntryEventImpl clientEvent,
+      boolean returnTombstones) throws TimeoutException, CacheLoaderException {
 
     @Retained
     Object result = null;
@@ -1567,7 +1578,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
     } finally {
       if (result != null) {
         VersionTag tag = clientEvent == null ? null : clientEvent.getVersionTag();
-        thisFuture.set(new Object[]{result, tag});
+        thisFuture.set(new Object[] {result, tag});
       } else {
         thisFuture.set(null);
       }
@@ -1581,6 +1592,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Returns true if get should give a copy; false if a reference.
+   *
    * @since GemFire 4.0
    */
   @Override
@@ -1592,6 +1604,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Makes a copy, if copy-on-get is enabled, of the specified object.
+   *
    * @since GemFire 4.0
    */
   Object conditionalCopy(Object o) {
@@ -1684,7 +1697,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
     // was modified to call the other EntryEventImpl constructor so that
     // an id will be generated by default. Null was passed in anyway.
     // generate EventID
-    @Retained final EntryEventImpl event = EntryEventImpl.create(this, Operation.UPDATE, key, value,
+    @Retained
+    final EntryEventImpl event = EntryEventImpl.create(this, Operation.UPDATE, key, value,
         aCallbackArgument, false, getMyId());
     boolean eventReturned = false;
     try {
@@ -1733,7 +1747,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
           } else if (this instanceof DistributedRegion
               && !((DistributedRegion) this).scope.isDistributedNoAck()
               && !((CacheDistributionAdvisee) this).getCacheDistributionAdvisor().adviseCacheOp()
-              .isEmpty()) {
+                  .isEmpty()) {
             extractDelta = true;
           }
           if (!extractDelta && ClientHealthMonitor.getInstance() != null) {
@@ -1808,6 +1822,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   /**
    * Just like getEntry but also updates the stats that get would have depending on a flag. See bug
    * 42410. Also skips discovering JTA
+   *
    * @return the entry if it exists; otherwise null.
    */
   public Entry accessEntry(Object key, boolean updateStats) {
@@ -1938,6 +1953,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Flavor of keys that will not do repeatable read
+   *
    * @since GemFire 5.5
    */
   public Set testHookKeys() {
@@ -1957,6 +1973,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * return a set of the keys in this region
+   *
    * @param allowTombstones whether destroyed entries should be included
    * @return the keys
    */
@@ -2075,6 +2092,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   /**
    * Convenience method to get region name for logging/exception messages. if this region is an
    * instanceof bucket region, it returns the bucket region name
+   *
    * @return name of the region or the owning partitioned region
    */
   @Override
@@ -2089,6 +2107,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * Returns the number of entries in this region. Note that because of the concurrency properties
    * of the {@link RegionMap}, the number of entries is only an approximate. That is, other threads
    * may change the number of entries in this region while this method is being invoked.
+   *
    * @see RegionMap#size
    */
   public int entryCount() {
@@ -2122,6 +2141,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Returns the {@code DiskRegion} that this region uses to access data on disk.
+   *
    * @return {@code null} if disk regions are not being used
    * @since GemFire 3.2
    */
@@ -2340,13 +2360,14 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * Do any extra initialization required. Region is already visible in parent's subregion map. This
    * method releases the initialization Latches, so subclasses should call this super method last
    * after performing additional initialization.
+   *
    * @param imageTarget ignored, used by subclass for get initial image
    * @see DistributedRegion#initialize(InputStream, InternalDistributedMember,
-   * InternalRegionArguments)
+   *      InternalRegionArguments)
    */
   @Override
   public void initialize(InputStream snapshotInputStream, InternalDistributedMember imageTarget,
-                         InternalRegionArguments internalRegionArgs)
+      InternalRegionArguments internalRegionArgs)
       throws TimeoutException, IOException, ClassNotFoundException {
     if (!isInternalRegion()) {
       // Subclasses may have already called this method, but this is
@@ -2558,6 +2579,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Called after we have delivered our REGION_CREATE event.
+   *
    * @since GemFire 5.0
    */
   private void releaseAfterRegionCreateEventLatch() {
@@ -2566,6 +2588,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Used to cause cache listener events to wait until the after region create event is delivered.
+   *
    * @since GemFire 5.0
    */
   private void waitForRegionCreateEvent() {
@@ -2585,8 +2608,9 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Removes entries and recursively destroys subregions.
+   *
    * @param eventSet collects the events for all destroyed regions if null, then we're closing so
-   * don't send events to callbacks or destroy the disk region
+   *        don't send events to callbacks or destroy the disk region
    */
   private void recursiveDestroyRegion(Set eventSet, RegionEventImpl regionEvent, boolean cacheWrite)
       throws CacheWriterException, TimeoutException {
@@ -2636,7 +2660,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
       // call recursiveDestroyRegion on each subregion and remove it
       // from this subregion map
       Collection values = this.subregions.values();
-      for (Iterator itr = values.iterator(); itr.hasNext(); ) {
+      for (Iterator itr = values.iterator(); itr.hasNext();) {
         // element is a LocalRegion
         Object element = itr.next();
         LocalRegion region;
@@ -2775,6 +2799,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * This method should be called when the caller cannot locate an entry and that condition is
    * unexpected. This will first double check the cache and region state before throwing an
    * EntryNotFoundException. EntryNotFoundException should be a last resort exception.
+   *
    * @param entryKey the missing entry's key.
    */
   @Override
@@ -2789,17 +2814,18 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * Search for the value in a server (if one exists), then try a loader.
    *
    * If we find a value, we put it in the cache.
+   *
    * @param preferCD return the CacheDeserializable, if that's what the value is.
    * @param requestingClient the client making the request, if any
    * @param clientEvent the client's event, if any. If not null, we set the version tag
    * @return the deserialized value
    */
   protected Object findObjectInSystem(KeyInfo keyInfo, boolean isCreate, TXStateInterface tx,
-                                      boolean generateCallbacks, Object localValue,
-                                      boolean disableCopyOnRead, boolean preferCD,
-                                      ClientProxyMembershipID requestingClient,
-                                      EntryEventImpl clientEvent,
-                                      boolean returnTombstones)
+      boolean generateCallbacks, Object localValue,
+      boolean disableCopyOnRead, boolean preferCD,
+      ClientProxyMembershipID requestingClient,
+      EntryEventImpl clientEvent,
+      boolean returnTombstones)
       throws CacheLoaderException, TimeoutException {
 
     final Object key = keyInfo.getKey();
@@ -2917,7 +2943,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   @SuppressWarnings({"rawtypes", "unchecked"})
   protected Object callCacheLoader(CacheLoader loader, final Object key,
-                                   final Object aCallbackArgument, boolean preferCD) {
+      final Object aCallbackArgument, boolean preferCD) {
     LoaderHelper loaderHelper = this.loaderHelperFactory.createLoaderHelper(key, aCallbackArgument,
         false /* netSearchAllowed */, true /* netloadAllowed */, null /* searcher */);
     Object result = loader.load(loaderHelper);
@@ -2933,6 +2959,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * Returns true if the cache already has this key as an invalid entry with a version >= the one in
    * the given event. This is used in cache-miss processing to avoid overwriting the entry when it
    * is not necessary, so that we avoid invoking cache listeners.
+   *
    * @return whether the entry is already invalid
    */
   private boolean alreadyInvalid(Object key, EntryEventImpl event) {
@@ -3089,6 +3116,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Destroy an entry on the server given its event.
+   *
    * @since GemFire 5.7
    */
   void serverDestroy(EntryEventImpl event, Object expectedOldValue) {
@@ -3146,8 +3174,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   @Override
   public void cacheWriteBeforePut(EntryEventImpl event, Set netWriteRecipients,
-                                  CacheWriter localWriter, boolean requireOldValue,
-                                  Object expectedOldValue)
+      CacheWriter localWriter, boolean requireOldValue,
+      Object expectedOldValue)
       throws CacheWriterException, TimeoutException {
     Assert.assertTrue(netWriteRecipients == null);
     Operation operation = event.getOperation();
@@ -3285,7 +3313,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   }
 
   private void scheduleTombstone(RegionEntry entry, VersionTag destroyedVersion,
-                                 boolean reschedule) {
+      boolean reschedule) {
     if (destroyedVersion == null) {
       throw new NullPointerException("destroyed version tag cannot be null");
     }
@@ -3334,11 +3362,12 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * remove any tombstones from the given member that are <= the given version
+   *
    * @param eventID event identifier for the GC operation
    * @param clientRouting routing info (if null a routing is computed)
    */
   public void expireTombstones(Map<VersionSource, Long> regionGCVersions, EventID eventID,
-                               FilterInfo clientRouting) {
+      FilterInfo clientRouting) {
     if (!this.getConcurrencyChecksEnabled()) {
       return;
     }
@@ -3368,12 +3397,13 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * pass tombstone garbage-collection info to clients
+   *
    * @param eventID the ID of the event (see bug #50683)
    * @param routing routing info (routing is computed if this is null)
    */
   protected void notifyClientsOfTombstoneGC(Map<VersionSource, Long> regionGCVersions,
-                                            Set<Object> keysRemoved, EventID eventID,
-                                            FilterInfo routing) {
+      Set<Object> keysRemoved, EventID eventID,
+      FilterInfo routing) {
     if (CacheClientNotifier.getInstance() != null) {
       // Only route the event to clients interested in the partitioned region.
       // We do this by constructing a region-level event and then use it to
@@ -3424,6 +3454,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * validate attributes of subregion being created, sent to parent
+   *
    * @throws IllegalArgumentException if attrs is null
    * @throws IllegalStateException if attributes are invalid
    */
@@ -3445,6 +3476,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * CachedDeserializable}received from another VM, that object will be returned. If the value does
    * not reside in the VM because it has been overflowed to disk, {@code null} will be returned.
    * This method is intended for testing.testing purposes only.
+   *
    * @throws EntryNotFoundException No entry with {@code key} exists
    * @see RegionMap#getEntry
    * @since GemFire 3.2
@@ -3485,8 +3517,9 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * This is a test hook method used to find out what keys the current tx has read or written.
+   *
    * @return an unmodifiable set of keys that have been read or written by the transaction on this
-   * thread.
+   *         thread.
    * @throws IllegalStateException if not tx in progress
    * @since GemFire 5.5
    */
@@ -3511,6 +3544,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * Returns the value of the entry with the given key as it is stored on disk. While the value may
    * be read from disk, it is <b>not</b> stored into the entry in the VM. This method is intended
    * for testing purposes only.
+   *
    * @throws EntryNotFoundException No entry with {@code key} exists
    * @throws IllegalStateException If this region does not write to disk
    * @see RegionEntry#getValueOnDisk
@@ -3543,6 +3577,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * disk. While the value may be read from disk or buffer, it is <b>not</b> stored into the entry
    * in the VM. This is different from getValueonDisk in that it checks for a value both in asynch
    * buffers ( subject to asynch mode enabled) as well as Disk
+   *
    * @throws EntryNotFoundException No entry with {@code key} exists
    * @throws IllegalStateException If this region does not write to disk
    * @see RegionEntry#getValueOnDisk
@@ -3560,6 +3595,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   /**
    * Does a get that attempts to not fault values in from disk or make the entry the most recent in
    * the LRU.
+   *
    * @param adamant fault in and affect LRU as a last resort
    * @param allowTombstone also return Token.TOMBSTONE if the entry is deleted
    * @param serializedFormOkay if the serialized form can be returned
@@ -3740,8 +3776,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   // TODO: this is distressingly similar to code in the client.internal package
   private void processSingleInterest(Object key, int interestType,
-                                     InterestResultPolicy interestResultPolicy, boolean isDurable,
-                                     boolean receiveUpdatesAsInvalidates) {
+      InterestResultPolicy interestResultPolicy, boolean isDurable,
+      boolean receiveUpdatesAsInvalidates) {
     final ServerRegionProxy proxy = getServerProxy();
     if (proxy == null) {
       throw new UnsupportedOperationException(
@@ -3892,7 +3928,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   @Override
   public void registerInterest(Object key, InterestResultPolicy policy, boolean isDurable,
-                               boolean receiveValues) {
+      boolean receiveValues) {
     processSingleInterest(key, InterestType.KEY, policy, isDurable, !receiveValues);
   }
 
@@ -3923,7 +3959,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   @Override
   public void registerInterestRegex(String regex, InterestResultPolicy policy, boolean isDurable,
-                                    boolean receiveValues) {
+      boolean receiveValues) {
     processSingleInterest(regex, InterestType.REGULAR_EXPRESSION, policy, isDurable,
         !receiveValues);
   }
@@ -4009,6 +4045,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   /**
    * finds the keys in this region using the given interestType and argument. Currently only
    * InterestType.REGULAR_EXPRESSION and InterestType.KEY are supported
+   *
    * @param interestType an InterestType value
    * @param interestArg the associated argument (regex string, key or key list, etc)
    * @param allowTombstones whether to return destroyed entries
@@ -4046,7 +4083,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
       if (interestArg instanceof List) {
         ret = new HashSet(); // TODO optimize initial size
         List keyList = (List) interestArg;
-        for (Iterator it = keyList.iterator(); it.hasNext(); ) {
+        for (Iterator it = keyList.iterator(); it.hasNext();) {
           Object entryKey = it.next();
           if (this.containsKey(entryKey) || allowTombstones && this.containsTombstone(entryKey)) {
             ret.add(entryKey);
@@ -4169,6 +4206,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Do localDestroy on a list of keys, if they exist
+   *
    * @param keys the list of arrays of keys to invalidate
    * @see #registerInterest(Object)
    */
@@ -4196,13 +4234,14 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * do a localDestroy on all matching keys
+   *
    * @param key the regular expression to match on
    * @see #registerInterestRegex(String)
    */
   private void clearViaRegEx(String key) {
     // TODO: if (key.equals(".*)) then cmnClearRegionNoCallbacks
     Pattern keyPattern = Pattern.compile(key);
-    for (Iterator it = this.entrySet(false).iterator(); it.hasNext(); ) {
+    for (Iterator it = this.entrySet(false).iterator(); it.hasNext();) {
       Region.Entry entry = (Region.Entry) it.next();
       try {
         Object entryKey = entry.getKey();
@@ -4222,6 +4261,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * do a localDestroy on all matching keys
+   *
    * @param key the regular expression to match on
    */
   private void clearViaFilterClass(String key) {
@@ -4268,7 +4308,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * Refresh local entries based on server's list of keys
    */
   public void refreshEntriesFromServerKeys(Connection con, List serverKeys,
-                                           InterestResultPolicy interestResultPolicy) {
+      InterestResultPolicy interestResultPolicy) {
     if (serverKeys == null) {
       return;
     }
@@ -4436,12 +4476,13 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * Remove values in local cache before registering interest
    *
    * TODO: interestResultPolicy is never used
+   *
    * @param key the interest key
    * @param interestType the interest type from {@link InterestType}
    * @param interestResultPolicy the policy from {@link InterestResultPolicy}
    */
   public void clearKeysOfInterest(Object key, int interestType,
-                                  InterestResultPolicy interestResultPolicy) {
+      InterestResultPolicy interestResultPolicy) {
     switch (interestType) {
       case InterestType.FILTER_CLASS:
         clearViaFilterClass((String) key);
@@ -4473,6 +4514,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * the region will go through regular GetInitialImage if it is a mirrored replicate.
    * <p>
    * Acquires and releases the DestroyLock.
+   *
    * @since GemFire 5.0
    */
   void reinitialize(InputStream inputStream, RegionEventImpl event)
@@ -4571,9 +4613,9 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
       if (snapshotVersion != SNAPSHOT_VERSION) {
         throw new IllegalArgumentException(
             LocalizedStrings.LocalRegion_UNSUPPORTED_SNAPSHOT_VERSION_0_ONLY_VERSION_1_IS_SUPPORTED
-                .toLocalizedString(new Object[]{snapshotVersion, SNAPSHOT_VERSION}));
+                .toLocalizedString(new Object[] {snapshotVersion, SNAPSHOT_VERSION}));
       }
-      for (; ; ) {
+      for (;;) {
         Object key = DataSerializer.readObject(in);
         if (key == null) {
           break;
@@ -4609,6 +4651,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Blocks until initialization is complete.
+   *
    * @param destroyedRegionOk true if it is okay to return a region that isDestroyed
    * @see DestroyRegionOperation
    */
@@ -4810,6 +4853,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   /**
    * Return true if invalidation occurred; false if it did not, for example if it was already
    * invalidated
+   *
    * @see DistributedRegion#basicInvalidate(EntryEventImpl)
    */
   public void basicInvalidate(EntryEventImpl event) throws EntryNotFoundException {
@@ -4866,7 +4910,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   @Override
   public void recordRecoveredVersonHolder(VersionSource member, RegionVersionHolder versionHolder,
-                                          boolean latestOplog) {
+      boolean latestOplog) {
     if (this.getConcurrencyChecksEnabled()) {
       // We need to update the RVV in memory
       this.versionVector.initRecoveredVersion(member, versionHolder, latestOplog);
@@ -4970,11 +5014,12 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   /**
    * basicInvalidate is overridden in HARegion to abort expiry of Events which have key as Long , if
    * it is not able to destroy from availableIDs
+   *
    * @param forceNewEntry true if we are a mirror and still in the initialization phase. Called from
-   * InvalidateOperation.InvalidateMessage
+   *        InvalidateOperation.InvalidateMessage
    */
   void basicInvalidate(final EntryEventImpl event, boolean invokeCallbacks,
-                       final boolean forceNewEntry) throws EntryNotFoundException {
+      final boolean forceNewEntry) throws EntryNotFoundException {
     if (!event.isOriginRemote() && !event.isDistributed() && getScope().isDistributed()
         && getDataPolicy().withReplication() && invokeCallbacks) {
       // catches case where being called by (distributed) invalidateRegion
@@ -5001,7 +5046,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   }
 
   void basicInvalidatePart2(RegionEntry regionEntry, EntryEventImpl event,
-                            boolean conflictWithClear, boolean invokeCallbacks) {
+      boolean conflictWithClear, boolean invokeCallbacks) {
     updateStatsForInvalidate();
 
     if (invokeCallbacks) {
@@ -5033,8 +5078,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    */
   @Override
   public void invokeInvalidateCallbacks(final EnumListenerEvent eventType,
-                                        final EntryEventImpl event,
-                                        final boolean callDispatchListenerEvent) {
+      final EntryEventImpl event,
+      final boolean callDispatchListenerEvent) {
     // Notify bridge clients (if this is a CacheServer)
     event.setEventType(eventType);
     notifyBridgeClients(event);
@@ -5054,12 +5099,12 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    */
   @Override
   public void txApplyInvalidate(Object key, Object newValue, boolean didDestroy,
-                                TransactionId transactionId, TXRmtEvent event, boolean localOp,
-                                EventID eventId,
-                                Object aCallbackArgument, List<EntryEventImpl> pendingCallbacks,
-                                FilterRoutingInfo filterRoutingInfo,
-                                ClientProxyMembershipID bridgeContext,
-                                TXEntryState txEntryState, VersionTag versionTag, long tailKey) {
+      TransactionId transactionId, TXRmtEvent event, boolean localOp,
+      EventID eventId,
+      Object aCallbackArgument, List<EntryEventImpl> pendingCallbacks,
+      FilterRoutingInfo filterRoutingInfo,
+      ClientProxyMembershipID bridgeContext,
+      TXEntryState txEntryState, VersionTag versionTag, long tailKey) {
 
     this.entries.txApplyInvalidate(key, newValue, didDestroy, transactionId, event, localOp,
         eventId, aCallbackArgument, pendingCallbacks, filterRoutingInfo, bridgeContext,
@@ -5071,7 +5116,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * completed its part of the basic destroy
    */
   void txApplyInvalidatePart2(RegionEntry regionEntry, Object key, boolean didDestroy,
-                              boolean didInvalidate) {
+      boolean didInvalidate) {
     if (this.testCallable != null) {
       this.testCallable.call(this, Operation.INVALIDATE, regionEntry);
     }
@@ -5100,20 +5145,21 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   /**
    * Allows null as new value to accomodate create with a null value. Assumes all key, value, and
    * callback validations have been performed.
+   *
    * @param event the event object for this operation, with the exception that the oldValue
-   * parameter is not yet filled in. The oldValue will be filled in by this operation.
+   *        parameter is not yet filled in. The oldValue will be filled in by this operation.
    * @param ifNew true if this operation must not overwrite an existing key
    * @param ifOld true if this operation must not create a new key
    * @param expectedOldValue only succeed if old value is equal to this value. If null, then doesn't
-   * matter what old value is. If INVALID token, must be INVALID.
+   *        matter what old value is. If INVALID token, must be INVALID.
    * @param requireOldValue true if the oldValue should be set in event even if ifNew and entry
-   * exists
+   *        exists
    * @return false if ifNew is true and there is an existing key or if ifOld is true and
-   * expectedOldValue does not match the current value in the cache. Otherwise return true.
+   *         expectedOldValue does not match the current value in the cache. Otherwise return true.
    */
   @Override
   public boolean basicPut(EntryEventImpl event, boolean ifNew, boolean ifOld,
-                          Object expectedOldValue, boolean requireOldValue)
+      Object expectedOldValue, boolean requireOldValue)
       throws TimeoutException, CacheWriterException {
     return getDataView().putEntry(event, ifNew, ifOld, expectedOldValue, requireOldValue, 0L,
         false);
@@ -5132,12 +5178,12 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    */
   @Override
   public void txApplyPut(Operation putOp, Object key, Object newValue, boolean didDestroy,
-                         TransactionId transactionId, TXRmtEvent event, EventID eventId,
-                         Object aCallbackArgument,
-                         List<EntryEventImpl> pendingCallbacks, FilterRoutingInfo filterRoutingInfo,
-                         ClientProxyMembershipID bridgeContext, TXEntryState txEntryState,
-                         VersionTag versionTag,
-                         long tailKey) {
+      TransactionId transactionId, TXRmtEvent event, EventID eventId,
+      Object aCallbackArgument,
+      List<EntryEventImpl> pendingCallbacks, FilterRoutingInfo filterRoutingInfo,
+      ClientProxyMembershipID bridgeContext, TXEntryState txEntryState,
+      VersionTag versionTag,
+      long tailKey) {
     long startPut = CachePerfStats.getStatTime();
     this.entries.txApplyPut(putOp, key, newValue, didDestroy, transactionId, event, eventId,
         aCallbackArgument, pendingCallbacks, filterRoutingInfo, bridgeContext, txEntryState,
@@ -5157,7 +5203,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   @Override
   public void txApplyPutPart2(RegionEntry regionEntry, Object key, long lastModified,
-                              boolean isCreate, boolean didDestroy, boolean clearConflict) {
+      boolean isCreate, boolean didDestroy, boolean clearConflict) {
     if (this.testCallable != null) {
       Operation op = isCreate ? Operation.CREATE : Operation.UPDATE;
       this.testCallable.call(this, op, regionEntry);
@@ -5188,9 +5234,9 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   }
 
   public boolean basicBridgeCreate(final Object key, final byte[] value, boolean isObject,
-                                   Object callbackArg, final ClientProxyMembershipID client,
-                                   boolean fromClient,
-                                   EntryEventImpl clientEvent, boolean throwEntryExists)
+      Object callbackArg, final ClientProxyMembershipID client,
+      boolean fromClient,
+      EntryEventImpl clientEvent, boolean throwEntryExists)
       throws TimeoutException, EntryExistsException, CacheWriterException {
 
     EventID eventId = clientEvent.getEventId();
@@ -5205,7 +5251,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
       }
     }
 
-    @Released final EntryEventImpl event = EntryEventImpl.create(this, Operation.CREATE, key, value,
+    @Released
+    final EntryEventImpl event = EntryEventImpl.create(this, Operation.CREATE, key, value,
         theCallbackArg, false /* origin remote */, client.getDistributedMember(),
         true /* generateCallbacks */, eventId);
 
@@ -5264,9 +5311,9 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   }
 
   public boolean basicBridgePut(Object key, Object value, byte[] deltaBytes, boolean isObject,
-                                Object callbackArg, ClientProxyMembershipID memberId,
-                                boolean fromClient,
-                                EntryEventImpl clientEvent)
+      Object callbackArg, ClientProxyMembershipID memberId,
+      boolean fromClient,
+      EntryEventImpl clientEvent)
       throws TimeoutException, CacheWriterException {
 
     EventID eventID = clientEvent.getEventId();
@@ -5281,7 +5328,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
       }
     }
 
-    @Released final EntryEventImpl event = EntryEventImpl.create(this, Operation.UPDATE, key,
+    @Released
+    final EntryEventImpl event = EntryEventImpl.create(this, Operation.UPDATE, key,
         null /* new value */, theCallbackArg, false /* origin remote */,
         memberId.getDistributedMember(), true /* generateCallbacks */, eventID);
 
@@ -5340,7 +5388,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
       this.concurrencyMessageIssued = true;
       logger.info(LocalizedMessage.create(
           LocalizedStrings.LocalRegion_SERVER_HAS_CONCURRENCY_CHECKS_ENABLED_0_BUT_CLIENT_HAS_1_FOR_REGION_2,
-          new Object[]{!this.getConcurrencyChecksEnabled(), this.getConcurrencyChecksEnabled(),
+          new Object[] {!this.getConcurrencyChecksEnabled(), this.getConcurrencyChecksEnabled(),
               this}));
     }
   }
@@ -5350,10 +5398,10 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * bridge server and should not be distributed back to it.
    */
   public void basicBridgeClientUpdate(DistributedMember serverId, Object key, Object value,
-                                      byte[] deltaBytes, boolean isObject, Object callbackArgument,
-                                      boolean isCreate,
-                                      boolean processedMarker, EntryEventImpl event,
-                                      EventID eventID)
+      byte[] deltaBytes, boolean isObject, Object callbackArgument,
+      boolean isCreate,
+      boolean processedMarker, EntryEventImpl event,
+      EventID eventID)
       throws TimeoutException, CacheWriterException {
 
     if (isCacheContentProxy()) {
@@ -5411,8 +5459,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * distributed back to it.
    */
   public void basicBridgeClientInvalidate(DistributedMember serverId, Object key,
-                                          Object callbackArgument, boolean processedMarker,
-                                          EventID eventID, VersionTag versionTag)
+      Object callbackArgument, boolean processedMarker,
+      EventID eventID, VersionTag versionTag)
       throws EntryNotFoundException {
 
     if (!isCacheContentProxy()) {
@@ -5461,8 +5509,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * distributed back to it.
    */
   public void basicBridgeClientDestroy(DistributedMember serverId, Object key,
-                                       Object callbackArgument, boolean processedMarker,
-                                       EventID eventID, VersionTag versionTag)
+      Object callbackArgument, boolean processedMarker,
+      EventID eventID, VersionTag versionTag)
       throws EntryNotFoundException {
 
     if (!isCacheContentProxy()) {
@@ -5509,8 +5557,9 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Clear the region from a server request.
+   *
    * @param callbackArgument The callback argument. This is currently null since {@link
-   * java.util.Map#clear} supports no parameters.
+   *        java.util.Map#clear} supports no parameters.
    * @param processedMarker Whether the marker has been processed (for durable clients)
    */
   public void basicBridgeClientClear(Object callbackArgument, boolean processedMarker) {
@@ -5531,7 +5580,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   }
 
   public void basicBridgeDestroy(Object key, Object callbackArg, ClientProxyMembershipID memberId,
-                                 boolean fromClient, EntryEventImpl clientEvent)
+      boolean fromClient, EntryEventImpl clientEvent)
       throws TimeoutException, EntryNotFoundException, CacheWriterException {
 
     Object theCallbackArg = callbackArg;
@@ -5544,7 +5593,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
     }
 
     // Create an event and put the entry
-    @Released final EntryEventImpl event = EntryEventImpl.create(this, Operation.DESTROY, key, null,
+    @Released
+    final EntryEventImpl event = EntryEventImpl.create(this, Operation.DESTROY, key, null,
         theCallbackArg, false, memberId.getDistributedMember(), true, clientEvent.getEventId());
 
     try {
@@ -5568,8 +5618,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   // TODO: fromClient is always true
   public void basicBridgeInvalidate(Object key, Object callbackArg,
-                                    ClientProxyMembershipID memberId, boolean fromClient,
-                                    EntryEventImpl clientEvent)
+      ClientProxyMembershipID memberId, boolean fromClient,
+      EntryEventImpl clientEvent)
       throws TimeoutException, EntryNotFoundException, CacheWriterException {
 
     Object theCallbackArg = callbackArg;
@@ -5582,8 +5632,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
     }
 
     // Create an event and put the entry
-    @Released final EntryEventImpl
-        event =
+    @Released
+    final EntryEventImpl event =
         EntryEventImpl.create(this, Operation.INVALIDATE, key, null,
             theCallbackArg, false, memberId.getDistributedMember(), true, clientEvent.getEventId());
 
@@ -5607,8 +5657,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   // TODO: fromClient is always false and never used
   // TODO: callbackArg is never used
   public void basicBridgeUpdateVersionStamp(Object key, Object callbackArg,
-                                            ClientProxyMembershipID memberId, boolean fromClient,
-                                            EntryEventImpl clientEvent) {
+      ClientProxyMembershipID memberId, boolean fromClient,
+      EntryEventImpl clientEvent) {
 
     // Create an event and update version stamp of the entry
     @Released
@@ -5648,21 +5698,23 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Allows null as new value to accommodate create with a null value.
+   *
    * @param event the event object for this operation, with the exception that the oldValue
-   * parameter is not yet filled in. The oldValue will be filled in by this operation.
+   *        parameter is not yet filled in. The oldValue will be filled in by this operation.
    * @param ifNew true if this operation must not overwrite an existing key
    * @param ifOld true if this operation must not create a new entry
    * @param lastModified the lastModified time to set with the value; if 0L, then the lastModified
-   * time will be set to now.
+   *        time will be set to now.
    * @param overwriteDestroyed true if okay to overwrite the DESTROYED token: when this is true has
-   * the following effect: even when ifNew is true will write over DESTROYED token when
-   * overwriteDestroyed is false and ifNew or ifOld is true then if the put doesn't occur because
-   * there is a DESTROYED token present then the entry flag blockedDestroyed is set.
+   *        the following effect: even when ifNew is true will write over DESTROYED token when
+   *        overwriteDestroyed is false and ifNew or ifOld is true then if the put doesn't occur
+   *        because
+   *        there is a DESTROYED token present then the entry flag blockedDestroyed is set.
    * @return false if ifNew is true and there is an existing key, or ifOld is true and there is no
-   * existing entry; otherwise return true.
+   *         existing entry; otherwise return true.
    */
   boolean basicUpdate(final EntryEventImpl event, final boolean ifNew, final boolean ifOld,
-                      final long lastModified, final boolean overwriteDestroyed)
+      final long lastModified, final boolean overwriteDestroyed)
       throws TimeoutException, CacheWriterException {
 
     // check validity of key against keyConstraint
@@ -5685,9 +5737,9 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    */
   @Override
   public boolean virtualPut(final EntryEventImpl event, final boolean ifNew, final boolean ifOld,
-                            Object expectedOldValue, boolean requireOldValue,
-                            final long lastModified,
-                            final boolean overwriteDestroyed)
+      Object expectedOldValue, boolean requireOldValue,
+      final long lastModified,
+      final boolean overwriteDestroyed)
       throws TimeoutException, CacheWriterException {
 
     if (!MemoryThresholds.isLowMemoryExceptionDisabled()) {
@@ -5749,6 +5801,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   /**
    * Checks to see if the event should be rejected because of sick state either due to exceeding
    * local critical threshold or a remote member exceeding critical threshold
+   *
    * @param key the key for the operation
    * @throws LowMemoryException if the target member for this operation is sick
    */
@@ -5770,10 +5823,11 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Allows null as new value to accommodate create with a null value.
+   *
    * @param event the event object for this operation, with the exception that the oldValue
-   * parameter is not yet filled in. The oldValue will be filled in by this operation.
+   *        parameter is not yet filled in. The oldValue will be filled in by this operation.
    * @param lastModified the lastModified time to set with the value; if 0L then the lastModified
-   * time will be set to now.
+   *        time will be set to now.
    * @return null if put not done; otherwise the put entry
    */
   protected RegionEntry basicPutEntry(final EntryEventImpl event, final long lastModified)
@@ -5801,7 +5855,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   @Override
   public long basicPutPart2(EntryEventImpl event, RegionEntry entry, boolean isInitialized,
-                            long lastModified, boolean clearConflict) {
+      long lastModified, boolean clearConflict) {
 
     final boolean isNewKey = event.getOperation().isCreate();
 
@@ -5878,7 +5932,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * message distribution and before waiting for responses.
    *
    * Warning: Even if you comment out bucket condition in following method, getLocalRoutingInfo()
-   * does NOT process CQs for bucket regions internally. See {@link FilterProfile#getFilterRoutingInfoPart2(FilterRoutingInfo,
+   * does NOT process CQs for bucket regions internally. See
+   * {@link FilterProfile#getFilterRoutingInfoPart2(FilterRoutingInfo,
    * CacheEvent)} .
    */
   protected void generateLocalFilterRouting(InternalCacheEvent event) {
@@ -5968,9 +6023,9 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   @Override
   public void basicPutPart3(EntryEventImpl event, RegionEntry entry, boolean isInitialized,
-                            long lastModified, boolean invokeCallbacks, boolean ifNew,
-                            boolean ifOld,
-                            Object expectedOldValue, boolean requireOldValue) {
+      long lastModified, boolean invokeCallbacks, boolean ifNew,
+      boolean ifOld,
+      Object expectedOldValue, boolean requireOldValue) {
 
     // We used to dispatch listener events here which is moved to part2 to be in RE lock #45520.
     if (invokeCallbacks) {
@@ -5982,7 +6037,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   @Override
   public void invokePutCallbacks(final EnumListenerEvent eventType, final EntryEventImpl event,
-                                 final boolean callDispatchListenerEvent, boolean notifyGateways) {
+      final boolean callDispatchListenerEvent, boolean notifyGateways) {
 
     // disallow callbacks on import
     if (!event.isGenerateCallbacks()) {
@@ -6023,6 +6078,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * Record the event state encapsulated in the given Map.
    * <p>
    * This is intended for state transfer during GII.
+   *
    * @param provider the member that provided this state
    * @param state a Map obtained from getEventState()
    */
@@ -6058,6 +6114,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * has the Region's event state seen this event?
+   *
    * @return true if the Region's event state has seen the event
    */
   @Override
@@ -6067,6 +6124,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * tries to find the version tag for a event
+   *
    * @return the version tag, if known. Null if not
    */
   @Override
@@ -6076,6 +6134,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * tries to find the version tag for a replayed client event
+   *
    * @return the version tag, if known. Null if not
    */
   public VersionTag findVersionTagForClientBulkOp(EventID eventId) {
@@ -6087,6 +6146,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * Event, not an ID, but with transactions we do not have an event at the time the check needs to
    * be made. Consequently, this method may cause events to be recorded that would otherwise be
    * ignored.
+   *
    * @param eventID the identifier of the event
    * @return true if the Region's event state has seen the event
    */
@@ -6098,6 +6158,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   /**
    * A routine to provide synchronization running based on <memberShipID, threadID> of the
    * requesting client for the region's event state
+   *
    * @param task - a Runnable to wrap the processing of the bulk op
    * @param eventId - the base event ID of the bulk op
    * @since GemFire 5.7
@@ -6244,7 +6305,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   }
 
   void basicDestroyRegion(RegionEventImpl event, boolean cacheWrite, boolean lock,
-                          boolean callbackEvents) throws CacheWriterException, TimeoutException {
+      boolean callbackEvents) throws CacheWriterException, TimeoutException {
     preDestroyChecks();
 
     final TXStateProxy tx = this.cache.getTXMgr().pauseTransaction();
@@ -6378,6 +6439,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Called after this region has been completely created
+   *
    * @see DistributedRegion#postDestroyRegion(boolean, RegionEventImpl)
    * @since GemFire 5.0
    */
@@ -6465,12 +6527,12 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * @param cacheWrite true if cacheWrite should be performed or false if cacheWrite should not be
-   * performed
+   *        performed
    * @see DistributedRegion#basicDestroy(EntryEventImpl, boolean, Object)
    */
   @Override
   public void basicDestroy(final EntryEventImpl event, final boolean cacheWrite,
-                           Object expectedOldValue)
+      Object expectedOldValue)
       throws EntryNotFoundException, CacheWriterException, TimeoutException {
 
     if (!event.isOriginRemote()) {
@@ -6505,6 +6567,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   /**
    * Do the expensive work of discovering an existing JTA transaction Only needs to be called at
    * Region.Entry entry points e.g. Region.put, Region.invalidate, etc.
+   *
    * @since GemFire tx
    */
   void discoverJTA() {
@@ -6534,12 +6597,12 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * @param expectedOldValue if this is non-null, only destroy if key exists and old value is equal
-   * to expectedOldValue
+   *        to expectedOldValue
    * @return true if a the destroy was done; false if it was not needed
    */
   @Override
   public boolean mapDestroy(final EntryEventImpl event, final boolean cacheWrite,
-                            final boolean isEviction, Object expectedOldValue)
+      final boolean isEviction, Object expectedOldValue)
       throws CacheWriterException, EntryNotFoundException, TimeoutException {
 
     final boolean inGII = lockGII();
@@ -6553,9 +6616,9 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   }
 
   private boolean mapDestroy(final EntryEventImpl event, final boolean cacheWrite,
-                             final boolean isEviction, Object expectedOldValue,
-                             boolean needTokensForGII,
-                             boolean removeRecoveredEntry) {
+      final boolean isEviction, Object expectedOldValue,
+      boolean needTokensForGII,
+      boolean removeRecoveredEntry) {
 
     // When register interest is in progress ,
     // We should not remove the key from the
@@ -6625,10 +6688,11 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   /**
    * @param dae DiskAccessException encountered by the thread
    * @param duringInitialization indicates that this exception occurred during region
-   * initialization. Instead of closing the cache here, we rely on the region initialization to
-   * clean things up.
+   *        initialization. Instead of closing the cache here, we rely on the region initialization
+   *        to
+   *        clean things up.
    * @see DistributedRegion#initialize(InputStream, InternalDistributedMember,
-   * InternalRegionArguments)
+   *      InternalRegionArguments)
    * @see LocalRegion#initialize(InputStream, InternalDistributedMember, InternalRegionArguments)
    * @see InitialImageOperation#processChunk
    */
@@ -6662,6 +6726,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   /**
    * Creates an event for EVICT_DESTROY operations. It is intended that this method be overridden to
    * allow for special handling of Partitioned Regions.
+   *
    * @param key - the key that this event is related to
    * @return an event for EVICT_DESTROY
    */
@@ -6684,7 +6749,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   boolean evictDestroy(EvictableEntry entry) {
     checkReadiness();
 
-    @Released final EntryEventImpl event = generateEvictDestroyEvent(entry.getKey());
+    @Released
+    final EntryEventImpl event = generateEvictDestroyEvent(entry.getKey());
 
     try {
       return mapDestroy(event, false, // cacheWrite
@@ -6715,6 +6781,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * <bold>and</bold> while the entry remains in the map. Once the entry is removed from the map,
    * then other operations synchronize on a new entry, allow for ordering problems between {@link
    * #create(Object, Object, Object)} and {@link #destroy(Object, Object)} operations.
+   *
    * @param entry the Region entry being destroyed
    * @param event the event describing the destroy operation
    * @since GemFire 5.1
@@ -6730,8 +6797,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    */
   @Override
   public void basicDestroyPart2(RegionEntry re, EntryEventImpl event, boolean inTokenMode,
-                                boolean conflictWithClear, boolean duringRI,
-                                boolean invokeCallbacks) {
+      boolean conflictWithClear, boolean duringRI,
+      boolean invokeCallbacks) {
     if (!(this instanceof HARegion)) {
       if (logger.isTraceEnabled()) {
         logger.trace("basicDestroyPart2(inTokenMode={},conflictWithClear={},duringRI={}) event={}",
@@ -6799,8 +6866,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    */
   @Override
   public void basicDestroyPart3(RegionEntry re, EntryEventImpl event, boolean inTokenMode,
-                                boolean duringRI, boolean invokeCallbacks,
-                                Object expectedOldValue) {
+      boolean duringRI, boolean invokeCallbacks,
+      Object expectedOldValue) {
 
     if (invokeCallbacks) {
       if (event.isBulkOpInProgress()) {
@@ -6831,8 +6898,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   @Override
   public void invokeDestroyCallbacks(final EnumListenerEvent eventType, final EntryEventImpl event,
-                                     final boolean callDispatchListenerEvent,
-                                     boolean notifyGateways) {
+      final boolean callDispatchListenerEvent,
+      boolean notifyGateways) {
 
     // The spec for ConcurrentMap support requires that operations be mapped
     // to non-CM counterparts
@@ -6851,7 +6918,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   @Override
   public void invokeTXCallbacks(final EnumListenerEvent eventType, final EntryEventImpl event,
-                                final boolean callDispatchListenerEvent) {
+      final boolean callDispatchListenerEvent) {
 
     // The spec for ConcurrentMap support requires that operations be mapped
     // to non-CM counterparts
@@ -6885,7 +6952,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * @param rmtOrigin true if transaction being applied had a remote origin
    * @param event filled in if operation performed
    * @param needTokensForGII true if caller has determined we are in destroy token mode and will
-   * keep us in that mode while this call is executing.
+   *        keep us in that mode while this call is executing.
    * @param isOriginRemote whether the event originated in a peer or in this vm
    * @param txEntryState for passing up versionTag - only on near side
    * @param versionTag tag generated by txCoordinator - only on far side
@@ -6893,13 +6960,13 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    */
   @Override
   public void txApplyDestroy(Object key, TransactionId rmtOrigin, TXRmtEvent event,
-                             boolean needTokensForGII, Operation op, EventID eventId,
-                             Object aCallbackArgument,
-                             List<EntryEventImpl> pendingCallbacks,
-                             FilterRoutingInfo filterRoutingInfo,
-                             ClientProxyMembershipID bridgeContext, boolean isOriginRemote,
-                             TXEntryState txEntryState,
-                             VersionTag versionTag, long tailKey) {
+      boolean needTokensForGII, Operation op, EventID eventId,
+      Object aCallbackArgument,
+      List<EntryEventImpl> pendingCallbacks,
+      FilterRoutingInfo filterRoutingInfo,
+      ClientProxyMembershipID bridgeContext, boolean isOriginRemote,
+      TXEntryState txEntryState,
+      VersionTag versionTag, long tailKey) {
 
     final boolean inRI = !needTokensForGII && lockRIReadLock();
     final boolean needRIDestroyToken = inRI && this.riCnt > 0;
@@ -6921,7 +6988,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * completed its part of the basic destroy
    */
   void txApplyDestroyPart2(RegionEntry re, Object key, boolean inTokenMode, boolean clearConflict,
-                           boolean alreadyDestroyedOrRemoved) {
+      boolean alreadyDestroyedOrRemoved) {
     if (this.testCallable != null) {
       this.testCallable.call(this, Operation.DESTROY, re);
     }
@@ -7186,6 +7253,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   /**
    * PRECONDITIONS: Synchronized on updateMonitor for this key in order to guarantee write-through
    * to map entry, and key must be in map
+   *
    * @param lastModified time, may be 0 in which case uses now instead
    * @return the actual lastModifiedTime used.
    */
@@ -7218,6 +7286,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Returns a region in the subregion map first, then looks in the reinitializing region registry.
+   *
    * @return the region or null if not found, may be destroyed
    */
   private LocalRegion basicGetSubregion(String name) {
@@ -7244,6 +7313,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Make a LocalRegion from an element in the subregion map Sent to parent region.
+   *
    * @return This method may return null or a destroyed region if the region was just destroyed
    */
   private LocalRegion toRegion(Object element) {
@@ -7257,6 +7327,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Update the API statistics appropriately for returning this value from get.
+   *
    * @param regionEntry the entry whose value was accessed
    */
   public void updateStatsForGet(final RegionEntry regionEntry, final boolean hit) {
@@ -7315,6 +7386,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Release the client connection pool if we have one
+   *
    * @since GemFire 5.7
    */
   private void detachPool() {
@@ -7482,6 +7554,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * For each region entry in this region call the callback
+   *
    * @since GemFire prPersistSprint2
    */
   @Override
@@ -7493,6 +7566,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Used by {@link #foreachRegionEntry}.
+   *
    * @since GemFire prPersistSprint2
    */
   public interface RegionEntryCallback {
@@ -7617,7 +7691,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   }
 
   protected DiskStoreImpl findDiskStore(RegionAttributes regionAttributes,
-                                        InternalRegionArguments internalRegionArgs) {
+      InternalRegionArguments internalRegionArgs) {
     // validate that persistent type registry is persistent
     if (getAttributes().getDataPolicy().withPersistence()) {
       getCache().getPdxRegistry().creatingPersistentRegion();
@@ -7663,6 +7737,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   /**
    * Creates a new {@code DiskRegion} for this region. We assume that the attributes and the name of
    * the region have been set.
+   *
    * @return {@code null} is a disk region is not desired
    * @since GemFire 3.2
    */
@@ -7706,6 +7781,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Returns the object sizer on this region or null if it has no sizer.
+   *
    * @since GemFire 6.1.2.9
    */
   @Override
@@ -7970,6 +8046,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
     /**
      * Returns the entry's RegionEntry if it "checks" out. The check is to see if the region entry
      * still exists.
+     *
      * @throws EntryNotFoundException if the RegionEntry has been removed.
      */
     private RegionEntry getCheckedRegionEntry() throws EntryNotFoundException {
@@ -8275,6 +8352,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   /**
    * Callers of this method should always follow the call with: if (lockGII()) { try { } finally {
    * unlockGII(); } }
+   *
    * @return true if lock obtained and unlock needs to be called
    */
   @Override
@@ -8303,6 +8381,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   /**
    * Callers of this method should always follow the call with: if (lockRIReadLock()) { try { }
    * finally { unlockRIReadLock(); } }
+   *
    * @return true if lock obtained and unlock needs to be called
    */
   private boolean lockRIReadLock() {
@@ -8363,7 +8442,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   @Override
   public TXEntryState createReadEntry(TXRegionState txRegionState, KeyInfo keyInfo,
-                                      boolean createIfAbsent) {
+      boolean createIfAbsent) {
     TXEntryState result = null;
     final RegionEntry regionEntry = this.basicGetTXEntry(keyInfo);
     if (regionEntry != null) {
@@ -8516,7 +8595,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   // package-private to avoid synthetic accessor
   static void dispatchEvent(LocalRegion region, InternalCacheEvent event,
-                            EnumListenerEvent operation) {
+      EnumListenerEvent operation) {
 
     CacheListener[] listeners = region.fetchCacheListenersField();
     if (event.getOperation().isCreate()) {
@@ -8720,7 +8799,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
     public Object[] toArray() {
       List temp = new ArrayList(this.size());
       // do NOT use addAll or this results in stack overflow - must use iterator()
-      for (Iterator iter = iterator(); iter.hasNext(); ) {
+      for (Iterator iter = iterator(); iter.hasNext();) {
         temp.add(iter.next());
       }
       return temp.toArray();
@@ -8730,7 +8809,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
     public Object[] toArray(Object[] array) {
       List temp = new ArrayList(this.size());
       // do NOT use addAll or this results in stack overflow - must use iterator()
-      for (Iterator iter = iterator(); iter.hasNext(); ) {
+      for (Iterator iter = iterator(); iter.hasNext();) {
         temp.add(iter.next());
       }
       return temp.toArray(array);
@@ -8816,6 +8895,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
     /**
      * To get the value from region in serialized form
+     *
      * @return {@link VMCachedDeserializable}
      */
     public Object getRawValue() {
@@ -8910,6 +8990,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * because of the concurrency properties of the {@link RegionMap}, the number of entries is only
    * an approximate. That is, other threads may change the number of entries in this region while
    * this method is being invoked.
+   *
    * @see LocalRegion#entryCount()
    */
   @Override
@@ -8935,6 +9016,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * returns an estimate of the number of entries in this region. This method should be preferred
    * over size() for hdfs regions where an accurate size is not needed. This method is not supported
    * on a client
+   *
    * @return the estimated size of this region
    */
   public int sizeEstimate() {
@@ -9016,7 +9098,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   // TODO: fromClient is always true
   public void basicBridgeDestroyRegion(Object callbackArg, final ClientProxyMembershipID client,
-                                       boolean fromClient, EventID eventId)
+      boolean fromClient, EventID eventId)
       throws TimeoutException, EntryExistsException, CacheWriterException {
 
     if (fromClient) {
@@ -9034,7 +9116,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   }
 
   public void basicBridgeClear(Object callbackArg, final ClientProxyMembershipID client,
-                               boolean fromClient, EventID eventId)
+      boolean fromClient, EventID eventId)
       throws TimeoutException, EntryExistsException, CacheWriterException {
 
     if (fromClient) {
@@ -9079,7 +9161,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * The clear operation will have immediate committed state.
    */
   void clearRegionLocally(RegionEventImpl regionEvent, boolean cacheWrite,
-                          RegionVersionVector vector) {
+      RegionVersionVector vector) {
     final boolean isRvvDebugEnabled = logger.isTraceEnabled(LogMarker.RVV_VERBOSE);
 
     RegionVersionVector rvv = vector;
@@ -9280,7 +9362,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
           // to keep the original collection intact
           keysList = new ArrayList(keys);
         }
-        for (Iterator iterator = keysList.iterator(); iterator.hasNext(); ) {
+        for (Iterator iterator = keysList.iterator(); iterator.hasNext();) {
           Object key = iterator.next();
           Object value;
           Region.Entry entry = accessEntry(key, true);
@@ -9302,7 +9384,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
         }
 
         // Add remote results to local cache and all results if successful
-        for (VersionedObjectList.Iterator it = remoteResults.iterator(); it.hasNext(); ) {
+        for (VersionedObjectList.Iterator it = remoteResults.iterator(); it.hasNext();) {
           VersionedObjectList.Entry entry = it.next();
           Object key = entry.getKey();
           boolean notOnServer = entry.isKeyNotOnServer();
@@ -9423,15 +9505,17 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Called on a bridge server when it has a received a putAll command from a client.
+   *
    * @param map a map of key->value for the entries we are putting
    * @param retryVersions a map of key->version tag. If any of the entries are the result of a
-   * retried client event, we need to make sure we send the original version tag along with the
-   * event.
+   *        retried client event, we need to make sure we send the original version tag along with
+   *        the
+   *        event.
    * @param callbackArg callback argument from client
    */
   public VersionedObjectList basicBridgePutAll(Map map, Map<Object, VersionTag> retryVersions,
-                                               ClientProxyMembershipID memberId, EventID eventId,
-                                               boolean skipCallbacks, Object callbackArg)
+      ClientProxyMembershipID memberId, EventID eventId,
+      boolean skipCallbacks, Object callbackArg)
       throws TimeoutException, CacheWriterException {
 
     long startPut = CachePerfStats.getStatTime();
@@ -9439,8 +9523,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
       callbackArg = new GatewaySenderEventCallbackArgument(callbackArg);
     }
 
-    @Released final EntryEventImpl
-        event =
+    @Released
+    final EntryEventImpl event =
         EntryEventImpl.create(this, Operation.PUTALL_CREATE, null,
             null /* new value */, callbackArg, false /* origin remote */,
             memberId.getDistributedMember(), !skipCallbacks /* generateCallbacks */, eventId);
@@ -9462,16 +9546,18 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Called on a bridge server when it has a received a removeAll command from a client.
+   *
    * @param keys a collection of the keys we are putting
    * @param retryVersions a collection of version tags. If the client is retrying a key then that
-   * keys slot will be non-null in this collection. Note that keys and retryVersions are parallel
-   * lists.
+   *        keys slot will be non-null in this collection. Note that keys and retryVersions are
+   *        parallel
+   *        lists.
    * @param callbackArg callback argument from client
    */
   public VersionedObjectList basicBridgeRemoveAll(List<Object> keys,
-                                                  ArrayList<VersionTag> retryVersions,
-                                                  ClientProxyMembershipID memberId, EventID eventId,
-                                                  Object callbackArg)
+      ArrayList<VersionTag> retryVersions,
+      ClientProxyMembershipID memberId, EventID eventId,
+      Object callbackArg)
       throws TimeoutException, CacheWriterException {
 
     long startOp = CachePerfStats.getStatTime();
@@ -9479,8 +9565,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
       callbackArg = new GatewaySenderEventCallbackArgument(callbackArg);
     }
 
-    @Released final EntryEventImpl
-        event =
+    @Released
+    final EntryEventImpl event =
         EntryEventImpl.create(this, Operation.REMOVEALL_DESTROY, null,
             null /* new value */, callbackArg, false /* origin remote */,
             memberId.getDistributedMember(), true /* generateCallbacks */, eventId);
@@ -9575,8 +9661,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   // TODO: refactor basicPutAll
   public VersionedObjectList basicPutAll(final Map<?, ?> map,
-                                         final DistributedPutAllOperation putAllOp,
-                                         final Map<Object, VersionTag> retryVersions) {
+      final DistributedPutAllOperation putAllOp,
+      final Map<Object, VersionTag> retryVersions) {
 
     final boolean isDebugEnabled = logger.isDebugEnabled();
     final EntryEventImpl event = putAllOp.getBaseEvent();
@@ -9750,7 +9836,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
           partialKeys.setSucceededKeysAndVersions(succeeded);
           logger
               .info(LocalizedMessage.create(LocalizedStrings.Region_PutAll_Applied_PartialKeys_0_1,
-                  new Object[]{getFullPath(), partialKeys}));
+                  new Object[] {getFullPath(), partialKeys}));
           if (isDebugEnabled) {
             logger.debug(partialKeys.detailString());
           }
@@ -9794,8 +9880,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   }
 
   VersionedObjectList basicRemoveAll(final Collection<Object> keys,
-                                     final DistributedRemoveAllOperation removeAllOp,
-                                     final List<VersionTag> retryVersions) {
+      final DistributedRemoveAllOperation removeAllOp,
+      final List<VersionTag> retryVersions) {
 
     final boolean isDebugEnabled = logger.isDebugEnabled();
     final boolean isTraceEnabled = logger.isTraceEnabled();
@@ -9973,7 +10059,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
           partialKeys.setSucceededKeysAndVersions(succeeded);
           logger.info(
               LocalizedMessage.create(LocalizedStrings.Region_RemoveAll_Applied_PartialKeys_0_1,
-                  new Object[]{getFullPath(), partialKeys}));
+                  new Object[] {getFullPath(), partialKeys}));
           if (isDebugEnabled) {
             logger.debug(partialKeys.detailString());
           }
@@ -10076,7 +10162,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   }
 
   private DistributedRemoveAllOperation newRemoveAllOperation(Collection<?> keys,
-                                                              Object callbackArg) {
+      Object callbackArg) {
     if (keys == null) {
       throw new NullPointerException("The keys Collection passed to removeAll was null.");
     }
@@ -10100,6 +10186,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * This performs the putAll operation for a specific key and value
+   *
    * @param key the cache key
    * @param value the cache value
    * @param putallOp the DistributedPutAllOperation associated with the event
@@ -10108,7 +10195,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * @throws CacheWriterException if a cache writer objects to the update
    */
   private void basicEntryPutAll(Object key, Object value, DistributedPutAllOperation putallOp,
-                                int offset, EntryEventImpl tagHolder)
+      int offset, EntryEventImpl tagHolder)
       throws TimeoutException, CacheWriterException {
 
     assert putallOp != null;
@@ -10154,7 +10241,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   }
 
   private void basicEntryRemoveAll(Object key, DistributedRemoveAllOperation op, int offset,
-                                   EntryEventImpl tagHolder)
+      EntryEventImpl tagHolder)
       throws TimeoutException, CacheWriterException {
 
     assert op != null;
@@ -10208,7 +10295,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   @Override
   public void postPutAllFireEvents(DistributedPutAllOperation putAllOp,
-                                   VersionedObjectList successfulPuts) {
+      VersionedObjectList successfulPuts) {
 
     if (!this.getDataPolicy().withStorage() && this.getConcurrencyChecksEnabled()
         && putAllOp.getBaseEvent().isBridgeEvent()) {
@@ -10223,7 +10310,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
       successfulKeys.add(key);
     }
 
-    for (Iterator it = putAllOp.eventIterator(); it.hasNext(); ) {
+    for (Iterator it = putAllOp.eventIterator(); it.hasNext();) {
       @Unretained
       EntryEventImpl event = (EntryEventImpl) it.next();
 
@@ -10232,18 +10319,18 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
             : EnumListenerEvent.AFTER_UPDATE;
         invokePutCallbacks(op, event, !event.callbacksInvoked() && !event.isPossibleDuplicate(),
             this.isUsedForPartitionedRegionBucket
-            /*
-             * If this is replicated region, use "false". We must notify gateways inside RegionEntry
-             * lock, NOT here, to preserve the order of events sent by gateways for same key. If this is
-             * bucket region, use "true", because the event order is guaranteed
-             */);
+        /*
+         * If this is replicated region, use "false". We must notify gateways inside RegionEntry
+         * lock, NOT here, to preserve the order of events sent by gateways for same key. If this is
+         * bucket region, use "true", because the event order is guaranteed
+         */);
       }
     }
   }
 
   @Override
   public void postRemoveAllFireEvents(DistributedRemoveAllOperation removeAllOp,
-                                      VersionedObjectList successfulOps) {
+      VersionedObjectList successfulOps) {
 
     if (!this.getDataPolicy().withStorage() && this.getConcurrencyChecksEnabled()
         && removeAllOp.getBaseEvent().isBridgeEvent()) {
@@ -10258,7 +10345,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
       successfulKeys.add(key);
     }
 
-    for (Iterator it = removeAllOp.eventIterator(); it.hasNext(); ) {
+    for (Iterator it = removeAllOp.eventIterator(); it.hasNext();) {
       @Unretained
       EntryEventImpl event = (EntryEventImpl) it.next();
 
@@ -10266,31 +10353,32 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
         invokeDestroyCallbacks(EnumListenerEvent.AFTER_DESTROY, event,
             !event.callbacksInvoked() && !event.isPossibleDuplicate(),
             this.isUsedForPartitionedRegionBucket
-            /*
-             * If this is replicated region, use "false". We must notify gateways inside RegionEntry
-             * lock, NOT here, to preserve the order of events sent by gateways for same key. If this is
-             * bucket region, use "true", because the event order is guaranteed
-             */);
+        /*
+         * If this is replicated region, use "false". We must notify gateways inside RegionEntry
+         * lock, NOT here, to preserve the order of events sent by gateways for same key. If this is
+         * bucket region, use "true", because the event order is guaranteed
+         */);
       }
     }
   }
 
   @Override
   public long postPutAllSend(DistributedPutAllOperation putAllOp,
-                             VersionedObjectList successfulPuts) {
+      VersionedObjectList successfulPuts) {
     /* No-op for local region of course */
     return -1;
   }
 
   @Override
   public long postRemoveAllSend(DistributedRemoveAllOperation op,
-                                VersionedObjectList successfulOps) {
+      VersionedObjectList successfulOps) {
     /* No-op for local region of course */
     return -1;
   }
 
   /**
    * DistributedRegion overrides isCurrentlyLockGrantor
+   *
    * @see DistributedRegion#isCurrentlyLockGrantor()
    */
   @Override
@@ -10301,15 +10389,16 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   /**
    * Handle a local region destroy or a region close that was done on this region in a remote vm.
    * Currently the only thing needed is to have the advisor
+   *
    * @param sender the id of the member that did the remote operation
    * @param topSerial the remote serialNumber for the top region (maybe root)
    * @param subregionSerialNumbers map of remote subregions to serialNumbers
    * @param regionDestroyed true if the region was destroyed on the remote host (as opposed to
-   * closed)
+   *        closed)
    * @since GemFire 5.0
    */
   void handleRemoteLocalRegionDestroyOrClose(InternalDistributedMember sender, int topSerial,
-                                             Map subregionSerialNumbers, boolean regionDestroyed) {
+      Map subregionSerialNumbers, boolean regionDestroyed) {
 
     // go through initialization latches
     final int oldLevel = setThreadInitLevelRequirement(LocalRegion.ANY_INIT);
@@ -10323,15 +10412,16 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Does the core work for handleRemoteLocalRegionDestroyOrClose.
+   *
    * @param sender the id of the member that did the remote operation
    * @param topSerial the remote serialNumber for the top region (maybe root)
    * @param subregionSerialNumbers remote map of subregions to serialNumbers
    * @since GemFire 5.0
    */
   private void basicHandleRemoteLocalRegionDestroyOrClose(InternalDistributedMember sender,
-                                                          int topSerial, Map subregionSerialNumbers,
-                                                          boolean subregion,
-                                                          boolean regionDestroyed) {
+      int topSerial, Map subregionSerialNumbers,
+      boolean subregion,
+      boolean regionDestroyed) {
 
     // use topSerial unless this region is in subregionSerialNumbers map
     int serialForThisRegion = topSerial;
@@ -10364,10 +10454,11 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Remove the specified sender from this regions advisor.
+   *
    * @since GemFire 5.0
    */
   protected void removeSenderFromAdvisor(InternalDistributedMember sender, int serial,
-                                         boolean regionDestroyed) {
+      boolean regionDestroyed) {
     // nothing needs to be done here since LocalRegion does not have an advisor.
   }
 
@@ -10379,6 +10470,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   /**
    * This method determines whether this region should synchronize with peer replicated regions when
    * the given member has crashed.
+   *
    * @param id the crashed member
    * @return true if synchronization should be attempted
    */
@@ -10390,6 +10482,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * forces the diskRegion to switch the oplog
+   *
    * @since GemFire 5.1
    */
   @Override
@@ -10401,6 +10494,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * TODO: return value is never used
+   *
    * @deprecated as of prPersistSprint1 use forceCompaction instead
    */
   @Deprecated
@@ -10415,8 +10509,9 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * @return int array containing the IDs of the oplogs which will potentially get rolled else null
-   * if no oplogs were available at the time of signal or region is not having disk persistence. Pls
-   * note that the actual number of oplogs rolled may be more than what is indicated
+   *         if no oplogs were available at the time of signal or region is not having disk
+   *         persistence. Pls
+   *         note that the actual number of oplogs rolled may be more than what is indicated
    * @since GemFire prPersistSprint1
    */
   @Override
@@ -10498,8 +10593,9 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * A convenience method to get the PartitionedRegion for a Bucket
+   *
    * @return If this is an instance of {@link BucketRegion}, returns the {@link PartitionedRegion}
-   * otherwise throws an IllegalArgumentException
+   *         otherwise throws an IllegalArgumentException
    */
   @Override
   public PartitionedRegion getPartitionedRegion() {
@@ -10546,8 +10642,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   @Override
   public LoaderHelper createLoaderHelper(Object key, Object callbackArgument,
-                                         boolean netSearchAllowed, boolean netLoadAllowed,
-                                         SearchLoadAndWriteProcessor searcher) {
+      boolean netSearchAllowed, boolean netLoadAllowed,
+      SearchLoadAndWriteProcessor searcher) {
     return new LoaderHelperImpl(this, key, callbackArgument, netSearchAllowed, netLoadAllowed,
         searcher);
   }
@@ -10559,7 +10655,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
       new DistributionAdvisor.ProfileVisitor<Void>() {
         @Override
         public boolean visit(DistributionAdvisor advisor, Profile profile, int profileIndex,
-                             int numProfiles, Void aggregate) {
+            int numProfiles, Void aggregate) {
           assert profile instanceof CacheProfile;
           final CacheProfile prof = (CacheProfile) profile;
 
@@ -10580,7 +10676,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
       new DistributionAdvisor.ProfileVisitor<Void>() {
         @Override
         public boolean visit(DistributionAdvisor advisor, Profile profile, int profileIndex,
-                             int numProfiles, Void aggregate) {
+            int numProfiles, Void aggregate) {
           assert profile instanceof CacheProfile;
           final CacheProfile prof = (CacheProfile) profile;
 
@@ -10621,6 +10717,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   /**
    * Used to prevent notification of bridge clients, typically used for internal "meta" regions and
    * if the cache doesn't have any bridge servers
+   *
    * @return true only if it's cache has bridge servers and this is nt a meta region
    */
   @Override
@@ -10631,6 +10728,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Check if the region has has a Listener or not
+   *
    * @return true only if this region has a Listener
    */
   @Override
@@ -10640,6 +10738,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Called by ccn when a client goes away
+   *
    * @since GemFire 5.7
    */
   @Override
@@ -10669,6 +10768,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * subregion full paths to SerialNumbers. Return is defined as HashMap because
    * DestroyRegionOperation will provide the map to DataSerializer.writeHashMap which requires
    * HashMap. Returns {@link #destroyedSubregionSerialNumbers}.
+   *
    * @return HashMap of subregions to SerialNumbers
    * @throws IllegalStateException if this region has not been destroyed
    */
@@ -10684,6 +10784,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   /**
    * Returns a map of subregion full paths to SerialNumbers. Caller must have acquired the
    * destroyLock if a stable view is desired. Key is String, value is Integer.
+   *
    * @return HashMap of subregions to SerialNumbers
    */
   private HashMap collectSubregionSerialNumbers() {
@@ -10694,6 +10795,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Iterates over all subregions to put the full path and serial number into the provided map.
+   *
    * @param map the map to put the full path and serial number into for each subregion
    */
   private void addSubregionSerialNumbers(Map map) {
@@ -10774,12 +10876,13 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * Execute the provided named function in all locations that contain the given keys. So function
    * can be executed on just one fabric node, executed in parallel on a subset of nodes in parallel
    * across all the nodes.
+   *
    * @since GemFire 5.8Beta
    */
   public ResultCollector executeFunction(final DistributedRegionFunctionExecutor execution,
-                                         final Function function, final Object args,
-                                         final ResultCollector rc, final Set filter,
-                                         final ServerToClientFunctionResultSender sender) {
+      final Function function, final Object args,
+      final ResultCollector rc, final Set filter,
+      final ServerToClientFunctionResultSender sender) {
 
     if (function.optimizeForWrite() && this.memoryThresholdReached.get()
         && !MemoryThresholds.isLowMemoryExceptionDisabled()) {
@@ -10822,12 +10925,12 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
     if (event.isLocal()) {
       if (event.getState().isCritical() && !event.getPreviousState().isCritical()
           && (event.getType() == ResourceType.HEAP_MEMORY
-          || (event.getType() == ResourceType.OFFHEAP_MEMORY && getOffHeap()))) {
+              || (event.getType() == ResourceType.OFFHEAP_MEMORY && getOffHeap()))) {
         // start rejecting operations
         this.memoryThresholdReached.set(true);
       } else if (!event.getState().isCritical() && event.getPreviousState().isCritical()
           && (event.getType() == ResourceType.HEAP_MEMORY
-          || (event.getType() == ResourceType.OFFHEAP_MEMORY && getOffHeap()))) {
+              || (event.getType() == ResourceType.OFFHEAP_MEMORY && getOffHeap()))) {
         this.memoryThresholdReached.set(false);
       }
     }
@@ -10882,7 +10985,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   @Override
   public void initializeStats(long numEntriesInVM, long numOverflowOnDisk,
-                              long numOverflowBytesOnDisk) {
+      long numOverflowBytesOnDisk) {
     getDiskRegion().getStats().incNumEntriesInVM(numEntriesInVM);
     getDiskRegion().getStats().incNumOverflowOnDisk(numOverflowOnDisk);
   }
@@ -10900,13 +11003,15 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * Initialize the set of remote members whose memory state is critical. This is called when
    * registering using {@link InternalResourceManager#addResourceListener(ResourceType,
    * ResourceListener)}. It should only be called once and very early in this region's lifetime.
+   *
    * @param localMemoryIsCritical true if the local memory is in a critical state
    * @param criticalMembers set of members whose memory is in a critical state
-   * @see ResourceManager#setCriticalHeapPercentage(float) and ResourceManager#setCriticalOffHeapPercentage(float)
+   * @see ResourceManager#setCriticalHeapPercentage(float) and
+   *      ResourceManager#setCriticalOffHeapPercentage(float)
    * @since GemFire 6.0
    */
   public void initialCriticalMembers(boolean localMemoryIsCritical,
-                                     Set<InternalDistributedMember> criticalMembers) {
+      Set<InternalDistributedMember> criticalMembers) {
     assert getScope().isLocal();
     if (localMemoryIsCritical) {
       this.memoryThresholdReached.set(true);
@@ -10944,6 +11049,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   /**
    * Fetch the Region which stores the given key The resulting Region will be used for a read
    * operation e.g. Region.get
+   *
    * @param entryKey key to evaluate to determine the returned region
    * @return region that stores the key
    */
@@ -10955,6 +11061,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
   /**
    * Fetch the Region which stores the given key. The resulting Region will be used for a write
    * operation e.g. Region.put
+   *
    * @param entryKey key to evaluate to determine the returned region
    * @return region that stores the key
    */
@@ -10965,7 +11072,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * @return a set of keys, intended for use by the various Region set operations such as {@link
-   * EntriesSet}
+   *         EntriesSet}
    */
   Set getRegionKeysForIteration() {
     return getRegionMap().keySet();
@@ -10977,8 +11084,9 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   /**
    * Used to bootstrap txState.
+   *
    * @return localMember for local and distributedRegions, member with primary bucket for
-   * partitionedRegions
+   *         partitionedRegions
    */
   @Override
   public DistributedMember getOwnerForKey(KeyInfo key) {
@@ -11599,7 +11707,7 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
     // This check allows NORMAL with local scope to fix bug 44856
     if (this.serverRegionProxy == null
         && (this.getDataPolicy() == DataPolicy.NORMAL && this.scope.isDistributed()
-        || this.getDataPolicy() == DataPolicy.EMPTY)) {
+            || this.getDataPolicy() == DataPolicy.EMPTY)) {
       // the functional spec says these data policies do not support concurrent map
       // operations
       throw new UnsupportedOperationException();
@@ -11630,15 +11738,17 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * If this method does not modify the region then no listeners or other callbacks are executed. If
    * a modification does occur, then the behavior with respect to callbacks is the same as {@link
    * Region#create(Object, Object)}.
+   *
    * @param key key with which the specified value is to be associated.
    * @param value the value for the new entry, which may be null meaning the new entry starts as if
-   * it had been locally invalidated.
+   *        it had been locally invalidated.
    * @return previous value associated with specified key, or <tt>null</tt> if there was no mapping
-   * for key. A <tt>null</tt> return can also indicate that the entry in the region was previously
-   * in an invalidated state.
+   *         for key. A <tt>null</tt> return can also indicate that the entry in the region was
+   *         previously
+   *         in an invalidated state.
    * @throws ClassCastException if key does not satisfy the keyConstraint
    * @throws IllegalArgumentException if the key or value is not serializable and this is a
-   * distributed region
+   *         distributed region
    * @throws TimeoutException if timed out getting distributed lock for {@code Scope.GLOBAL}
    * @throws NullPointerException if key is <tt>null</tt>
    * @throws PartitionedRegionStorageException if the operation could not be completed.
@@ -11850,9 +11960,9 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   // TODO: fromClient is always null
   public Object basicBridgePutIfAbsent(final Object key, Object value, boolean isObject,
-                                       Object callbackArg, final ClientProxyMembershipID client,
-                                       boolean fromClient,
-                                       EntryEventImpl clientEvent)
+      Object callbackArg, final ClientProxyMembershipID client,
+      boolean fromClient,
+      EntryEventImpl clientEvent)
       throws TimeoutException, EntryExistsException, CacheWriterException {
 
     EventID eventId = clientEvent.getEventId();
@@ -11865,8 +11975,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
       }
     }
 
-    @Released final EntryEventImpl
-        event =
+    @Released
+    final EntryEventImpl event =
         EntryEventImpl.create(this, Operation.PUT_IF_ABSENT, key, null,
             callbackArg, false, client.getDistributedMember(), true, eventId);
 
@@ -11933,9 +12043,9 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   // TODO: fromClient is always true
   public boolean basicBridgeReplace(final Object key, Object expectedOldValue, Object value,
-                                    boolean isObject, Object callbackArg,
-                                    final ClientProxyMembershipID client,
-                                    boolean fromClient, EntryEventImpl clientEvent)
+      boolean isObject, Object callbackArg,
+      final ClientProxyMembershipID client,
+      boolean fromClient, EntryEventImpl clientEvent)
       throws TimeoutException, EntryExistsException, CacheWriterException {
 
     EventID eventId = clientEvent.getEventId();
@@ -11948,7 +12058,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
       }
     }
 
-    @Released final EntryEventImpl event = EntryEventImpl.create(this, Operation.REPLACE, key, null,
+    @Released
+    final EntryEventImpl event = EntryEventImpl.create(this, Operation.REPLACE, key, null,
         callbackArg, false, client.getDistributedMember(), true, eventId);
 
     try {
@@ -11993,9 +12104,9 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   // TODO: fromClient is always true
   public Object basicBridgeReplace(final Object key, Object value, boolean isObject,
-                                   Object callbackArg, final ClientProxyMembershipID client,
-                                   boolean fromClient,
-                                   EntryEventImpl clientEvent)
+      Object callbackArg, final ClientProxyMembershipID client,
+      boolean fromClient,
+      EntryEventImpl clientEvent)
       throws TimeoutException, EntryExistsException, CacheWriterException {
 
     EventID eventId = clientEvent.getEventId();
@@ -12008,7 +12119,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
       }
     }
 
-    @Released final EntryEventImpl event = EntryEventImpl.create(this, Operation.REPLACE, key, null,
+    @Released
+    final EntryEventImpl event = EntryEventImpl.create(this, Operation.REPLACE, key, null,
         callbackArg, false, client.getDistributedMember(), true, eventId);
 
     try {
@@ -12063,8 +12175,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
 
   // TODO: fromClient is always true
   public void basicBridgeRemove(Object key, Object expectedOldValue, Object callbackArg,
-                                ClientProxyMembershipID memberId, boolean fromClient,
-                                EntryEventImpl clientEvent)
+      ClientProxyMembershipID memberId, boolean fromClient,
+      EntryEventImpl clientEvent)
       throws TimeoutException, EntryNotFoundException, CacheWriterException {
 
     if (fromClient) {
@@ -12076,7 +12188,8 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
     }
 
     // Create an event and put the entry
-    @Released final EntryEventImpl event = EntryEventImpl.create(this, Operation.REMOVE, key, null,
+    @Released
+    final EntryEventImpl event = EntryEventImpl.create(this, Operation.REMOVE, key, null,
         callbackArg, false, memberId.getDistributedMember(), true, clientEvent.getEventId());
 
     try {
@@ -12109,13 +12222,14 @@ public class LocalRegion extends AbstractRegion implements LoaderHelperFactory,
    * that was previously persisted, depending on whether the index previously existed.
    *
    * TODO: none of the parameters are ever used
+   *
    * @param indexName the name of the index
    * @param indexedExpression the index expression
    * @param fromClause the from clause.
    * @return The index map.
    * @throws IllegalStateException if this region is not using soplog persistence
    * @throws IllegalStateException if this index was previously persisted with a different
-   * expression or from clause.
+   *         expression or from clause.
    */
   public IndexMap getIndexMap(String indexName, String indexedExpression, String fromClause) {
     return new IndexMapImpl();
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/OfflineCompactionDiskRegion.java b/geode-core/src/main/java/org/apache/geode/internal/cache/OfflineCompactionDiskRegion.java
index 7dd2f86..2d479db 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/OfflineCompactionDiskRegion.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/OfflineCompactionDiskRegion.java
@@ -28,7 +28,8 @@ import org.apache.geode.statistics.StatsFactory;
 public class OfflineCompactionDiskRegion extends DiskRegion implements DiskRecoveryStore {
   private OfflineCompactionDiskRegion(DiskStoreImpl ds, DiskRegionView drv) {
     super(ds, drv.getName(), drv.isBucket(), true, false, true,
-        StatsFactory.createDiskRegionStatsImpl(ds.getCache().getDistributedSystem().getStatisticsFactory(), drv.getName()),
+        StatsFactory.createDiskRegionStatsImpl(
+            ds.getCache().getDistributedSystem().getStatisticsFactory(), drv.getName()),
         new DummyCancelCriterion(), new DummyDiskExceptionHandler(), null, drv.getFlags(),
         drv.getPartitionName(), drv.getStartingBucketId(), drv.getCompressorClassName(),
         drv.getOffHeap());
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/PartitionedRegion.java b/geode-core/src/main/java/org/apache/geode/internal/cache/PartitionedRegion.java
index 8df9654..c3c8b0d 100755
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/PartitionedRegion.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/PartitionedRegion.java
@@ -53,7 +53,6 @@ import org.apache.logging.log4j.Logger;
 
 import org.apache.geode.CancelException;
 import org.apache.geode.InternalGemFireException;
-import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.SystemFailure;
 import org.apache.geode.cache.AttributesFactory;
 import org.apache.geode.cache.AttributesMutator;
@@ -251,6 +250,7 @@ import org.apache.geode.internal.sequencelog.RegionLogger;
 import org.apache.geode.internal.size.Sizeable;
 import org.apache.geode.internal.util.TransformUtils;
 import org.apache.geode.internal.util.concurrent.StoppableCountDownLatch;
+import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatsFactory;
 
 /**
@@ -741,7 +741,8 @@ public class PartitionedRegion extends LocalRegion
     super(regionName, regionAttributes, parentRegion, cache, internalRegionArgs);
 
     this.node = initializeNode();
-    this.prStats = StatsFactory.createPartitionedRegionStatsImpl(cache.getDistributedSystem().getStatisticsFactory(), getFullPath());
+    this.prStats = StatsFactory.createPartitionedRegionStatsImpl(
+        cache.getDistributedSystem().getStatisticsFactory(), getFullPath());
     this.regionIdentifier = getFullPath().replace('/', '#');
 
     if (logger.isDebugEnabled()) {
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/PartitionedRegionDataStore.java b/geode-core/src/main/java/org/apache/geode/internal/cache/PartitionedRegionDataStore.java
index d874c0c..903570e 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/PartitionedRegionDataStore.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/PartitionedRegionDataStore.java
@@ -74,7 +74,6 @@ import org.apache.geode.distributed.internal.membership.InternalDistributedMembe
 import org.apache.geode.i18n.StringId;
 import org.apache.geode.internal.Assert;
 import org.apache.geode.internal.cache.BucketRegion.RawValue;
-import org.apache.geode.internal.cache.LocalRegion.RegionPerfStatsImpl;
 import org.apache.geode.internal.cache.PartitionedRegion.BucketLock;
 import org.apache.geode.internal.cache.PartitionedRegion.SizeEntry;
 import org.apache.geode.internal.cache.backup.BackupService;
@@ -189,6 +188,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
   /**
    * Creates PartitionedRegionDataStore for dataStorage of PR and starts a PartitionService to
    * handle remote operations on this DataStore from other participating nodes.
+   *
    * @param pr PartitionedRegion associated with this DataStore.
    */
   PartitionedRegionDataStore(final PartitionedRegion pr) {
@@ -217,10 +217,11 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * This method creates a PartitionedRegionDataStore be invoking the PRDS Constructor.
+   *
    * @return @throws PartitionedRegionException
    */
   static PartitionedRegionDataStore createDataStore(Cache cache, PartitionedRegion pr,
-                                                    PartitionAttributes pa)
+      PartitionAttributes pa)
       throws PartitionedRegionException {
     PartitionedRegionDataStore prd = new PartitionedRegionDataStore(pr);
     return prd;
@@ -232,6 +233,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * Test to determine if this data store is managing a bucket
+   *
    * @param bucketId the id of the bucket
    * @return true if the provided bucket is being managed
    */
@@ -277,13 +279,13 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
   }
 
   private boolean areAllColocatedPartitionedRegionsReady(int bucketId,
-                                                         List<PartitionedRegion> colocatedWithList) {
+      List<PartitionedRegion> colocatedWithList) {
     return colocatedWithList.stream().allMatch(
         partitionedRegion -> isColocatedPartitionedRegionInitialized(partitionedRegion, bucketId));
   }
 
   private boolean isColocatedPartitionedRegionInitialized(PartitionedRegion partitionedRegion,
-                                                          final int bucketId) {
+      final int bucketId) {
     if (!partitionedRegion.isInitialized()) {
       return false;
     }
@@ -304,13 +306,13 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
    */
 
   protected CreateBucketResult grabFreeBucketRecursively(final int bucketId,
-                                                         final PartitionedRegion pr,
-                                                         final InternalDistributedMember moveSource,
-                                                         final boolean forceCreation,
-                                                         final boolean isRebalance,
-                                                         final boolean replaceOfflineData,
-                                                         final InternalDistributedMember creationRequestor,
-                                                         final boolean isDiskRecovery) {
+      final PartitionedRegion pr,
+      final InternalDistributedMember moveSource,
+      final boolean forceCreation,
+      final boolean isRebalance,
+      final boolean replaceOfflineData,
+      final InternalDistributedMember creationRequestor,
+      final boolean isDiskRecovery) {
     CreateBucketResult grab;
     DistributedMember dm = pr.getMyId();
     List<PartitionedRegion> colocatedWithList = ColocationHelper.getColocatedChildRegions(pr);
@@ -355,8 +357,9 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
   /**
    * Attempts to map a bucket id to this node. Creates real storage for the bucket by adding a new
    * Region to bucket2Map. Bucket creation is done under the d-lock on b2n region.
+   *
    * @param possiblyFreeBucketId the identity of the bucket + @param mustBeNew boolean enforcing
-   * that the bucket must not already exist
+   *        that the bucket must not already exist
    * @param sender the member requesting the bucket
    * @param moveSource Where we are moving the bucket from, if this is a move.
    * @param forceCreation avoid any checks (with in reason) which might prevent bucket creation
@@ -364,11 +367,11 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
    * @return true if successful
    */
   CreateBucketResult grabFreeBucket(final int possiblyFreeBucketId, final DistributedMember sender,
-                                    final InternalDistributedMember moveSource,
-                                    final boolean forceCreation,
-                                    final boolean isRebalance, final boolean lockRedundancyLock,
-                                    boolean replaceOffineData,
-                                    InternalDistributedMember creationRequestor) {
+      final InternalDistributedMember moveSource,
+      final boolean forceCreation,
+      final boolean isRebalance, final boolean lockRedundancyLock,
+      boolean replaceOffineData,
+      InternalDistributedMember creationRequestor) {
 
     final boolean isDebugEnabled = logger.isDebugEnabled();
 
@@ -593,7 +596,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
   }
 
   public Object lockRedundancyLock(InternalDistributedMember moveSource, int bucketId,
-                                   boolean replaceOffineData) {
+      boolean replaceOffineData) {
     // TODO prperist - Make this thing easier to find!
     final PartitionedRegion.BucketLock bl = partitionedRegion.getRegionAdvisor()
         .getBucketAdvisor(bucketId).getProxyBucketRegion().getBucketLock();
@@ -639,6 +642,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * Returns false if this region is colocated and parent bucket does not exist.
+   *
    * @return true if ok to make bucket
    */
   private boolean okToCreateChildBucket(int bucketId) {
@@ -680,6 +684,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
   /**
    * This method creates bucket regions, based on redundancy level. If redundancy level is: a) = 1
    * it creates a local region b) >1 it creates a distributed region
+   *
    * @return @throws CacheException
    */
   private BucketRegion createBucketRegion(int bucketId) {
@@ -777,7 +782,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
     if (Boolean.getBoolean(DistributionConfig.GEMFIRE_PREFIX + "PRDebug")) {
       logger.info(LocalizedMessage.create(
           LocalizedStrings.PartitionedRegionDataStore_CREATEBUCKETREGION_CREATING_BUCKETID_0_NAME_1,
-          new Object[]{this.partitionedRegion.bucketStringForLogs(bucketId), bucketRegionName}));
+          new Object[] {this.partitionedRegion.bucketStringForLogs(bucketId), bucketRegionName}));
     }
     try {
       final Bucket proxyBucket = this.partitionedRegion.getRegionAdvisor().getBucket(bucketId);
@@ -852,7 +857,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
     // Build index info thats used to create indexes on bucket regions.
     indexes = new ArrayList();
     Set indexSet = indexMap.entrySet();
-    for (Iterator it = indexSet.iterator(); it.hasNext(); ) {
+    for (Iterator it = indexSet.iterator(); it.hasNext();) {
       try {
         Map.Entry indexEntry = (Map.Entry) it.next();
         PartitionedIndex index = (PartitionedIndex) indexEntry.getValue();
@@ -920,23 +925,17 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
             event.getRegion().getFullPath(), event.getKey(), event.getDistributedMember());
       }
 
-      public void afterRegionInvalidate(RegionEvent event) {
-      }
+      public void afterRegionInvalidate(RegionEvent event) {}
 
-      public void afterRegionDestroy(RegionEvent event) {
-      }
+      public void afterRegionDestroy(RegionEvent event) {}
 
-      public void afterRegionClear(RegionEvent event) {
-      }
+      public void afterRegionClear(RegionEvent event) {}
 
-      public void afterRegionCreate(RegionEvent event) {
-      }
+      public void afterRegionCreate(RegionEvent event) {}
 
-      public void afterRegionLive(RegionEvent event) {
-      }
+      public void afterRegionLive(RegionEvent event) {}
 
-      public void close() {
-      }
+      public void close() {}
     };
   }
 
@@ -982,6 +981,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
    * is the current memory (MB) watermark for data in this PR.
    *
    * If eviction to disk is enabled, this does not reflect the size of entries on disk.
+   *
    * @return the total memory size in bytes for all the Map's values
    */
   public long currentAllocatedMemory() {
@@ -991,13 +991,14 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
   /**
    * Checks if this PartitionedRegionDataStore has the capacity to handle the bucket creation
    * request. If so, creates the real storage for the bucket.
+   *
    * @param bucketId the bucket id
    * @param size the size in bytes of the bucket to create locally
    * @param forceCreation ignore local maximum buckets check
    * @return true if already managing the bucket or if the bucket has been created
    */
   public boolean handleManageBucketRequest(int bucketId, int size, InternalDistributedMember sender,
-                                           boolean forceCreation) {
+      boolean forceCreation) {
 
     // check maxMemory setting
     if (!this.partitionedRegion.isDataStore()) {
@@ -1047,6 +1048,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
   /**
    * Determine if the ratio of buckets this VM should host is appropriate given its localMaxMemory
    * setting as compared to others
+   *
    * @return true if this data store can host another bucket
    */
   boolean canAccommodateAnotherBucket() {
@@ -1097,6 +1099,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * Checks if this PartitionedRegionDataStore has the capacity to handle the rebalancing size.
+   *
    * @param size the size in bytes of the bucket to be rebalanced
    * @return true if size can be accommodated without exceeding ratioFull
    */
@@ -1110,6 +1113,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * Handles rebalance by accepting new bucket contents and storing it.
+   *
    * @param bucketId the id of the bucket to rebalance
    * @param obj the contents of the bucket
    * @param regionName the name of the PR
@@ -1124,6 +1128,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
    * storage for the bucket and a bucket2Node Region mapping. These two operations are done as a
    * logical unit so that the node can immediately begin handling remote requests once the
    * bucket2Node mapping becomes visible.
+   *
    * @param bucketId the bucket id
    */
   boolean handleRemoteCreateBackupRegion(int bucketId) {
@@ -1132,6 +1137,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * Return the size in bytes for a given bucket.
+   *
    * @return size in bytes
    */
   public long getBucketSize(int bucketId) {
@@ -1171,6 +1177,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * This method returns name of this Partitioned Region
+   *
    * @return Partitioned Region name
    */
   private String getName() {
@@ -1182,11 +1189,17 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
   // ////////////////////////////////////
 
   /**
-   * Puts the object with the given key locally. <br> Step: <br> 1) It finds out the bucket region
-   * for the bucket id. <br> 2) If from step 1 it gets null, that means the bucket is re-mapped.
-   * <br> 3) If it finds the bucket region from step 1, it tries to put the key-value on the region.
-   * <br> 4) updateBucket2Size if bucket is on more than 1 node or else bucket listeners would take
+   * Puts the object with the given key locally. <br>
+   * Step: <br>
+   * 1) It finds out the bucket region
+   * for the bucket id. <br>
+   * 2) If from step 1 it gets null, that means the bucket is re-mapped.
+   * <br>
+   * 3) If it finds the bucket region from step 1, it tries to put the key-value on the region.
+   * <br>
+   * 4) updateBucket2Size if bucket is on more than 1 node or else bucket listeners would take
    * care of size update. <br>
+   *
    * @param bucketId the bucket id of the key
    * @param event the operation event
    * @param ifNew whether a create must be performed
@@ -1197,17 +1210,17 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
    * @throws PrimaryBucketException if the bucket in this data store is not the primary bucket
    */
   public boolean putLocally(final Integer bucketId, final EntryEventImpl event, boolean ifNew,
-                            boolean ifOld, Object expectedOldValue, boolean requireOldValue,
-                            final long lastModified)
+      boolean ifOld, Object expectedOldValue, boolean requireOldValue,
+      final long lastModified)
       throws PrimaryBucketException, ForceReattemptException {
     final BucketRegion br = getInitializedBucketForId(event.getKey(), bucketId);
     return putLocally(br, event, ifNew, ifOld, expectedOldValue, requireOldValue, lastModified);
   }
 
   public boolean putLocally(final BucketRegion bucketRegion, final EntryEventImpl event,
-                            boolean ifNew, boolean ifOld, Object expectedOldValue,
-                            boolean requireOldValue,
-                            final long lastModified)
+      boolean ifNew, boolean ifOld, Object expectedOldValue,
+      boolean requireOldValue,
+      final long lastModified)
       throws PrimaryBucketException, ForceReattemptException {
     boolean didPut = false; // false if entry put fails
 
@@ -1261,7 +1274,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
         }
       }
       if (logStr != null) {
-        Object[] logArgs = new Object[]{this.partitionedRegion.getFullPath(), logStr,
+        Object[] logArgs = new Object[] {this.partitionedRegion.getFullPath(), logStr,
             Long.valueOf(this.partitionedRegion.getLocalMaxMemory()),
             Long.valueOf(locBytes / PartitionedRegionHelper.BYTES_PER_MB)};
         if (this.exceededLocalMaxMemoryLimit) {
@@ -1275,6 +1288,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * Checks whether there is room in this Map to accommodate more data.
+   *
    * @param bytes the size to check in bytes
    */
   boolean canAccommodateMoreBytesSafely(int bytes) {
@@ -1319,17 +1333,22 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
   }
 
   /**
-   * Handles the remote request to remove the key from this Map. <br> Step: <br> 1) Locates the
+   * Handles the remote request to remove the key from this Map. <br>
+   * Step: <br>
+   * 1) Locates the
    * bucket region. If it doesnt find the actual bucket, it means that this bucket is remapped to
-   * some other node and remappedBucketException is thrown <br> 2) Invokes destroy on that bucket
-   * region <br> 3) updateBucket2Size if bucket is on more than 1 node or else bucket listners would
+   * some other node and remappedBucketException is thrown <br>
+   * 2) Invokes destroy on that bucket
+   * region <br>
+   * 3) updateBucket2Size if bucket is on more than 1 node or else bucket listners would
    * take care of size update.
+   *
    * @param bucketId for the key
    * @param event the event causing this action
    * @param expectedOldValue if non-null, then only succeed if current value
    * @return the removed value
    * @throws EntryNotFoundException if entry is not found for the key or expectedOldValue is not
-   * null and current value is not equal to it
+   *         null and current value is not equal to it
    * @throws PrimaryBucketException if the locally managed bucket is not the primary
    * @throws ForceReattemptException if the bucket region is null
    */
@@ -1444,7 +1463,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
               } catch (Exception ex) {
                 logger.warn(LocalizedMessage.create(
                     LocalizedStrings.PartitionedRegion_PARTITIONEDREGION_0_CLEANUP_PROBLEM_DESTROYING_BUCKET_1,
-                    new Object[]{this.partitionedRegion.getFullPath(),
+                    new Object[] {this.partitionedRegion.getFullPath(),
                         Integer.valueOf(buk.getId())}),
                     ex);
               }
@@ -1528,6 +1547,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
    *
    * This method is now also used by the PartitionManager. For the PartitionManager, it does remove
    * the primary bucket.
+   *
    * @param bucketId the id of the bucket to remove
    * @return true if the bucket was removed; false if unable to remove or if bucket is not hosted
    */
@@ -1678,6 +1698,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
   /**
    * This calls removeBucket on every colocated child that is directly colocated to this bucket's
    * PR. Those each in turn do the same to their child buckets and so on before returning.
+   *
    * @param bucketId the bucket to remove
    * @param forceRemovePrimary true if we should remove the bucket, even if it is primary.
    * @return true if bucket was removed from all children
@@ -1702,12 +1723,13 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
   /**
    * Create a new backup of the bucket, allowing redundancy to be exceeded. All colocated child
    * buckets will also be created.
+   *
    * @param bucketId the bucket to create
    * @param isRebalance true if bucket creation is directed by rebalancing
    * @return true if the bucket and its colocated chain of children are created
    */
   public CreateBucketResult createRedundantBucket(int bucketId, boolean isRebalance,
-                                                  InternalDistributedMember moveSource) {
+      InternalDistributedMember moveSource) {
     // recurse down to create each tier of children after creating leader bucket
     return grabBucket(bucketId, moveSource, true, false, isRebalance, null, false);
   }
@@ -1717,12 +1739,13 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
    *
    * If the bucket is the leader bucket then it will recursively create all colocated children and
    * then remove all colocated children as well from the <code>source</code> member.
+   *
    * @param bucketId the bucket to move
    * @param source the member to move the bucket from
    * @return true if bucket was successfully moved to this datastore
    */
   public boolean moveBucket(int bucketId, InternalDistributedMember source,
-                            final boolean isRebalance) {
+      final boolean isRebalance) {
 
     if (createRedundantBucket(bucketId, isRebalance, source) != CreateBucketResult.CREATED) {
       if (logger.isDebugEnabled()) {
@@ -1761,6 +1784,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * Fetch a BucketRegion, but do not return until it is initialized
+   *
    * @param key optional for error reporting; if none, no key available.
    * @param bucketId the bucket to fetch
    * @return the region
@@ -1778,7 +1802,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
       ForceReattemptException fre = new BucketNotFoundException(
           LocalizedStrings.PartitionedRegionDataStore_BUCKET_ID_0_NOT_FOUND_ON_VM_1
               .toLocalizedString(
-                  new Object[]{this.partitionedRegion.bucketStringForLogs(bucketId.intValue()),
+                  new Object[] {this.partitionedRegion.bucketStringForLogs(bucketId.intValue()),
                       this.partitionedRegion.getMyId()}));
       if (key != null) {
         fre.setHash(key.hashCode());
@@ -1791,6 +1815,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * Returns the local BucketRegion given an bucketId. Returns null if no BucketRegion exists.
+   *
    * @since GemFire 6.1.2.9
    */
   public BucketRegion getLocalBucketById(Integer bucketId) {
@@ -1814,6 +1839,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
   /**
    * Test hook to return the per entry overhead for a bucket region. PRECONDITION: a bucket must
    * exist and be using LRU.
+   *
    * @since GemFire 6.1.2.9
    */
   public int getPerEntryLRUOverhead() {
@@ -1823,6 +1849,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * Fetch a BucketRegion, but do not return until it is initialized and the primary is known.
+   *
    * @return the initialized region
    * @see #getInitializedBucketForId(Object, Integer)
    */
@@ -1835,6 +1862,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * Checks if this instance contains a value for the key locally.
+   *
    * @param bucketId for the key
    * @param key the key, whose value needs to be checks
    * @return true if there is a non-null value for the given key
@@ -1885,6 +1913,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * Checks if this instance contains a key.
+   *
    * @param bucketId the bucketId for the key
    * @param key the key to look for
    * @return true if there is an entry with the given key
@@ -1944,6 +1973,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * Returns value corresponding to this key.
+   *
    * @param key the key to look for
    * @param requestingClient the client making the request, or null
    * @param clientEvent client's event (for returning version tag)
@@ -1954,9 +1984,9 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
    * @throws PRLocallyDestroyedException if the PartitionRegion is locally destroyed
    */
   public Object getLocally(int bucketId, final Object key, final Object aCallbackArgument,
-                           boolean disableCopyOnRead, boolean preferCD,
-                           ClientProxyMembershipID requestingClient,
-                           EntryEventImpl clientEvent, boolean returnTombstones)
+      boolean disableCopyOnRead, boolean preferCD,
+      ClientProxyMembershipID requestingClient,
+      EntryEventImpl clientEvent, boolean returnTombstones)
       throws PrimaryBucketException, ForceReattemptException, PRLocallyDestroyedException {
     return getLocally(bucketId, key, aCallbackArgument, disableCopyOnRead, preferCD,
         requestingClient, clientEvent, returnTombstones, false);
@@ -1964,22 +1994,23 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * Returns value corresponding to this key.
+   *
    * @param key the key to look for
    * @param requestingClient the client making the request, or null
    * @param clientEvent client's event (for returning version tag)
    * @param returnTombstones whether tombstones should be returned
    * @param opScopeIsLocal if true then just check local storage for a value; if false then try to
-   * find the value if it is not local
+   *        find the value if it is not local
    * @return value from the bucket region
    * @throws ForceReattemptException if bucket region is null
    * @throws PrimaryBucketException if the locally managed bucket is not primary
    * @throws PRLocallyDestroyedException if the PartitionRegion is locally destroyed
    */
   public Object getLocally(int bucketId, final Object key, final Object aCallbackArgument,
-                           boolean disableCopyOnRead, boolean preferCD,
-                           ClientProxyMembershipID requestingClient,
-                           EntryEventImpl clientEvent, boolean returnTombstones,
-                           boolean opScopeIsLocal)
+      boolean disableCopyOnRead, boolean preferCD,
+      ClientProxyMembershipID requestingClient,
+      EntryEventImpl clientEvent, boolean returnTombstones,
+      boolean opScopeIsLocal)
       throws PrimaryBucketException, ForceReattemptException, PRLocallyDestroyedException {
     final BucketRegion bucketRegion = getInitializedBucketForId(key, Integer.valueOf(bucketId));
     // check for primary (when a loader is present) done deeper in the BucketRegion
@@ -2013,6 +2044,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * Return a value from the bucket region, always serialized
+   *
    * @param keyInfo TODO
    * @param clientEvent a "client" event that will hold version information about the entry
    * @param returnTombstones TODO
@@ -2020,12 +2052,12 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
    * @throws ForceReattemptException if bucket region is null
    * @throws PrimaryBucketException if the locally managed bucket is not primary
    * @see #getLocally(int, Object, Object, boolean, boolean, ClientProxyMembershipID,
-   * EntryEventImpl, boolean)
+   *      EntryEventImpl, boolean)
    */
   public RawValue getSerializedLocally(KeyInfo keyInfo, boolean doNotLockEntry,
-                                       ClientProxyMembershipID requestingClient,
-                                       EntryEventImpl clientEvent,
-                                       boolean returnTombstones)
+      ClientProxyMembershipID requestingClient,
+      EntryEventImpl clientEvent,
+      boolean returnTombstones)
       throws PrimaryBucketException, ForceReattemptException {
     final BucketRegion bucketRegion =
         getInitializedBucketForId(keyInfo.getKey(), keyInfo.getBucketId());
@@ -2058,6 +2090,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * Finds the local bucket corresponding to the given key and retrieves the key's Region.Entry
+   *
    * @param key the key to look for
    * @param access true if caller wants last accessed time updated
    * @param allowTombstones whether a tombstoned entry can be returned
@@ -2067,7 +2100,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
    * @throws PRLocallyDestroyedException if the PartitionRegion is locally destroyed
    */
   public EntrySnapshot getEntryLocally(int bucketId, final Object key, boolean access,
-                                       boolean allowTombstones)
+      boolean allowTombstones)
       throws EntryNotFoundException, PrimaryBucketException,
       ForceReattemptException, PRLocallyDestroyedException {
     final BucketRegion bucketRegion = getInitializedBucketForId(key, Integer.valueOf(bucketId));
@@ -2122,13 +2155,14 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * Handle a remote request for keys for the provided bucketId
+   *
    * @param allowTombstones whether to return destroyed entries
    * @return The <code>Set</code> of keys for bucketId or {@link Collections#EMPTY_SET}if no keys
-   * are present
+   *         are present
    * @throws PRLocallyDestroyedException if the PartitionRegion is locally destroyed
    */
   public Set handleRemoteGetKeys(Integer bucketId, int interestType, Object interestArg,
-                                 boolean allowTombstones)
+      boolean allowTombstones)
       throws PRLocallyDestroyedException, ForceReattemptException {
     if (logger.isDebugEnabled()) {
       logger.debug("handleRemoteGetKeys: bucketId: {}{}{} with tombstones {}",
@@ -2168,9 +2202,10 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
    * Get the local keys for a given bucket. This operation should be as efficient as possible, by
    * avoiding making copies of the returned set. The returned set can and should reflect concurrent
    * changes (no ConcurrentModificationExceptions).
+   *
    * @param allowTombstones whether to include destroyed entries in the result
    * @return The <code>Set</code> of keys for bucketId or {@link Collections#EMPTY_SET} if no keys
-   * are present
+   *         are present
    * @throws PRLocallyDestroyedException if the PartitionRegion is locally destroyed
    */
   public Set getKeysLocally(Integer bucketId, boolean allowTombstones)
@@ -2220,12 +2255,19 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
   }
 
   /**
-   * Creates the entry with the given key locally. <br> Steps: <br> 1) It finds out the bucket
-   * region for the bucket id. <br> 2) If from step 1 it gets null, that means the bucket is
-   * remapped. <br> 3) If it finds the bucket region from step 1, it tries to creates the entry on
-   * the region. <br> 4) If entry already exists, for the key, step 3 would throw
-   * EntryExistsException or else it will create an entry <br> 5) updateBucket2Size if bucket is on
+   * Creates the entry with the given key locally. <br>
+   * Steps: <br>
+   * 1) It finds out the bucket
+   * region for the bucket id. <br>
+   * 2) If from step 1 it gets null, that means the bucket is
+   * remapped. <br>
+   * 3) If it finds the bucket region from step 1, it tries to creates the entry on
+   * the region. <br>
+   * 4) If entry already exists, for the key, step 3 would throw
+   * EntryExistsException or else it will create an entry <br>
+   * 5) updateBucket2Size if bucket is on
    * more than 1 node or else bucket listners would take care of size update. <br>
+   *
    * @param bucketRegion the bucket to do the create in
    * @param event the particulars of the operation
    * @param ifNew whether a new entry can be created
@@ -2235,8 +2277,8 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
    * @throws EntryExistsException if an entry with this key already exists
    */
   public boolean createLocally(final BucketRegion bucketRegion, final EntryEventImpl event,
-                               boolean ifNew, boolean ifOld, boolean requireOldValue,
-                               final long lastModified)
+      boolean ifNew, boolean ifOld, boolean requireOldValue,
+      final long lastModified)
       throws ForceReattemptException {
     boolean result = false;
     try {
@@ -2268,13 +2310,20 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
   }
 
   /**
-   * Handles the local request to invalidate the key from this region. <br> Steps: <br> 1) Locates
+   * Handles the local request to invalidate the key from this region. <br>
+   * Steps: <br>
+   * 1) Locates
    * the bucket region. If it doesnt find the actual bucket, it means that this bucket is remapped
-   * to some other node and remappedBucketException is thrown <br> 2) get the existing value for the
-   * key from bucket region <br> 3) if step 2 returns null, throw EntryNotFoundException <br> 4) if
+   * to some other node and remappedBucketException is thrown <br>
+   * 2) get the existing value for the
+   * key from bucket region <br>
+   * 3) if step 2 returns null, throw EntryNotFoundException <br>
+   * 4) if
    * step 2 returns non-null value, perform invalidate on the bucket region and use value from step
-   * 2 in step 5 <br> 5) updateBucket2Size if bucket is on more than 1 node or else bucket listners
+   * 2 in step 5 <br>
+   * 5) updateBucket2Size if bucket is on more than 1 node or else bucket listners
    * would take care of size update. <br>
+   *
    * @param bucketId the bucketId for the key
    * @param event the event that prompted this action
    * @throws ForceReattemptException if bucket region is null
@@ -2308,13 +2357,19 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * This method iterates over localBucket2RegionMap and returns collective size of the bucket
-   * regions. <br> Steps: <br> 1) Check if localBucket2RegionMap is empty. If it is, return 0.<br>
-   * 2) If localBucket2RegionMap is not empty, get keyset of all these bucket IDs. <br> 3) Get the
-   * nodeList for the bucket ID from Bucket2Node region. <br> 4) If first node from the node list is
-   * current node, increment the size counter. <br> 5) Step#4 takes care of the problem of
+   * regions. <br>
+   * Steps: <br>
+   * 1) Check if localBucket2RegionMap is empty. If it is, return 0.<br>
+   * 2) If localBucket2RegionMap is not empty, get keyset of all these bucket IDs. <br>
+   * 3) Get the
+   * nodeList for the bucket ID from Bucket2Node region. <br>
+   * 4) If first node from the node list is
+   * current node, increment the size counter. <br>
+   * 5) Step#4 takes care of the problem of
    * recounting the size of redundant buckets. <br>
+   *
    * @return the map of bucketIds and their associated sizes, or {@link Collections#EMPTY_MAP}when
-   * the size is zero
+   *         the size is zero
    */
   public Map getSizeLocally() {
     return getSizeLocally(false);
@@ -2334,7 +2389,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
     Map.Entry<Integer, BucketRegion> me;
     BucketRegion r;
     for (Iterator<Map.Entry<Integer, BucketRegion>> itr =
-         this.localBucket2RegionMap.entrySet().iterator(); itr.hasNext(); ) {
+        this.localBucket2RegionMap.entrySet().iterator(); itr.hasNext();) {
       me = itr.next();
       try {
         r = me.getValue();
@@ -2362,8 +2417,9 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
   /**
    * This method iterates over localBucket2RegionMap and returns collective size of the primary
    * bucket regions.
+   *
    * @return the map of bucketIds and their associated sizes, or {@link Collections#EMPTY_MAP}when
-   * the size is zero
+   *         the size is zero
    */
   public Map<Integer, SizeEntry> getSizeForLocalBuckets() {
     return getSizeLocallyForBuckets(this.localBucket2RegionMap.keySet());
@@ -2379,6 +2435,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * This calculates size of all the primary bucket regions for the list of bucketIds.
+   *
    * @return the size of all the primary bucket regions for the list of bucketIds.
    */
   public Map<Integer, SizeEntry> getSizeLocallyForBuckets(Collection<Integer> bucketIds) {
@@ -2390,7 +2447,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
   }
 
   private Map<Integer, SizeEntry> getSizeLocallyForPrimary(Collection<Integer> bucketIds,
-                                                           boolean estimate) {
+      boolean estimate) {
     Map<Integer, SizeEntry> mySizeMap;
     if (this.localBucket2RegionMap.isEmpty()) {
       return Collections.emptyMap();
@@ -2450,7 +2507,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
   public void visitBuckets(final BucketVisitor bv) {
     if (this.localBucket2RegionMap.size() > 0) {
       Map.Entry me;
-      for (Iterator i = this.localBucket2RegionMap.entrySet().iterator(); i.hasNext(); ) {
+      for (Iterator i = this.localBucket2RegionMap.entrySet().iterator(); i.hasNext();) {
         me = (Map.Entry) i.next();
         Region r = (Region) me.getValue();
         // ConcurrentHashMap entrySet iterator does not guarantee an atomic snapshot
@@ -2468,9 +2525,9 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
   }
 
   private void visitBucket(final Integer bucketId, final LocalRegion bucket,
-                           final EntryVisitor ev) {
+      final EntryVisitor ev) {
     try {
-      for (Iterator ei = bucket.entrySet().iterator(); ei.hasNext(); ) {
+      for (Iterator ei = bucket.entrySet().iterator(); ei.hasNext();) {
         ev.visit(bucketId, (Region.Entry) ei.next());
       }
     } catch (CacheRuntimeException ignore) {
@@ -2494,7 +2551,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
       public void visit(Integer bucketId, Region buk) {
         try {
           ((LocalRegion) buk).waitForData();
-          for (Iterator ei = buk.entrySet().iterator(); ei.hasNext(); ) {
+          for (Iterator ei = buk.entrySet().iterator(); ei.hasNext();) {
             knock.visit(bucketId, (Region.Entry) ei.next());
           }
         } catch (CacheRuntimeException ignore) {
@@ -2506,6 +2563,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * <i>Test Method</i> Return the list of PartitionedRegion entries contained in this data store
+   *
    * @return a List of all entries gathered across all buckets in this data store
    */
   public List getEntries() {
@@ -2519,8 +2577,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
       }
 
       @Override
-      public void finishedVisiting() {
-      }
+      public void finishedVisiting() {}
     });
     return al;
   }
@@ -2528,6 +2585,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
   /**
    * <i>Test Method</i> Dump all the entries in all the buckets to the logger, validate that the
    * bucket-to-node meta region contains all bhe buckets managed by this data store
+   *
    * @param validateOnly only perform bucket-to-node validation
    */
   public void dumpEntries(final boolean validateOnly) {
@@ -2640,6 +2698,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * <i>Test Method</i> Dump the entries in this given bucket to the logger
+   *
    * @param bucketId the id of the bucket to dump
    * @param bucket the Region containing the bucket data
    */
@@ -2664,6 +2723,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * Fetch the entries for the given bucket
+   *
    * @param bucketId the id of the bucket
    * @return a Map containing all the entries
    */
@@ -2685,6 +2745,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
   /**
    * Return a set of local buckets. Iterators may include entries with null values (but non-null
    * keys).
+   *
    * @return an unmodifiable set of Map.Entry objects
    */
   public Set<Map.Entry<Integer, BucketRegion>> getAllLocalBuckets() {
@@ -2697,6 +2758,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * Returns a set of local buckets.
+   *
    * @return an unmodifiable set of BucketRegion
    */
   public Set<BucketRegion> getAllLocalBucketRegions() {
@@ -2807,11 +2869,12 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
   /**
    * Checks for RegionDestroyedException in case of remoteEvent & localDestroy OR isClosed throws a
    * ForceReattemptException
+   *
    * @param br the bucket that we are trying to operate on
    * @param isOriginRemote true the event we are processing has a remote origin.
    */
   public void checkRegionDestroyedOnBucket(final BucketRegion br, final boolean isOriginRemote,
-                                           RegionDestroyedException rde)
+      RegionDestroyedException rde)
       throws ForceReattemptException {
     if (isOriginRemote) {
       if (logger.isDebugEnabled()) {
@@ -2839,24 +2902,25 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * Create a redundancy bucket on this member
+   *
    * @param bucketId the id of the bucket to create
    * @param moveSource the member id of where the bucket is being copied from, if this is a bucket
-   * move. Setting this field means that we are allowed to go over redundancy.
+   *        move. Setting this field means that we are allowed to go over redundancy.
    * @param forceCreation force the bucket creation, even if it would provide better balance if the
-   * bucket was placed on another node.
+   *        bucket was placed on another node.
    * @param replaceOffineData create the bucket, even if redundancy is satisfied when considering
-   * offline members.
+   *        offline members.
    * @param isRebalance true if this is a rebalance
    * @param creationRequestor the id of the member that is atomically creating this bucket on all
-   * members, if this is an atomic bucket creation.
+   *        members, if this is an atomic bucket creation.
    * @return the status of the bucket creation.
    */
   public CreateBucketResult grabBucket(final int bucketId,
-                                       final InternalDistributedMember moveSource,
-                                       final boolean forceCreation,
-                                       final boolean replaceOffineData, final boolean isRebalance,
-                                       final InternalDistributedMember creationRequestor,
-                                       final boolean isDiskRecovery) {
+      final InternalDistributedMember moveSource,
+      final boolean forceCreation,
+      final boolean replaceOffineData, final boolean isRebalance,
+      final InternalDistributedMember creationRequestor,
+      final boolean isDiskRecovery) {
     CreateBucketResult grab = grabFreeBucket(bucketId, partitionedRegion.getMyId(), moveSource,
         forceCreation, isRebalance, true, replaceOffineData, creationRequestor);
     if (!grab.nowExists()) {
@@ -2909,8 +2973,9 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
 
   /**
    * Checks consistency of bucket and meta data before attempting to grab the bucket.
+   *
    * @return false if bucket should not be grabbed, else true. TODO prpersist - move this to
-   * BucketRegion
+   *         BucketRegion
    */
   public boolean verifyBucketBeforeGrabbing(final int buckId) {
     // Consistency checks
@@ -2921,7 +2986,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
         Set owners = partitionedRegion.getRegionAdvisor().getBucketOwners(buckId);
         logger.info(LocalizedMessage.create(
             LocalizedStrings.PartitionedRegionDataStore_VERIFIED_NODELIST_FOR_BUCKETID_0_IS_1,
-            new Object[]{partitionedRegion.bucketStringForLogs(buckId),
+            new Object[] {partitionedRegion.bucketStringForLogs(buckId),
                 PartitionedRegionHelper.printCollection(owners)}));
         Assert.assertTrue(false,
             " This node " + partitionedRegion.getNode() + " is managing the bucket with bucketId= "
@@ -2954,11 +3019,11 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
   }
 
   public void executeOnDataStore(final Set localKeys, final Function function, final Object object,
-                                 final int prid, final Set<Integer> bucketSet,
-                                 final boolean isReExecute,
-                                 final PartitionedRegionFunctionStreamingMessage msg, long time,
-                                 ServerConnection servConn,
-                                 int transactionID) {
+      final int prid, final Set<Integer> bucketSet,
+      final boolean isReExecute,
+      final PartitionedRegionFunctionStreamingMessage msg, long time,
+      ServerConnection servConn,
+      int transactionID) {
 
     if (!areAllBucketsHosted(bucketSet)) {
       throw new BucketMovedException(
@@ -2972,7 +3037,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
     final RegionFunctionContextImpl prContext =
         new RegionFunctionContextImpl(getPartitionedRegion().getCache(), function.getId(),
             this.partitionedRegion, object, localKeys, ColocationHelper
-            .constructAndGetAllColocatedLocalDataSet(this.partitionedRegion, bucketSet),
+                .constructAndGetAllColocatedLocalDataSet(this.partitionedRegion, bucketSet),
             bucketSet, resultSender, isReExecute);
 
     FunctionStats stats = FunctionStatsImpl.getFunctionStats(function.getId(), dm.getSystem());
@@ -3036,7 +3101,7 @@ public class PartitionedRegionDataStore implements HasCachePerfStats {
     }
     synchronized (keysOfInterestLock) {
       boolean isRegister = event.isRegister();
-      for (Iterator i = event.getKeysOfInterest().iterator(); i.hasNext(); ) {
+      for (Iterator i = event.getKeysOfInterest().iterator(); i.hasNext();) {
         Object key = i.next();
         // Get the reference counter for this key
         AtomicInteger references = (AtomicInteger) this.keysOfInterest.get(key);
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/PartitionedRegionHelper.java b/geode-core/src/main/java/org/apache/geode/internal/cache/PartitionedRegionHelper.java
index a2a7f91..94a5506 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/PartitionedRegionHelper.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/PartitionedRegionHelper.java
@@ -261,7 +261,8 @@ public class PartitionedRegionHelper {
       RegionAttributes ra = factory.create();
       // Create anonymous stats holder for Partitioned Region meta data
       final HasCachePerfStats prMetaStatsHolder =
-          () -> StatsFactory.createCachePerfStatsImpl(cache.getDistributedSystem().getStatisticsFactory(), "partitionMetaData");
+          () -> StatsFactory.createCachePerfStatsImpl(
+              cache.getDistributedSystem().getStatisticsFactory(), "partitionMetaData");
 
       try {
         root = (DistributedRegion) cache.createVMRegion(PR_ROOT_REGION_NAME, ra,
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/PartitionedRegionStatsImpl.java b/geode-core/src/main/java/org/apache/geode/internal/cache/PartitionedRegionStatsImpl.java
index b1bdaad..50a28e9 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/PartitionedRegionStatsImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/PartitionedRegionStatsImpl.java
@@ -19,12 +19,12 @@ import java.util.Collections;
 import java.util.Map;
 import java.util.concurrent.ConcurrentHashMap;
 
+import org.apache.geode.cache.Region;
 import org.apache.geode.statistics.GFSStatsImplementor;
 import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.cache.Region;
 
 /**
  * Represents a statistics type that can be archived to vsd. Loading of this class automatically
@@ -38,6 +38,7 @@ import org.apache.geode.cache.Region;
  * <p>
  *
  * To manipulate the statistic values, use (inc|dec|set|get)&lt;fieldName&gt; methods.
+ *
  * @since GemFire 5.0
  */
 public class PartitionedRegionStatsImpl implements PartitionedRegionStats, GFSStatsImplementor {
@@ -177,7 +178,7 @@ public class PartitionedRegionStatsImpl implements PartitionedRegionStats, GFSSt
     final boolean largerIsBetter = true;
     type = factory.createType("PartitionedRegionStats",
         "Statistics for operations and connections in the Partitioned Region",
-        new StatisticDescriptor[]{
+        new StatisticDescriptor[] {
 
             factory.createIntGauge("bucketCount", "Number of buckets in this node.", "buckets"),
             factory.createIntCounter("putsCompleted", "Number of puts completed.", "operations",
@@ -1038,6 +1039,7 @@ public class PartitionedRegionStatsImpl implements PartitionedRegionStats, GFSSt
 
   /**
    * Statistic to track the {@link Region#getEntry(Object)} call
+   *
    * @param startTime the time the getEntry operation started
    */
   @Override
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/PoolStatsImpl.java b/geode-core/src/main/java/org/apache/geode/internal/cache/PoolStatsImpl.java
index d4e1b67..78a2da3 100755
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/PoolStatsImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/PoolStatsImpl.java
@@ -21,10 +21,10 @@ import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.distributed.internal.DistributionStats;
 
 /**
  * GemFire statistics about a Pool
+ *
  * @since GemFire 5.7
  */
 public class PoolStatsImpl implements PoolStats, GFSStatsImplementor {
@@ -77,7 +77,7 @@ public class PoolStatsImpl implements PoolStats, GFSStatsImplementor {
     String statName = "PoolStats";
 
     _type = factory.createType(statName, statName,
-        new StatisticDescriptor[]{
+        new StatisticDescriptor[] {
             factory.createIntGauge(INITIAL_CONTACTS, "Number of contacts initially by user",
                 "contacts"),
             factory.createIntGauge(KNOWN_LOCATORS, "Current number of locators discovered",
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/TXRemoteCommitMessage.java b/geode-core/src/main/java/org/apache/geode/internal/cache/TXRemoteCommitMessage.java
index 5be23d7..a296976 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/TXRemoteCommitMessage.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/TXRemoteCommitMessage.java
@@ -28,7 +28,6 @@ import org.apache.geode.distributed.DistributedMember;
 import org.apache.geode.distributed.internal.ClusterDistributionManager;
 import org.apache.geode.distributed.internal.DistributionManager;
 import org.apache.geode.distributed.internal.DistributionMessage;
-import org.apache.geode.distributed.internal.DistributionStats;
 import org.apache.geode.distributed.internal.DistributionStatsImpl;
 import org.apache.geode.distributed.internal.InternalDistributedSystem;
 import org.apache.geode.distributed.internal.ReplyMessage;
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/ValidatingDiskRegion.java b/geode-core/src/main/java/org/apache/geode/internal/cache/ValidatingDiskRegion.java
index 2d5d19c..964a93e 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/ValidatingDiskRegion.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/ValidatingDiskRegion.java
@@ -50,14 +50,22 @@ import org.apache.geode.statistics.StatsFactory;
 public class ValidatingDiskRegion extends DiskRegion implements DiskRecoveryStore {
   protected ValidatingDiskRegion(DiskStoreImpl diskStore, DiskRegionView diskRegionView) {
     super(diskStore, diskRegionView.getName(), diskRegionView.isBucket(), true, false, true,
-        StatsFactory.createDiskRegionStatsImpl(diskStore.getCache().getDistributedSystem().getStatisticsFactory(), diskRegionView.getName()),
-        new DummyCancelCriterion(), new DummyDiskExceptionHandler(), null, diskRegionView.getFlags(),
-        diskRegionView.getPartitionName(), diskRegionView.getStartingBucketId(), diskRegionView.getCompressorClassName(),
+        StatsFactory.createDiskRegionStatsImpl(
+            diskStore.getCache().getDistributedSystem().getStatisticsFactory(),
+            diskRegionView.getName()),
+        new DummyCancelCriterion(), new DummyDiskExceptionHandler(), null,
+        diskRegionView.getFlags(),
+        diskRegionView.getPartitionName(), diskRegionView.getStartingBucketId(),
+        diskRegionView.getCompressorClassName(),
+        diskRegionView.getOffHeap());
+    setConfig(diskRegionView.getLruAlgorithm(), diskRegionView.getLruAction(),
+        diskRegionView.getLruLimit(),
+        diskRegionView.getConcurrencyLevel(), diskRegionView.getInitialCapacity(),
+        diskRegionView.getLoadFactor(),
+        diskRegionView.getStatisticsEnabled(), diskRegionView.isBucket(), diskRegionView.getFlags(),
+        diskRegionView.getPartitionName(),
+        diskRegionView.getStartingBucketId(), diskRegionView.getCompressorClassName(),
         diskRegionView.getOffHeap());
-    setConfig(diskRegionView.getLruAlgorithm(), diskRegionView.getLruAction(), diskRegionView.getLruLimit(),
-        diskRegionView.getConcurrencyLevel(), diskRegionView.getInitialCapacity(), diskRegionView.getLoadFactor(),
-        diskRegionView.getStatisticsEnabled(), diskRegionView.isBucket(), diskRegionView.getFlags(), diskRegionView.getPartitionName(),
-        diskRegionView.getStartingBucketId(), diskRegionView.getCompressorClassName(), diskRegionView.getOffHeap());
   }
 
   static ValidatingDiskRegion create(DiskStoreImpl dsi, DiskRegionView drv) {
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/client/protocol/ClientProtocolService.java b/geode-core/src/main/java/org/apache/geode/internal/cache/client/protocol/ClientProtocolService.java
index 8e4d979..2e4c81c 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/client/protocol/ClientProtocolService.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/client/protocol/ClientProtocolService.java
@@ -15,10 +15,10 @@
 
 package org.apache.geode.internal.cache.client.protocol;
 
-import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.distributed.internal.InternalLocator;
 import org.apache.geode.internal.cache.InternalCache;
 import org.apache.geode.internal.security.SecurityService;
+import org.apache.geode.statistics.StatisticsFactory;
 
 /**
  * Provides a convenient location for a client protocol service to be loaded into the system.
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/control/HeapMemoryMonitor.java b/geode-core/src/main/java/org/apache/geode/internal/cache/control/HeapMemoryMonitor.java
index 712e5e0..c2c662d 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/control/HeapMemoryMonitor.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/control/HeapMemoryMonitor.java
@@ -280,7 +280,8 @@ public class HeapMemoryMonitor implements NotificationListener, MemoryMonitor {
       }
 
       // Stop the stats listener
-      final GemFireStatSampler sampler = this.cache.getInternalDistributedSystem().getInternalDistributedSystemStats().getStatSampler();
+      final GemFireStatSampler sampler = this.cache.getInternalDistributedSystem()
+          .getInternalDistributedSystemStats().getStatSampler();
       if (sampler != null) {
         sampler.removeLocalStatListener(this.statListener);
       }
@@ -295,7 +296,8 @@ public class HeapMemoryMonitor implements NotificationListener, MemoryMonitor {
    * @return True of the listener was correctly started, false otherwise.
    */
   private boolean startCacheStatListener() {
-    final GemFireStatSampler sampler = this.cache.getInternalDistributedSystem().getInternalDistributedSystemStats().getStatSampler();
+    final GemFireStatSampler sampler = this.cache.getInternalDistributedSystem()
+        .getInternalDistributedSystemStats().getStatSampler();
     if (sampler == null) {
       return false;
     }
@@ -303,7 +305,8 @@ public class HeapMemoryMonitor implements NotificationListener, MemoryMonitor {
     try {
       sampler.waitForInitialization();
       String tenuredPoolName = getTenuredMemoryPoolMXBean().getName();
-      List list = this.cache.getInternalDistributedSystem().getInternalDistributedSystemStats().getStatsList();
+      List list = this.cache.getInternalDistributedSystem().getInternalDistributedSystemStats()
+          .getStatsList();
       for (Object o : list) {
         if (o instanceof StatisticsImpl) {
           StatisticsImpl statisticsImpl = (StatisticsImpl) o;
@@ -311,7 +314,8 @@ public class HeapMemoryMonitor implements NotificationListener, MemoryMonitor {
               && statisticsImpl.getType().getName().contains("PoolStats")) {
             sampler.addLocalStatListener(this.statListener, statisticsImpl, "currentUsedMemory");
             if (this.cache.getLoggerI18n().fineEnabled()) {
-              this.cache.getLoggerI18n().fine("Registered stat listener for " + statisticsImpl.getTextId());
+              this.cache.getLoggerI18n()
+                  .fine("Registered stat listener for " + statisticsImpl.getTextId());
             }
 
             return true;
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/control/InternalResourceManager.java b/geode-core/src/main/java/org/apache/geode/internal/cache/control/InternalResourceManager.java
index 311c9b7..deed1f9 100755
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/control/InternalResourceManager.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/control/InternalResourceManager.java
@@ -115,7 +115,8 @@ public class InternalResourceManager implements ResourceManager {
   private InternalResourceManager(InternalCache cache) {
     this.cache = cache;
     this.resourceAdvisor = (ResourceAdvisor) cache.getDistributionAdvisor();
-    this.stats = StatsFactory.createResourceManagerStatsImpl(cache.getDistributedSystem().getStatisticsFactory());
+    this.stats = StatsFactory
+        .createResourceManagerStatsImpl(cache.getDistributedSystem().getStatisticsFactory());
 
     // Create a new executor that other classes may use for handling resource
     // related tasks
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/control/ResourceManagerStatsImpl.java b/geode-core/src/main/java/org/apache/geode/internal/cache/control/ResourceManagerStatsImpl.java
index fea2b3d..c904177 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/control/ResourceManagerStatsImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/control/ResourceManagerStatsImpl.java
@@ -14,15 +14,13 @@
  */
 package org.apache.geode.internal.cache.control;
 
+import org.apache.geode.distributed.internal.PoolStatHelper;
+import org.apache.geode.distributed.internal.QueueStatHelper;
 import org.apache.geode.statistics.GFSStatsImplementor;
 import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.statistics.StatisticsTypeFactory;
-import org.apache.geode.distributed.internal.PoolStatHelper;
-import org.apache.geode.distributed.internal.QueueStatHelper;
-import org.apache.geode.internal.statistics.StatisticsTypeFactoryImpl;
 
 /**
  * Contains methods for manipulating resource manager statistics.
@@ -76,7 +74,7 @@ public class ResourceManagerStatsImpl implements ResourceManagerStats, GFSStatsI
   @Override
   public void initializeStats(StatisticsFactory factory) {
     type = factory.createType("ResourceManagerStats", "Statistics about resource management",
-        new StatisticDescriptor[]{
+        new StatisticDescriptor[] {
             factory.createIntGauge("rebalancesInProgress",
                 "Current number of cache rebalance operations being directed by this process.",
                 "operations"),
@@ -589,7 +587,7 @@ public class ResourceManagerStatsImpl implements ResourceManagerStats, GFSStatsI
 
   /**
    * @return a {@link QueueStatHelper} so that we can record number of events in the
-   * thresholdEventProcessor queue.
+   *         thresholdEventProcessor queue.
    */
   @Override
   public QueueStatHelper getResourceEventQueueStatHelper() {
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/eviction/AbstractEvictionController.java b/geode-core/src/main/java/org/apache/geode/internal/cache/eviction/AbstractEvictionController.java
index 2ee9cef..a8ea786 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/eviction/AbstractEvictionController.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/eviction/AbstractEvictionController.java
@@ -14,12 +14,12 @@
  */
 package org.apache.geode.internal.cache.eviction;
 
-import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.cache.EvictionAction;
 import org.apache.geode.cache.EvictionAlgorithm;
 import org.apache.geode.cache.EvictionAttributes;
 import org.apache.geode.cache.util.ObjectSizer;
 import org.apache.geode.internal.cache.BucketRegion;
+import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatsFactory;
 
 /**
@@ -49,6 +49,7 @@ import org.apache.geode.statistics.StatsFactory;
  * result, an instance of {@code AbstractEvictionController} cannot be shared among multiple
  * regions. Attempts to create a region with a LRU-based capacity controller that has already been
  * used to create another region will result in an {@link IllegalStateException} being thrown.
+ *
  * @since GemFire 3.2
  */
 public abstract class AbstractEvictionController implements EvictionController {
@@ -57,7 +58,7 @@ public abstract class AbstractEvictionController implements EvictionController {
    * Create and return the appropriate eviction controller using the attributes provided.
    */
   public static EvictionController create(EvictionAttributes evictionAttributes, boolean isOffHeap,
-                                          StatisticsFactory statsFactory, String statsName) {
+      StatisticsFactory statsFactory, String statsName) {
     EvictionAlgorithm algorithm = evictionAttributes.getAlgorithm();
     EvictionAction action = evictionAttributes.getAction();
     ObjectSizer sizer = evictionAttributes.getObjectSizer();
@@ -101,7 +102,7 @@ public abstract class AbstractEvictionController implements EvictionController {
    * eviction action}.
    */
   protected AbstractEvictionController(EvictionCounters evictionCounters,
-                                       EvictionAction evictionAction, EvictionAlgorithm algorithm) {
+      EvictionAction evictionAction, EvictionAlgorithm algorithm) {
     this.counters = evictionCounters;
     this.evictionAction = evictionAction;
     this.algorithm = algorithm;
@@ -109,6 +110,7 @@ public abstract class AbstractEvictionController implements EvictionController {
 
   /**
    * Force subclasses to have a reasonable {@code toString}
+   *
    * @since GemFire 4.0
    */
   @Override
@@ -117,8 +119,9 @@ public abstract class AbstractEvictionController implements EvictionController {
   /**
    * Gets the action that is performed on the least recently used entry when it is evicted from the
    * VM.
+   *
    * @return one of the following constants: {@link EvictionAction#LOCAL_DESTROY}, {@link
-   * EvictionAction#OVERFLOW_TO_DISK}
+   *         EvictionAction#OVERFLOW_TO_DISK}
    */
   @Override
   public EvictionAction getEvictionAction() {
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/eviction/CountLRUStatisticsImpl.java b/geode-core/src/main/java/org/apache/geode/internal/cache/eviction/CountLRUStatisticsImpl.java
index 0bf3c25..f5608c9 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/eviction/CountLRUStatisticsImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/eviction/CountLRUStatisticsImpl.java
@@ -19,8 +19,6 @@ import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.statistics.StatisticsTypeFactory;
-import org.apache.geode.internal.statistics.StatisticsTypeFactoryImpl;
 
 public class CountLRUStatisticsImpl implements CountLRUEvictionStats, GFSStatsImplementor {
   private StatisticsType statType;
@@ -41,14 +39,15 @@ public class CountLRUStatisticsImpl implements CountLRUEvictionStats, GFSStatsIm
     final String lruEvaluationsDesc = "Number of entries evaluated during LRU operations.";
     final String lruGreedyReturnsDesc = "Number of non-LRU entries evicted during LRU operations";
 
-    statType = factory.createType("LRUStatistics", "Statistics relates to entry cout based eviction",
-        new StatisticDescriptor[] {
-            factory.createLongGauge("entriesAllowed", entriesAllowedDesc, "entries"),
-            factory.createLongGauge("entryCount", regionEntryCountDesc, "entries"),
-            factory.createLongCounter("lruEvictions", lruEvictionsDesc, "entries"),
-            factory.createLongCounter("lruDestroys", lruDestroysDesc, "entries"),
-            factory.createLongCounter("lruEvaluations", lruEvaluationsDesc, "entries"),
-            factory.createLongCounter("lruGreedyReturns", lruGreedyReturnsDesc, "entries")});
+    statType =
+        factory.createType("LRUStatistics", "Statistics relates to entry cout based eviction",
+            new StatisticDescriptor[] {
+                factory.createLongGauge("entriesAllowed", entriesAllowedDesc, "entries"),
+                factory.createLongGauge("entryCount", regionEntryCountDesc, "entries"),
+                factory.createLongCounter("lruEvictions", lruEvictionsDesc, "entries"),
+                factory.createLongCounter("lruDestroys", lruDestroysDesc, "entries"),
+                factory.createLongCounter("lruEvaluations", lruEvaluationsDesc, "entries"),
+                factory.createLongCounter("lruGreedyReturns", lruGreedyReturnsDesc, "entries")});
 
     limitId = statType.nameToId("entriesAllowed");
     counterId = statType.nameToId("entryCount");
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/eviction/HeapLRUStatisticsImpl.java b/geode-core/src/main/java/org/apache/geode/internal/cache/eviction/HeapLRUStatisticsImpl.java
index 4af3cc9..d7d6004 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/eviction/HeapLRUStatisticsImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/eviction/HeapLRUStatisticsImpl.java
@@ -19,8 +19,6 @@ import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.statistics.StatisticsTypeFactory;
-import org.apache.geode.internal.statistics.StatisticsTypeFactoryImpl;
 
 public class HeapLRUStatisticsImpl implements HeapLRUEvictionStats, GFSStatsImplementor {
   private StatisticsType statType;
@@ -35,14 +33,13 @@ public class HeapLRUStatisticsImpl implements HeapLRUEvictionStats, GFSStatsImpl
     final String entryBytesDesc =
         "The amount of memory currently used by regions configured for eviction.";
     final String lruEvictionsDesc = "Number of total entry evictions triggered by LRU.";
-    final String
-        lruDestroysDesc =
+    final String lruDestroysDesc =
         "Number of entries destroyed in the region through both destroy cache operations and eviction.";
     final String lruEvaluationsDesc = "Number of entries evaluated during LRU operations.";
     final String lruGreedyReturnsDesc = "Number of non-LRU entries evicted during LRU operations";
 
     statType = factory.createType("HeapLRUStatistics", "Statistics related to heap based eviction",
-        new StatisticDescriptor[]{factory.createLongGauge("entryBytes", entryBytesDesc, "bytes"),
+        new StatisticDescriptor[] {factory.createLongGauge("entryBytes", entryBytesDesc, "bytes"),
             factory.createLongCounter("lruEvictions", lruEvictionsDesc, "entries"),
             factory.createLongCounter("lruDestroys", lruDestroysDesc, "entries"),
             factory.createLongCounter("lruEvaluations", lruEvaluationsDesc, "entries"),
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/eviction/MemoryLRUStatisticsImpl.java b/geode-core/src/main/java/org/apache/geode/internal/cache/eviction/MemoryLRUStatisticsImpl.java
index 98798b0..464a484 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/eviction/MemoryLRUStatisticsImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/eviction/MemoryLRUStatisticsImpl.java
@@ -19,8 +19,6 @@ import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.statistics.StatisticsTypeFactory;
-import org.apache.geode.internal.statistics.StatisticsTypeFactoryImpl;
 
 public class MemoryLRUStatisticsImpl implements MemoryLRUEvictionStats, GFSStatsImplementor {
   private StatisticsType statType;
@@ -36,14 +34,13 @@ public class MemoryLRUStatisticsImpl implements MemoryLRUEvictionStats, GFSStats
     final String bytesAllowedDesc = "Number of total bytes allowed in this region.";
     final String byteCountDesc = "Number of bytes in region.";
     final String lruEvictionsDesc = "Number of total entry evictions triggered by LRU.";
-    final String
-        lruDestroysDesc =
+    final String lruDestroysDesc =
         "Number of entries destroyed in the region through both destroy cache operations and eviction.";
     final String lruEvaluationsDesc = "Number of entries evaluated during LRU operations.";
     final String lruGreedyReturnsDesc = "Number of non-LRU entries evicted during LRU operations";
 
     statType = factory.createType("MemLRUStatistics", "Statistics relates to memory based eviction",
-        new StatisticDescriptor[]{
+        new StatisticDescriptor[] {
             factory.createLongGauge("bytesAllowed", bytesAllowedDesc, "bytes"),
             factory.createLongGauge("byteCount", byteCountDesc, "bytes"),
             factory.createLongCounter("lruEvictions", lruEvictionsDesc, "entries"),
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/execute/FunctionServiceStats.java b/geode-core/src/main/java/org/apache/geode/internal/cache/execute/FunctionServiceStats.java
index ab36767..8912a76 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/execute/FunctionServiceStats.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/execute/FunctionServiceStats.java
@@ -18,10 +18,6 @@ import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.statistics.StatisticsTypeFactory;
-import org.apache.geode.distributed.internal.DistributionStats;
-import org.apache.geode.internal.statistics.DummyStatisticsImpl;
-import org.apache.geode.internal.statistics.StatisticsTypeFactoryImpl;
 
 public class FunctionServiceStats {
 
@@ -126,7 +122,8 @@ public class FunctionServiceStats {
             factory.createLongCounter(FUNCTION_EXECUTIONS_COMPLETED_PROCESSING_TIME,
                 "Total time consumed for all completed invocations", "nanoseconds"),
 
-            factory.createIntGauge(FUNCTION_EXECUTIONS_RUNNING, "number of currently running invocations",
+            factory.createIntGauge(FUNCTION_EXECUTIONS_RUNNING,
+                "number of currently running invocations",
                 "operations"),
 
             factory.createIntCounter(RESULTS_SENT_TO_RESULTCOLLECTOR,
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/execute/FunctionStatsImpl.java b/geode-core/src/main/java/org/apache/geode/internal/cache/execute/FunctionStatsImpl.java
index 215b25c..64685db 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/execute/FunctionStatsImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/execute/FunctionStatsImpl.java
@@ -14,14 +14,13 @@
  */
 package org.apache.geode.internal.cache.execute;
 
+import org.apache.geode.distributed.internal.DistributionConfig;
+import org.apache.geode.distributed.internal.InternalDistributedSystem;
 import org.apache.geode.statistics.GFSStatsImplementor;
 import org.apache.geode.statistics.StatisticDescriptor;
 import org.apache.geode.statistics.Statistics;
 import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatisticsType;
-import org.apache.geode.distributed.internal.DistributionConfig;
-import org.apache.geode.distributed.internal.DistributionStats;
-import org.apache.geode.distributed.internal.InternalDistributedSystem;
 
 public class FunctionStatsImpl implements FunctionStats, GFSStatsImplementor {
 
@@ -147,12 +146,13 @@ public class FunctionStatsImpl implements FunctionStats, GFSStatsImplementor {
 
   /**
    * Returns the Function Stats for the given function
+   *
    * @param functionID represents the function for which we are returning the function Stats
    * @param distributedSystem represents the Distributed System
    * @return object of the FunctionStats
    */
   public static FunctionStats getFunctionStats(String functionID,
-                                                   InternalDistributedSystem distributedSystem) {
+      InternalDistributedSystem distributedSystem) {
     return distributedSystem.getInternalDistributedSystemStats().getFunctionStats(functionID);
   }
 
@@ -180,7 +180,7 @@ public class FunctionStatsImpl implements FunctionStats, GFSStatsImplementor {
     String statDescription = "This is the stats for the individual Function's Execution";
 
     _type = factory.createType(statName, statDescription,
-        new StatisticDescriptor[]{factory.createIntCounter(FUNCTION_EXECUTIONS_COMPLETED,
+        new StatisticDescriptor[] {factory.createIntCounter(FUNCTION_EXECUTIONS_COMPLETED,
             "Total number of completed function.execute() calls for given function", "operations"),
 
             factory.createLongCounter(FUNCTION_EXECUTIONS_COMPLETED_PROCESSING_TIME,
@@ -237,8 +237,9 @@ public class FunctionStatsImpl implements FunctionStats, GFSStatsImplementor {
 
   /**
    * Constructor.
+   *
    * @param factory The <code>StatisticsFactory</code> which creates the <code>Statistics</code>
-   * instance
+   *        instance
    * @param name The name of the <code>Statistics</code>
    */
   public FunctionStatsImpl(StatisticsFactory factory, String name) {
@@ -259,6 +260,7 @@ public class FunctionStatsImpl implements FunctionStats, GFSStatsImplementor {
 
   /**
    * Returns the current value of the "Total number of completed function.execute() calls" stat.
+   *
    * @return the current value of the "function Executions completed" stat
    */
   @Override
@@ -277,6 +279,7 @@ public class FunctionStatsImpl implements FunctionStats, GFSStatsImplementor {
 
   /**
    * Returns the current value of the "Total time consumed for all completed invocations" stat.
+   *
    * @return the current value of the "functionExecutionCompleteProcessingTime" stat
    */
   @Override
@@ -286,6 +289,7 @@ public class FunctionStatsImpl implements FunctionStats, GFSStatsImplementor {
 
   /**
    * Returns the current value of the "number of currently running invocations" stat.
+   *
    * @return the current value of the "functionExecutionsRunning" stat
    */
   @Override
@@ -304,6 +308,7 @@ public class FunctionStatsImpl implements FunctionStats, GFSStatsImplementor {
 
   /**
    * Returns the current value of the "Total number of results sent to the ResultCollector" stat.
+   *
    * @return the current value of the "resultsReturned" stat
    */
   @Override
@@ -323,6 +328,7 @@ public class FunctionStatsImpl implements FunctionStats, GFSStatsImplementor {
   /**
    * Returns the current value of the "Total number of results received and passed to
    * ResultCollector" stat.
+   *
    * @return the current value of the "resultsReturned" stat
    */
   @Override
@@ -341,6 +347,7 @@ public class FunctionStatsImpl implements FunctionStats, GFSStatsImplementor {
 
   /**
    * Returns the current value of the "Total number of FunctionService...execute() calls" stat.
+   *
    * @return the current value of the "functionExecutionsCall" stat
    */
   @Override
@@ -360,6 +367,7 @@ public class FunctionStatsImpl implements FunctionStats, GFSStatsImplementor {
   /**
    * Returns the current value of the "Total time consumed for all completed execute() calls where
    * hasResult() returns true" stat.
+   *
    * @return the current value of the "functionExecutionHasResultCompleteProcessingTime" stat
    */
   @Override
@@ -370,6 +378,7 @@ public class FunctionStatsImpl implements FunctionStats, GFSStatsImplementor {
   /**
    * Returns the current value of the "A gauge indicating the number of currently active execute()
    * calls for functions where hasResult() returns true" stat.
+   *
    * @return the current value of the "functionExecutionHasResultRunning" stat
    */
   @Override
@@ -389,6 +398,7 @@ public class FunctionStatsImpl implements FunctionStats, GFSStatsImplementor {
   /**
    * Returns the current value of the "Total number of Exceptions Occurred while executing function"
    * stat.
+   *
    * @return the current value of the "functionExecutionHasResultRunning" stat
    */
   @Override
@@ -407,6 +417,7 @@ public class FunctionStatsImpl implements FunctionStats, GFSStatsImplementor {
 
   /**
    * Returns the current time (ns).
+   *
    * @return the current time (ns)
    */
   @Override
@@ -436,10 +447,11 @@ public class FunctionStatsImpl implements FunctionStats, GFSStatsImplementor {
   /**
    * Increments the "functionExecutionsCompleted" and "functionExecutionCompleteProcessingTime"
    * stats.
+   *
    * @param start The start of the functionExecution (which is decremented from the current time to
-   * determine the function Execution processing time).
+   *        determine the function Execution processing time).
    * @param haveResult haveResult=true then update the _functionExecutionHasResultRunningId and
-   * _functionExecutionHasResultCompleteProcessingTimeId
+   *        _functionExecutionHasResultCompleteProcessingTimeId
    */
   @Override
   public void endFunctionExecution(long start, boolean haveResult) {
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/execute/MemberFunctionResultSender.java b/geode-core/src/main/java/org/apache/geode/internal/cache/execute/MemberFunctionResultSender.java
index fa37363..45c958d 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/execute/MemberFunctionResultSender.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/execute/MemberFunctionResultSender.java
@@ -118,7 +118,8 @@ public class MemberFunctionResultSender implements InternalResultSender {
           // call a synchronized method as local node is also waiting to send lastResult
           lastResult(oneResult, rc, false, true, this.dm.getDistributionManagerId());
         }
-        FunctionStatsImpl.getFunctionStats(function.getId(), this.dm.getSystem()).incResultsReceived();
+        FunctionStatsImpl.getFunctionStats(function.getId(), this.dm.getSystem())
+            .incResultsReceived();
       }
     }
     FunctionStatsImpl.getFunctionStats(function.getId(), this.dm.getSystem()).incResultsReturned();
@@ -176,7 +177,8 @@ public class MemberFunctionResultSender implements InternalResultSender {
       } else {
         reply.addResult(memberID, oneResult);
       }
-      FunctionStatsImpl.getFunctionStats(function.getId(), this.dm.getSystem()).incResultsReceived();
+      FunctionStatsImpl.getFunctionStats(function.getId(), this.dm.getSystem())
+          .incResultsReceived();
     }
     FunctionStatsImpl.getFunctionStats(function.getId(), this.dm.getSystem()).incResultsReturned();
   }
@@ -206,10 +208,12 @@ public class MemberFunctionResultSender implements InternalResultSender {
         }
       } else {
         this.rc.addResult(this.dm.getDistributionManagerId(), oneResult);
-        FunctionStatsImpl.getFunctionStats(function.getId(), this.dm.getSystem()).incResultsReceived();
+        FunctionStatsImpl.getFunctionStats(function.getId(), this.dm.getSystem())
+            .incResultsReceived();
       }
       // incrementing result sent stats.
-      FunctionStatsImpl.getFunctionStats(function.getId(), this.dm.getSystem()).incResultsReturned();
+      FunctionStatsImpl.getFunctionStats(function.getId(), this.dm.getSystem())
+          .incResultsReturned();
     }
   }
 
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/execute/PartitionedRegionFunctionResultSender.java b/geode-core/src/main/java/org/apache/geode/internal/cache/execute/PartitionedRegionFunctionResultSender.java
index 8bd1d63..5954847 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/execute/PartitionedRegionFunctionResultSender.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/execute/PartitionedRegionFunctionResultSender.java
@@ -202,7 +202,8 @@ public class PartitionedRegionFunctionResultSender implements InternalResultSend
           // call a synchronized method as local node is also waiting to send lastResult
           lastResult(oneResult, rc, false, true, dm.getDistributionManagerId());
         }
-        FunctionStatsImpl.getFunctionStats(function.getId(), this.dm.getSystem()).incResultsReceived();
+        FunctionStatsImpl.getFunctionStats(function.getId(), this.dm.getSystem())
+            .incResultsReceived();
       }
       // incrementing result sent stats.
       // Bug : remote node as well as local node calls this method to send
@@ -210,7 +211,8 @@ public class PartitionedRegionFunctionResultSender implements InternalResultSend
       // time the stats for the result sent is again incremented : Once the PR team comes with the
       // concept of the Streaming FunctionOperation
       // for the partitioned Region then it will be simple to fix this problem.
-      FunctionStatsImpl.getFunctionStats(function.getId(), this.dm.getSystem()).incResultsReturned();
+      FunctionStatsImpl.getFunctionStats(function.getId(), this.dm.getSystem())
+          .incResultsReturned();
     }
   }
 
@@ -318,13 +320,15 @@ public class PartitionedRegionFunctionResultSender implements InternalResultSend
       if (this.dm == null) {
         FunctionStatsImpl.getFunctionStats(function.getId()).incResultsReceived();
       } else {
-        FunctionStatsImpl.getFunctionStats(function.getId(), this.dm.getSystem()).incResultsReceived();
+        FunctionStatsImpl.getFunctionStats(function.getId(), this.dm.getSystem())
+            .incResultsReceived();
       }
     }
     if (this.dm == null) {
       FunctionStatsImpl.getFunctionStats(function.getId()).incResultsReturned();
     } else {
-      FunctionStatsImpl.getFunctionStats(function.getId(), this.dm.getSystem()).incResultsReturned();
+      FunctionStatsImpl.getFunctionStats(function.getId(), this.dm.getSystem())
+          .incResultsReturned();
     }
   }
 
@@ -356,10 +360,12 @@ public class PartitionedRegionFunctionResultSender implements InternalResultSend
             "PartitionedRegionFunctionResultSender adding result to ResultCollector on local node {}",
             oneResult);
         this.rc.addResult(dm.getDistributionManagerId(), oneResult);
-        FunctionStatsImpl.getFunctionStats(function.getId(), this.dm.getSystem()).incResultsReceived();
+        FunctionStatsImpl.getFunctionStats(function.getId(), this.dm.getSystem())
+            .incResultsReceived();
       }
       // incrementing result sent stats.
-      FunctionStatsImpl.getFunctionStats(function.getId(), this.dm.getSystem()).incResultsReturned();
+      FunctionStatsImpl.getFunctionStats(function.getId(), this.dm.getSystem())
+          .incResultsReturned();
     }
   }
 
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/execute/ServerRegionFunctionExecutor.java b/geode-core/src/main/java/org/apache/geode/internal/cache/execute/ServerRegionFunctionExecutor.java
index a60314d..77ee0be 100755
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/execute/ServerRegionFunctionExecutor.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/execute/ServerRegionFunctionExecutor.java
@@ -199,7 +199,8 @@ public class ServerRegionFunctionExecutor extends AbstractExecution {
   private ResultCollector executeOnServer(Function function, ResultCollector collector,
       byte hasResult) throws FunctionException {
     ServerRegionProxy srp = getServerRegionProxy();
-    FunctionStats stats = FunctionStatsImpl.getFunctionStats(function.getId(), this.region.getSystem());
+    FunctionStats stats =
+        FunctionStatsImpl.getFunctionStats(function.getId(), this.region.getSystem());
     try {
       validateExecution(function, null);
       long start = stats.startTime();
@@ -241,7 +242,8 @@ public class ServerRegionFunctionExecutor extends AbstractExecution {
 
   private void executeOnServerNoAck(Function function, byte hasResult) throws FunctionException {
     ServerRegionProxy srp = getServerRegionProxy();
-    FunctionStats stats = FunctionStatsImpl.getFunctionStats(function.getId(), this.region.getSystem());
+    FunctionStats stats =
+        FunctionStatsImpl.getFunctionStats(function.getId(), this.region.getSystem());
     try {
       validateExecution(function, null);
       long start = stats.startTime();
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/ha/HARegionQueue.java b/geode-core/src/main/java/org/apache/geode/internal/cache/ha/HARegionQueue.java
index 9222e0d..0bfee21 100755
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/ha/HARegionQueue.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/ha/HARegionQueue.java
@@ -50,7 +50,6 @@ import org.apache.geode.CancelCriterion;
 import org.apache.geode.CancelException;
 import org.apache.geode.InternalGemFireError;
 import org.apache.geode.InternalGemFireException;
-import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.SystemFailure;
 import org.apache.geode.cache.AttributesFactory;
 import org.apache.geode.cache.CacheException;
@@ -103,6 +102,7 @@ import org.apache.geode.internal.util.concurrent.StoppableCondition;
 import org.apache.geode.internal.util.concurrent.StoppableReentrantLock;
 import org.apache.geode.internal.util.concurrent.StoppableReentrantReadWriteLock;
 import org.apache.geode.internal.util.concurrent.StoppableReentrantReadWriteLock.StoppableWriteLock;
+import org.apache.geode.statistics.StatisticsFactory;
 import org.apache.geode.statistics.StatsFactory;
 
 /**
@@ -374,7 +374,8 @@ public class HARegionQueue implements RegionQueue {
     createHARegion(processedRegionName, cache);
 
     initializeHARegionQueue(processedRegionName, this.region, haContainer, clientProxyId,
-        clientConflation, isPrimary, StatsFactory.createHARegionQueueStatsImpl(factory, processedRegionName),
+        clientConflation, isPrimary,
+        StatsFactory.createHARegionQueueStatsImpl(factory, processedRegionName),
         new StoppableReentrantReadWriteLock(cache.getCancelCriterion()),
         new StoppableReentrantReadWriteLock(region.getCancelCriterion()),
         this.region.getCancelCriterion(), true);
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/ha/HARegionQueueStatsImpl.java b/geode-core/src/main/java/org/apache/geode/internal/cache/ha/HARegionQueueStatsImpl.java
index 25cfa92..78122e0 100755
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/ha/HARegionQueueStatsImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/ha/HARegionQueueStatsImpl.java
@@ -106,7 +106,8 @@ public class HARegionQueueStatsImpl implements HARegionQueueStats, GFSStatsImple
   public void initializeStats(StatisticsFactory factory) {
     String statName = "ClientSubscriptionStats";
     _type = factory.createType(statName, statName, new StatisticDescriptor[] {
-        factory.createLongCounter(HARegionQueueStats.EVENTS_QUEUED, "Number of events added to queue.", "operations"),
+        factory.createLongCounter(HARegionQueueStats.EVENTS_QUEUED,
+            "Number of events added to queue.", "operations"),
 
         factory.createLongCounter(EVENTS_CONFLATED, "Number of events conflated for the queue.",
             "operations"),
@@ -117,7 +118,8 @@ public class HARegionQueueStatsImpl implements HARegionQueueStats, GFSStatsImple
         factory.createLongCounter(EVENTS_REMOVED, "Number of events removed from the queue.",
             "operations"),
 
-        factory.createLongCounter(EVENTS_TAKEN, "Number of events taken from the queue.", "operations"),
+        factory.createLongCounter(EVENTS_TAKEN, "Number of events taken from the queue.",
+            "operations"),
 
         factory.createLongCounter(EVENTS_EXPIRED, "Number of events expired from the queue.",
             "operations"),
@@ -125,7 +127,8 @@ public class HARegionQueueStatsImpl implements HARegionQueueStats, GFSStatsImple
         factory.createLongCounter(EVENTS_REMOVED_BY_QRM, "Number of events removed by QRM message.",
             "operations"),
 
-        factory.createIntCounter(THREAD_IDENTIFIERS, "Number of ThreadIdenfier objects for the queue.",
+        factory.createIntCounter(THREAD_IDENTIFIERS,
+            "Number of ThreadIdenfier objects for the queue.",
             "units"),
 
         factory.createLongCounter(EVENTS_DISPATCHED, "Number of events that have been dispatched.",
@@ -134,7 +137,8 @@ public class HARegionQueueStatsImpl implements HARegionQueueStats, GFSStatsImple
         factory.createLongCounter(NUM_VOID_REMOVALS, "Number of void removals from the queue.",
             "operations"),
 
-        factory.createLongCounter(NUM_SEQUENCE_VIOLATED, "Number of events that has violated sequence.",
+        factory.createLongCounter(NUM_SEQUENCE_VIOLATED,
+            "Number of events that has violated sequence.",
             "operations")});
 
     // Initialize id fields
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/partitioned/BucketSizeMessage.java b/geode-core/src/main/java/org/apache/geode/internal/cache/partitioned/BucketSizeMessage.java
index aaa9fc1..4a80bfc 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/partitioned/BucketSizeMessage.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/partitioned/BucketSizeMessage.java
@@ -26,7 +26,6 @@ import org.apache.geode.cache.CacheException;
 import org.apache.geode.distributed.internal.ClusterDistributionManager;
 import org.apache.geode.distributed.internal.DistributionManager;
 import org.apache.geode.distributed.internal.DistributionMessage;
-import org.apache.geode.distributed.internal.DistributionStats;
 import org.apache.geode.distributed.internal.HighPriorityDistributionMessage;
 import org.apache.geode.distributed.internal.InternalDistributedSystem;
 import org.apache.geode.distributed.internal.ReplyException;
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/partitioned/CreateBucketMessage.java b/geode-core/src/main/java/org/apache/geode/internal/cache/partitioned/CreateBucketMessage.java
index 3019ccb..edadb7e 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/partitioned/CreateBucketMessage.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/partitioned/CreateBucketMessage.java
@@ -27,7 +27,6 @@ import org.apache.geode.cache.PartitionedRegionStorageException;
 import org.apache.geode.distributed.internal.ClusterDistributionManager;
 import org.apache.geode.distributed.internal.DistributionManager;
 import org.apache.geode.distributed.internal.DistributionMessage;
-import org.apache.geode.distributed.internal.DistributionStats;
 import org.apache.geode.distributed.internal.InternalDistributedSystem;
 import org.apache.geode.distributed.internal.ReplyException;
 import org.apache.geode.distributed.internal.ReplyMessage;
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/partitioned/CreateMissingBucketsTask.java b/geode-core/src/main/java/org/apache/geode/internal/cache/partitioned/CreateMissingBucketsTask.java
index 79e87a3..862522d 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/partitioned/CreateMissingBucketsTask.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/partitioned/CreateMissingBucketsTask.java
@@ -18,7 +18,6 @@ import org.apache.geode.internal.cache.ColocationHelper;
 import org.apache.geode.internal.cache.PRHARedundancyProvider;
 import org.apache.geode.internal.cache.PartitionedRegion;
 import org.apache.geode.internal.cache.PartitionedRegion.RecoveryLock;
-import org.apache.geode.internal.cache.PartitionedRegionStats;
 
 /**
  * A task for creating buckets in a child colocated region that are present in the leader region.
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/partitioned/GetMessage.java b/geode-core/src/main/java/org/apache/geode/internal/cache/partitioned/GetMessage.java
index 6440d76..1286cc2 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/partitioned/GetMessage.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/partitioned/GetMessage.java
@@ -63,11 +63,13 @@ import org.apache.geode.internal.offheap.OffHeapHelper;
 import org.apache.geode.internal.util.BlobHelper;
 
 /**
- * This message is used as the request for a {@link org.apache.geode.cache.Region#get(Object)}operation.
+ * This message is used as the request for a
+ * {@link org.apache.geode.cache.Region#get(Object)}operation.
  * The reply is sent in a {@link org.apache.geode.internal.cache.partitioned.GetMessage}.
  *
  * Since the {@link org.apache.geode.cache.Region#get(Object)}operation is used <bold>very </bold>
  * frequently the performance of this class is critical.
+ *
  * @since GemFire 5.0
  */
 public class GetMessage extends PartitionMessageWithDirectReply {
@@ -96,13 +98,12 @@ public class GetMessage extends PartitionMessageWithDirectReply {
   /**
    * Empty constructor to satisfy {@link DataSerializer} requirements
    */
-  public GetMessage() {
-  }
+  public GetMessage() {}
 
   private GetMessage(InternalDistributedMember recipient, int regionId,
-                     DirectReplyProcessor processor, final Object key,
-                     final Object aCallbackArgument,
-                     ClientProxyMembershipID context, boolean returnTombstones) {
+      DirectReplyProcessor processor, final Object key,
+      final Object aCallbackArgument,
+      ClientProxyMembershipID context, boolean returnTombstones) {
     super(recipient, regionId, processor);
     this.key = key;
     this.cbArg = aCallbackArgument;
@@ -158,7 +159,7 @@ public class GetMessage extends PartitionMessageWithDirectReply {
 
   @Override
   protected boolean operateOnPartitionedRegion(final ClusterDistributionManager dm,
-                                               PartitionedRegion r, long startTime)
+      PartitionedRegion r, long startTime)
       throws ForceReattemptException {
     if (logger.isTraceEnabled(LogMarker.DM_VERBOSE)) {
       logger.trace(LogMarker.DM_VERBOSE, "GetMessage operateOnRegion: {}", r.getFullPath());
@@ -274,6 +275,7 @@ public class GetMessage extends PartitionMessageWithDirectReply {
 
   /**
    * Sends a PartitionedRegion {@link org.apache.geode.cache.Region#get(Object)} message
+   *
    * @param recipient the member that the get message is sent to
    * @param r the PartitionedRegion for which get was performed upon
    * @param key the object to which the value should be feteched
@@ -282,9 +284,9 @@ public class GetMessage extends PartitionMessageWithDirectReply {
    * @throws ForceReattemptException if the peer is no longer available
    */
   public static GetResponse send(InternalDistributedMember recipient, PartitionedRegion r,
-                                 final Object key, final Object aCallbackArgument,
-                                 ClientProxyMembershipID requestingClient,
-                                 boolean returnTombstones) throws ForceReattemptException {
+      final Object key, final Object aCallbackArgument,
+      ClientProxyMembershipID requestingClient,
+      boolean returnTombstones) throws ForceReattemptException {
     Assert.assertTrue(recipient != null, "PRDistribuedGetReplyMessage NULL reply message");
     GetResponse p = new GetResponse(r.getSystem(), Collections.singleton(recipient), key);
     GetMessage m = new GetMessage(recipient, r.getPRId(), p, key, aCallbackArgument,
@@ -300,11 +302,13 @@ public class GetMessage extends PartitionMessageWithDirectReply {
   }
 
   /**
-   * This message is used for the reply to a {@link org.apache.geode.cache.Region#get(Object)}operation
+   * This message is used for the reply to a
+   * {@link org.apache.geode.cache.Region#get(Object)}operation
    * This is the reply to a {@link GetMessage}.
    *
    * Since the {@link org.apache.geode.cache.Region#get(Object)}operation is used <bold>very </bold>
    * frequently the performance of this class is critical.
+   *
    * @since GemFire 5.0
    */
   public static class GetReplyMessage extends ReplyMessage {
@@ -344,8 +348,7 @@ public class GetMessage extends PartitionMessageWithDirectReply {
     /**
      * Empty constructor to conform to DataSerializable interface
      */
-    public GetReplyMessage() {
-    }
+    public GetReplyMessage() {}
 
     private GetReplyMessage(int processorId, RawValue val, VersionTag versionTag) {
       setProcessorId(processorId);
@@ -375,6 +378,7 @@ public class GetMessage extends PartitionMessageWithDirectReply {
      * Return the value from the get operation, serialize it bytes as late as possible to avoid
      * making un-neccesary byte[] copies. De-serialize those same bytes as late as possible to avoid
      * using precious threads (aka P2P readers).
+     *
      * @param recipient the origin VM that performed the get
      * @param processorId the processor on which the origin thread is waiting
      * @param val the raw value that will eventually be serialized
@@ -382,7 +386,7 @@ public class GetMessage extends PartitionMessageWithDirectReply {
      * @param versionTag the version of the object
      */
     public static void send(InternalDistributedMember recipient, int processorId, RawValue val,
-                            ReplySender replySender, VersionTag versionTag)
+        ReplySender replySender, VersionTag versionTag)
         throws ForceReattemptException {
       Assert.assertTrue(recipient != null, "PRDistribuedGetReplyMessage NULL reply message");
       GetReplyMessage m = new GetReplyMessage(processorId, val, versionTag);
@@ -392,6 +396,7 @@ public class GetMessage extends PartitionMessageWithDirectReply {
 
     /**
      * Processes this message. This method is invoked by the receiver of the message.
+     *
      * @param dm the distribution manager that is processing the message.
      */
     @Override
@@ -489,7 +494,9 @@ public class GetMessage extends PartitionMessageWithDirectReply {
   }
 
   /**
-   * A processor to capture the value returned by {@link org.apache.geode.internal.cache.partitioned.GetMessage.GetReplyMessage}
+   * A processor to capture the value returned by
+   * {@link org.apache.geode.internal.cache.partitioned.GetMessage.GetReplyMessage}
+   *
    * @since GemFire 5.0
    */
   public static class GetResponse extends PartitionResponse {
@@ -524,6 +531,7 @@ public class GetMessage extends PartitionMessageWithDirectReply {
     /**
      * De-seralize the value, if the value isn't already a byte array, this method should be called
      * in the context of the requesting thread for the best scalability
+     *
      * @return the value object
      * @see EntryEventImpl#deserialize(byte[])
      */
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/partitioned/ManageBackupBucketMessage.java b/geode-core/src/main/java/org/apache/geode/internal/cache/partitioned/ManageBackupBucketMessage.java
index 23257a5..9e33f4b 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/partitioned/ManageBackupBucketMessage.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/partitioned/ManageBackupBucketMessage.java
@@ -26,7 +26,6 @@ import org.apache.geode.cache.CacheClosedException;
 import org.apache.geode.distributed.internal.ClusterDistributionManager;
 import org.apache.geode.distributed.internal.DistributionManager;
 import org.apache.geode.distributed.internal.DistributionMessage;
-import org.apache.geode.distributed.internal.DistributionStats;
 import org.apache.geode.distributed.internal.InternalDistributedSystem;
 import org.apache.geode.distributed.internal.ReplyException;
 import org.apache.geode.distributed.internal.ReplyMessage;
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/tier/sockets/AcceptorImpl.java b/geode-core/src/main/java/org/apache/geode/internal/cache/tier/sockets/AcceptorImpl.java
index aee3045..f03fa76 100755
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/tier/sockets/AcceptorImpl.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/tier/sockets/AcceptorImpl.java
@@ -86,7 +86,6 @@ import org.apache.geode.internal.cache.partitioned.AllBucketProfilesUpdateMessag
 import org.apache.geode.internal.cache.tier.Acceptor;
 import org.apache.geode.internal.cache.tier.CachedRegionHelper;
 import org.apache.geode.internal.cache.tier.CommunicationMode;
-import org.apache.geode.internal.cache.wan.GatewayReceiverStats;
 import org.apache.geode.internal.i18n.LocalizedStrings;
 import org.apache.geode.internal.logging.LogService;
 import org.apache.geode.internal.logging.LoggingThreadGroup;
@@ -530,7 +529,7 @@ public class AcceptorImpl implements Acceptor, Runnable, CommBufferPool {
           LocalizedStrings.AcceptorImpl_CACHE_SERVER_CONNECTION_LISTENER_BOUND_TO_ADDRESS_0_WITH_BACKLOG_1,
           new Object[] {sockName, Integer.valueOf(backLog)}));
       if (isGatewayReceiver) {
-        this.stats = (CacheServerStats) StatsFactory.createGatewayReceiverStatsImpl(null,sockName);
+        this.stats = (CacheServerStats) StatsFactory.createGatewayReceiverStatsImpl(null, sockName);
       } else {
         this.stats = new CacheServerStatsImpl(sockName);
       }
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/tier/sockets/BaseCommand.java b/geode-core/src/main/java/org/apache/geode/internal/cache/tier/sockets/BaseCommand.java
index 5731b42..0ad152e 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/tier/sockets/BaseCommand.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/tier/sockets/BaseCommand.java
@@ -48,7 +48,6 @@ import org.apache.geode.cache.persistence.PartitionOfflineException;
 import org.apache.geode.cache.query.types.CollectionType;
 import org.apache.geode.distributed.DistributedSystemDisconnectedException;
 import org.apache.geode.distributed.internal.DistributionConfig;
-import org.apache.geode.distributed.internal.DistributionStats;
 import org.apache.geode.distributed.internal.membership.InternalDistributedMember;
 import org.apache.geode.internal.Assert;
 import org.apache.geode.internal.Version;
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/tier/sockets/BaseCommandQuery.java b/geode-core/src/main/java/org/apache/geode/internal/cache/tier/sockets/BaseCommandQuery.java
index 5659fd9..c6802d3 100644
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/tier/sockets/BaseCommandQuery.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/tier/sockets/BaseCommandQuery.java
@@ -34,7 +34,6 @@ import org.apache.geode.cache.query.internal.types.CollectionTypeImpl;
 import org.apache.geode.cache.query.internal.types.StructTypeImpl;
 import org.apache.geode.cache.query.types.CollectionType;
 import org.apache.geode.distributed.DistributedSystemDisconnectedException;
-import org.apache.geode.distributed.internal.DistributionStats;
 import org.apache.geode.internal.Version;
 import org.apache.geode.internal.cache.CachedDeserializable;
 import org.apache.geode.internal.cache.tier.CachedRegionHelper;
diff --git a/geode-core/src/main/java/org/apache/geode/internal/cache/tier/sockets/CacheClientNotifier.java b/geode-core/src/main/java/org/apache/geode/internal/cache/tier/sockets/CacheClientNotifier.java
index 905a211..bcd4abf 100755
--- a/geode-core/src/main/java/org/apache/geode/internal/cache/tier/sockets/CacheClientNotifier.java
+++ b/geode-core/src/main/java/org/apache/geode/internal/cache/tier/sockets/CacheClientNotifier.java
@@ -116,6 +116,7 @@ import org.apache.geode.statistics.StatisticsFactory;
 /**
  * Class <code>CacheClientNotifier</code> works on the server and manages client socket connections
  * to clients requesting notification of updates and notifies them when updates occur.
+ *
  * @since GemFire 3.2
  */
 @SuppressWarnings({"synthetic-access", "deprecation"})
@@ -126,16 +127,17 @@ public class CacheClientNotifier {
 
   /**
    * Factory method to construct a CacheClientNotifier <code>CacheClientNotifier</code> instance.
+   *
    * @param cache The GemFire <code>InternalCache</code>
    * @return A <code>CacheClientNotifier</code> instance
    */
   public static synchronized CacheClientNotifier getInstance(InternalCache cache,
-                                                             CacheServerStats acceptorStats,
-                                                             int maximumMessageCount,
-                                                             int messageTimeToLive,
-                                                             ConnectionListener listener,
-                                                             List overflowAttributesList,
-                                                             boolean isGatewayReceiver) {
+      CacheServerStats acceptorStats,
+      int maximumMessageCount,
+      int messageTimeToLive,
+      ConnectionListener listener,
+      List overflowAttributesList,
+      boolean isGatewayReceiver) {
     if (ccnSingleton == null) {
       ccnSingleton = new CacheClientNotifier(cache, acceptorStats, maximumMessageCount,
           messageTimeToLive, listener, overflowAttributesList, isGatewayReceiver);
@@ -155,6 +157,7 @@ public class CacheClientNotifier {
 
   /**
    * Writes a given message to the output stream
+   *
    * @param dos the <code>DataOutputStream</code> to use for writing the message
    * @param type a byte representing the message type
    * @param p_msg the message to be written; can be null
@@ -165,7 +168,7 @@ public class CacheClientNotifier {
   }
 
   private void writeHandshakeMessage(DataOutputStream dos, byte type, String p_msg,
-                                     Version clientVersion, byte endpointType, int queueSize)
+      Version clientVersion, byte endpointType, int queueSize)
       throws IOException {
     String msg = p_msg;
 
@@ -224,6 +227,7 @@ public class CacheClientNotifier {
 
   /**
    * Writes an exception message to the socket
+   *
    * @param dos the <code>DataOutputStream</code> to use for writing the message
    * @param type a byte representing the exception type
    * @param ex the exception to be written; should not be null
@@ -235,10 +239,11 @@ public class CacheClientNotifier {
 
   /**
    * Registers a new client updater that wants to receive updates with this server.
+   *
    * @param socket The socket over which the server communicates with the client.
    */
   public void registerClient(Socket socket, boolean isPrimary, long acceptorId,
-                             boolean notifyBySubscription) throws IOException {
+      boolean notifyBySubscription) throws IOException {
     // Since no remote ports were specified in the message, wait for them.
     long startTime = this.statistics.startTime();
     DataInputStream dis = new DataInputStream(socket.getInputStream());
@@ -286,9 +291,9 @@ public class CacheClientNotifier {
   }
 
   protected void registerGFEClient(DataInputStream dis, DataOutputStream dos, Socket socket,
-                                   boolean isPrimary, long startTime, Version clientVersion,
-                                   long acceptorId,
-                                   boolean notifyBySubscription) throws IOException {
+      boolean isPrimary, long startTime, Version clientVersion,
+      long acceptorId,
+      boolean notifyBySubscription) throws IOException {
     // Read the ports and throw them away. We no longer need them
     int numberOfPorts = dis.readInt();
     for (int i = 0; i < numberOfPorts; i++) {
@@ -314,7 +319,7 @@ public class CacheClientNotifier {
       String authenticator = sysProps.getProperty(SECURITY_CLIENT_AUTHENTICATOR);
 
       if (clientVersion.compareTo(Version.GFE_603) >= 0) {
-        byte[] overrides = Handshake.extractOverrides(new byte[]{(byte) dis.read()});
+        byte[] overrides = Handshake.extractOverrides(new byte[] {(byte) dis.read()});
         clientConflation = overrides[0];
       } else {
         clientConflation = (byte) dis.read();
@@ -362,7 +367,7 @@ public class CacheClientNotifier {
             if (principal == null) {
               securityLogWriter.warning(
                   LocalizedStrings.CacheClientNotifier_CACHECLIENTNOTIFIER_POST_PROCESS_AUTHORIZATION_CALLBACK_ENABLED_BUT_AUTHENTICATION_CALLBACK_0_RETURNED_WITH_NULL_CREDENTIALS_FOR_PROXYID_1,
-                  new Object[]{SECURITY_CLIENT_AUTHENTICATOR, proxyID});
+                  new Object[] {SECURITY_CLIENT_AUTHENTICATOR, proxyID});
             }
             Method authzMethod = ClassLoadUtil.methodFromName(postAuthzFactoryName);
             authzCallback = (AccessControl) authzMethod.invoke(null, (Object[]) null);
@@ -378,19 +383,19 @@ public class CacheClientNotifier {
     } catch (AuthenticationRequiredException ex) {
       securityLogWriter.warning(
           LocalizedStrings.CacheClientNotifier_AN_EXCEPTION_WAS_THROWN_FOR_CLIENT_0_1,
-          new Object[]{proxyID, ex});
+          new Object[] {proxyID, ex});
       writeException(dos, Handshake.REPLY_EXCEPTION_AUTHENTICATION_REQUIRED, ex, clientVersion);
       return;
     } catch (AuthenticationFailedException ex) {
       securityLogWriter.warning(
           LocalizedStrings.CacheClientNotifier_AN_EXCEPTION_WAS_THROWN_FOR_CLIENT_0_1,
-          new Object[]{proxyID, ex});
+          new Object[] {proxyID, ex});
       writeException(dos, Handshake.REPLY_EXCEPTION_AUTHENTICATION_FAILED, ex, clientVersion);
       return;
     } catch (CacheException e) {
       logger.warn(LocalizedMessage.create(
           LocalizedStrings.CacheClientNotifier_0_REGISTERCLIENT_EXCEPTION_ENCOUNTERED_IN_REGISTRATION_1,
-          new Object[]{this, e}), e);
+          new Object[] {this, e}), e);
       IOException io = new IOException(
           LocalizedStrings.CacheClientNotifier_EXCEPTION_OCCURRED_WHILE_TRYING_TO_REGISTER_INTEREST_DUE_TO_0
               .toLocalizedString(e.getMessage()));
@@ -399,7 +404,7 @@ public class CacheClientNotifier {
     } catch (Exception ex) {
       logger.warn(LocalizedMessage.create(
           LocalizedStrings.CacheClientNotifier_AN_EXCEPTION_WAS_THROWN_FOR_CLIENT_0_1,
-          new Object[]{proxyID, ""}), ex);
+          new Object[] {proxyID, ""}), ex);
       writeException(dos, CommunicationMode.UnsuccessfulServerToClient.getModeNumber(), ex,
           clientVersion);
       return;
@@ -410,16 +415,17 @@ public class CacheClientNotifier {
 
   /**
    * Registers a new client that wants to receive updates with this server.
+   *
    * @param socket The socket over which the server communicates with the client.
    * @param proxyId The distributed member id of the client being registered
    * @param proxy The <code>CacheClientProxy</code> of the given <code>proxyId</code>
    * @return CacheClientProxy for the registered client
    */
   private CacheClientProxy registerClient(Socket socket, ClientProxyMembershipID proxyId,
-                                          CacheClientProxy proxy, boolean isPrimary,
-                                          byte clientConflation, Version clientVersion,
-                                          long acceptorId, boolean notifyBySubscription,
-                                          Subject subject)
+      CacheClientProxy proxy, boolean isPrimary,
+      byte clientConflation, Version clientVersion,
+      long acceptorId, boolean notifyBySubscription,
+      Subject subject)
       throws IOException, CacheException {
     CacheClientProxy l_proxy = proxy;
 
@@ -513,7 +519,7 @@ public class CacheClientNotifier {
           // client is already using this durable id.
           unsuccessfulMsg =
               LocalizedStrings.CacheClientNotifier_CACHECLIENTNOTIFIER_THE_REQUESTED_DURABLE_CLIENT_HAS_THE_SAME_IDENTIFIER__0__AS_AN_EXISTING_DURABLE_CLIENT__1__DUPLICATE_DURABLE_CLIENTS_ARE_NOT_ALLOWED
-                  .toLocalizedString(new Object[]{proxyId.getDurableId(), proxy});
+                  .toLocalizedString(new Object[] {proxyId.getDurableId(), proxy});
           logger.warn(unsuccessfulMsg);
           // Set the unsuccessful response byte.
           responseByte = Handshake.REPLY_EXCEPTION_DUPLICATE_DURABLE_CLIENT;
@@ -555,7 +561,7 @@ public class CacheClientNotifier {
       responseByte = Handshake.REPLY_REFUSED;
       unsuccessfulMsg =
           LocalizedStrings.CacheClientNotifier_CACHECLIENTNOTIFIER_A_PREVIOUS_CONNECTION_ATTEMPT_FROM_THIS_CLIENT_IS_STILL_BEING_PROCESSED__0
-              .toLocalizedString(new Object[]{proxyId});
+              .toLocalizedString(new Object[] {proxyId});
       logger.warn(unsuccessfulMsg);
     }
 
@@ -605,7 +611,7 @@ public class CacheClientNotifier {
     } else {
       logger.warn(LocalizedMessage.create(
           LocalizedStrings.CacheClientNotifier_CACHECLIENTNOTIFIER_UNSUCCESSFULLY_REGISTERED_CLIENT_WITH_IDENTIFIER__0,
-          new Object[]{proxyId, responseByte}));
+          new Object[] {proxyId, responseByte}));
     }
     return l_proxy;
   }
@@ -641,8 +647,9 @@ public class CacheClientNotifier {
 
   /**
    * Makes Primary to this CacheClientProxy and start the dispatcher of the CacheClientProxy
+   *
    * @param isClientReady Whether the marker has already been processed. This value helps determine
-   * whether to start the dispatcher.
+   *        whether to start the dispatcher.
    */
   public void makePrimary(ClientProxyMembershipID proxyId, boolean isClientReady) {
     CacheClientProxy proxy = getClientProxy(proxyId);
@@ -686,6 +693,7 @@ public class CacheClientNotifier {
 
   /**
    * Sets keepalive on the proxy of the given membershipID
+   *
    * @param membershipID Uniquely identifies the client pool
    * @since GemFire 5.7
    */
@@ -706,6 +714,7 @@ public class CacheClientNotifier {
 
   /**
    * Unregisters an existing client from this server.
+   *
    * @param memberId Uniquely identifies the client
    */
   public void unregisterClient(ClientProxyMembershipID memberId, boolean normalShutdown) {
@@ -753,7 +762,7 @@ public class CacheClientNotifier {
     } catch (Exception e) {
       logger.fatal(LocalizedMessage.create(
           LocalizedStrings.CacheClientNotifier_CANNOT_NOTIFY_CLIENTS_TO_PERFORM_OPERATION_0_ON_EVENT_1,
-          new Object[]{operation, event}), e);
+          new Object[] {operation, event}), e);
     }
     return clientMessage;
   }
@@ -796,7 +805,7 @@ public class CacheClientNotifier {
 
     if ((filterInfo == null
         || (filterInfo.getCQs() == null && filterInfo.getInterestedClients() == null
-        && filterInfo.getInterestedClientsInv() == null))) {
+            && filterInfo.getInterestedClientsInv() == null))) {
       return;
     }
 
@@ -916,7 +925,7 @@ public class CacheClientNotifier {
   }
 
   private void removeDestroyTokensFromCqResultKeys(InternalCacheEvent event,
-                                                   FilterInfo filterInfo) {
+      FilterInfo filterInfo) {
     FilterProfile regionProfile = ((LocalRegion) event.getRegion()).getFilterProfile();
     if (event.getOperation().isEntry() && filterInfo.getCQs() != null) {
       EntryEventImpl entryEvent = (EntryEventImpl) event;
@@ -962,7 +971,7 @@ public class CacheClientNotifier {
    * this is for server side registration of client queue
    */
   public static void routeSingleClientMessage(ClientUpdateMessage clientMessage,
-                                              ClientProxyMembershipID clientProxyMembershipId) {
+      ClientProxyMembershipID clientProxyMembershipId) {
     CacheClientNotifier instance = ccnSingleton;
     if (instance != null) {
       instance.singletonRouteClientMessage(clientMessage,
@@ -971,7 +980,7 @@ public class CacheClientNotifier {
   }
 
   private void singletonRouteClientMessage(Conflatable conflatable,
-                                           Collection<ClientProxyMembershipID> filterClients) {
+      Collection<ClientProxyMembershipID> filterClients) {
 
     this.cache.getCancelCriterion().checkCancelInProgress(null); // bug #43942 - client notified
     // but no p2p distribution
@@ -1018,7 +1027,7 @@ public class CacheClientNotifier {
    * for proxies in initialization as well as fully initialized proxies.
    */
   public Set<ClientProxyMembershipID> getProxyIDs(Set mixedDurableAndNonDurableIDs,
-                                                  boolean proxyInInitMode) {
+      boolean proxyInInitMode) {
     Set<ClientProxyMembershipID> result = new HashSet();
     for (Object id : mixedDurableAndNonDurableIDs) {
       if (id instanceof String) {
@@ -1049,7 +1058,7 @@ public class CacheClientNotifier {
       // log alert with client info.
       logger.warn(
           LocalizedMessage.create(LocalizedStrings.CacheClientNotifier_CLIENT_0_IS_A_SLOW_RECEIVER,
-              new Object[]{proxy.getProxyID()}));
+              new Object[] {proxy.getProxyID()}));
       addToBlacklistedClient(proxy.getProxyID());
       InternalDistributedSystem ids =
           (InternalDistributedSystem) this.getCache().getDistributedSystem();
@@ -1086,6 +1095,7 @@ public class CacheClientNotifier {
 
   /**
    * Initializes a <code>ClientUpdateMessage</code> from an operation and event
+   *
    * @param operation The operation that occurred (e.g. AFTER_CREATE)
    * @param event The event containing the data to be updated
    * @return a <code>ClientUpdateMessage</code>
@@ -1147,6 +1157,7 @@ public class CacheClientNotifier {
 
   /**
    * Returns whether the <code>CacheClientNotifier</code> supports the input operation.
+   *
    * @param operation The operation that occurred (e.g. AFTER_CREATE)
    * @return whether the <code>CacheClientNotifier</code> supports the input operation
    */
@@ -1162,6 +1173,7 @@ public class CacheClientNotifier {
 
   /**
    * Registers client interest in the input region and key.
+   *
    * @param regionName The name of the region of interest
    * @param keyOfInterest The name of the key of interest
    * @param membershipID clients ID
@@ -1172,11 +1184,11 @@ public class CacheClientNotifier {
    * @param regionDataPolicy (0=empty)
    */
   public void registerClientInterest(String regionName, Object keyOfInterest,
-                                     ClientProxyMembershipID membershipID, int interestType,
-                                     boolean isDurable,
-                                     boolean sendUpdatesAsInvalidates, boolean manageEmptyRegions,
-                                     int regionDataPolicy,
-                                     boolean flushState)
+      ClientProxyMembershipID membershipID, int interestType,
+      boolean isDurable,
+      boolean sendUpdatesAsInvalidates, boolean manageEmptyRegions,
+      int regionDataPolicy,
+      boolean flushState)
       throws IOException, RegionDestroyedException {
 
     CacheClientProxy proxy = getClientProxy(membershipID, true);
@@ -1214,11 +1226,12 @@ public class CacheClientNotifier {
 
   /**
    * Store region and delta relation
+   *
    * @param regionDataPolicy (0==empty)
    * @since GemFire 6.1
    */
   public void updateMapOfEmptyRegions(Map regionsWithEmptyDataPolicy, String regionName,
-                                      int regionDataPolicy) {
+      int regionDataPolicy) {
     if (regionDataPolicy == 0) {
       if (!regionsWithEmptyDataPolicy.containsKey(regionName)) {
         regionsWithEmptyDataPolicy.put(regionName, Integer.valueOf(0));
@@ -1228,15 +1241,16 @@ public class CacheClientNotifier {
 
   /**
    * Unregisters client interest in the input region and key.
+   *
    * @param regionName The name of the region of interest
    * @param keyOfInterest The name of the key of interest
    * @param isClosing Whether the caller is closing
    * @param membershipID The <code>ClientProxyMembershipID</code> of the client no longer interested
-   * in this <code>Region</code> and key
+   *        in this <code>Region</code> and key
    */
   public void unregisterClientInterest(String regionName, Object keyOfInterest, int interestType,
-                                       boolean isClosing, ClientProxyMembershipID membershipID,
-                                       boolean keepalive) {
+      boolean isClosing, ClientProxyMembershipID membershipID,
+      boolean keepalive) {
     if (logger.isDebugEnabled()) {
       logger.debug(
           "CacheClientNotifier: Client {} unregistering interest in: {} -> {} (an instance of {})",
@@ -1251,16 +1265,17 @@ public class CacheClientNotifier {
 
   /**
    * Registers client interest in the input region and list of keys.
+   *
    * @param regionName The name of the region of interest
    * @param keysOfInterest The list of keys of interest
    * @param membershipID The <code>ClientProxyMembershipID</code> of the client no longer interested
-   * in this <code>Region</code> and key
+   *        in this <code>Region</code> and key
    */
   public void registerClientInterest(String regionName, List keysOfInterest,
-                                     ClientProxyMembershipID membershipID, boolean isDurable,
-                                     boolean sendUpdatesAsInvalidates,
-                                     boolean manageEmptyRegions, int regionDataPolicy,
-                                     boolean flushState)
+      ClientProxyMembershipID membershipID, boolean isDurable,
+      boolean sendUpdatesAsInvalidates,
+      boolean manageEmptyRegions, int regionDataPolicy,
+      boolean flushState)
       throws IOException, RegionDestroyedException {
     CacheClientProxy proxy = getClientProxy(membershipID, true);
 
@@ -1285,14 +1300,15 @@ public class CacheClientNotifier {
 
   /**
    * Unregisters client interest in the input region and list of keys.
+   *
    * @param regionName The name of the region of interest
    * @param keysOfInterest The list of keys of interest
    * @param isClosing Whether the caller is closing
    * @param membershipID The <code>ClientProxyMembershipID</code> of the client no longer interested
-   * in this <code>Region</code> and key
+   *        in this <code>Region</code> and key
    */
   public void unregisterClientInterest(String regionName, List keysOfInterest, boolean isClosing,
-                                       ClientProxyMembershipID membershipID, boolean keepalive) {
+      ClientProxyMembershipID membershipID, boolean keepalive) {
     if (logger.isDebugEnabled()) {
       logger.debug("CacheClientNotifier: Client {} unregistering interest in: {} -> {}",
           membershipID, regionName, keysOfInterest);
@@ -1306,6 +1322,7 @@ public class CacheClientNotifier {
 
   /**
    * Returns the <code>CacheClientProxy</code> associated to the membershipID *
+   *
    * @return the <code>CacheClientProxy</code> associated to the membershipID
    */
   public CacheClientProxy getClientProxy(ClientProxyMembershipID membershipID) {
@@ -1317,7 +1334,7 @@ public class CacheClientNotifier {
    * are initialized and those that are still in initialization mode.
    */
   public CacheClientProxy getClientProxy(ClientProxyMembershipID membershipID,
-                                         boolean proxyInInitMode) {
+      boolean proxyInInitMode) {
     CacheClientProxy proxy = getClientProxy(membershipID);
     if (proxyInInitMode && proxy == null) {
       proxy = (CacheClientProxy) this._initClientProxies.get(membershipID);
@@ -1327,6 +1344,7 @@ public class CacheClientNotifier {
 
   /**
    * Returns the <code>CacheClientProxy</code> associated to the durableClientId
+   *
    * @return the <code>CacheClientProxy</code> associated to the durableClientId
    */
   public CacheClientProxy getClientProxy(String durableClientId) {
@@ -1336,6 +1354,7 @@ public class CacheClientNotifier {
   /**
    * Returns the <code>CacheClientProxy</code> associated to the durableClientId. This version of
    * the method can check for initializing proxies as well as fully initialized proxies.
+   *
    * @return the <code>CacheClientProxy</code> associated to the durableClientId
    */
   public CacheClientProxy getClientProxy(String durableClientId, boolean proxyInInitMode) {
@@ -1346,7 +1365,7 @@ public class CacheClientNotifier {
       logger.debug("CacheClientNotifier: Determining client for {}", durableClientId);
     }
     CacheClientProxy proxy = null;
-    for (Iterator i = getClientProxies().iterator(); i.hasNext(); ) {
+    for (Iterator i = getClientProxies().iterator(); i.hasNext();) {
       CacheClientProxy clientProxy = (CacheClientProxy) i.next();
       if (isTraceEnabled) {
         logger.trace("CacheClientNotifier: Checking client {}", clientProxy);
@@ -1361,7 +1380,7 @@ public class CacheClientNotifier {
       }
     }
     if (proxy == null && proxyInInitMode) {
-      for (Iterator i = this._initClientProxies.values().iterator(); i.hasNext(); ) {
+      for (Iterator i = this._initClientProxies.values().iterator(); i.hasNext();) {
         CacheClientProxy clientProxy = (CacheClientProxy) i.next();
         if (isTraceEnabled) {
           logger.trace("CacheClientNotifier: Checking initializing client {}", clientProxy);
@@ -1382,6 +1401,7 @@ public class CacheClientNotifier {
 
   /**
    * Returns the <code>CacheClientProxySameDS</code> associated to the membershipID *
+   *
    * @return the <code>CacheClientProxy</code> associated to the same distributed system
    */
   public CacheClientProxy getClientProxySameDS(ClientProxyMembershipID membershipID) {
@@ -1393,7 +1413,7 @@ public class CacheClientNotifier {
           this, getClientProxies().size());
     }
     CacheClientProxy proxy = null;
-    for (Iterator i = getClientProxies().iterator(); i.hasNext(); ) {
+    for (Iterator i = getClientProxies().iterator(); i.hasNext();) {
       CacheClientProxy clientProxy = (CacheClientProxy) i.next();
       if (isDebugEnabled) {
         logger.debug("CacheClientNotifier: Checking client {}", clientProxy);
@@ -1473,6 +1493,7 @@ public class CacheClientNotifier {
 
   /**
    * Adds a new <code>CacheClientProxy</code> to the list of known client proxies
+   *
    * @param proxy The <code>CacheClientProxy</code> to add
    */
   protected void addClientProxy(CacheClientProxy proxy) throws IOException {
@@ -1511,11 +1532,12 @@ public class CacheClientNotifier {
   /**
    * Returns (possibly stale) set of memberIds for all clients being actively notified by this
    * server.
+   *
    * @return set of memberIds
    */
   public Set getActiveClients() {
     Set clients = new HashSet();
-    for (Iterator iter = getClientProxies().iterator(); iter.hasNext(); ) {
+    for (Iterator iter = getClientProxies().iterator(); iter.hasNext();) {
       CacheClientProxy proxy = (CacheClientProxy) iter.next();
       if (proxy.hasRegisteredInterested()) {
         ClientProxyMembershipID proxyID = proxy.getProxyID();
@@ -1527,11 +1549,12 @@ public class CacheClientNotifier {
 
   /**
    * Return (possibly stale) list of all clients and their status
+   *
    * @return Map, with CacheClientProxy as a key and CacheClientStatus as a value
    */
   public Map getAllClients() {
     Map clients = new HashMap();
-    for (Iterator iter = this._clientProxies.values().iterator(); iter.hasNext(); ) {
+    for (Iterator iter = this._clientProxies.values().iterator(); iter.hasNext();) {
       CacheClientProxy proxy = (CacheClientProxy) iter.next();
       ClientProxyMembershipID proxyID = proxy.getProxyID();
       clients.put(proxyID, new CacheClientStatus(proxyID));
@@ -1541,12 +1564,13 @@ public class CacheClientNotifier {
 
   /**
    * Checks if there is any proxy present for the given durable client
+   *
    * @param durableId - id for the durable-client
    * @return - true if a proxy is present for the given durable client
    * @since GemFire 5.6
    */
   public boolean hasDurableClient(String durableId) {
-    for (Iterator iter = this._clientProxies.values().iterator(); iter.hasNext(); ) {
+    for (Iterator iter = this._clientProxies.values().iterator(); iter.hasNext();) {
       CacheClientProxy proxy = (CacheClientProxy) iter.next();
       ClientProxyMembershipID proxyID = proxy.getProxyID();
       if (durableId.equals(proxyID.getDurableId())) {
@@ -1558,12 +1582,13 @@ public class CacheClientNotifier {
 
   /**
    * Checks if there is any proxy which is primary for the given durable client
+   *
    * @param durableId - id for the durable-client
    * @return - true if a primary proxy is present for the given durable client
    * @since GemFire 5.6
    */
   public boolean hasPrimaryForDurableClient(String durableId) {
-    for (Iterator iter = this._clientProxies.values().iterator(); iter.hasNext(); ) {
+    for (Iterator iter = this._clientProxies.values().iterator(); iter.hasNext();) {
       CacheClientProxy proxy = (CacheClientProxy) iter.next();
       ClientProxyMembershipID proxyID = proxy.getProxyID();
       if (durableId.equals(proxyID.getDurableId())) {
@@ -1579,11 +1604,12 @@ public class CacheClientNotifier {
 
   /**
    * Returns (possibly stale) map of queue sizes for all clients notified by this server.
+   *
    * @return map with CacheClientProxy as key, and Integer as a value
    */
   public Map getClientQueueSizes() {
     Map/* <ClientProxyMembershipID,Integer> */ queueSizes = new HashMap();
-    for (Iterator iter = this._clientProxies.values().iterator(); iter.hasNext(); ) {
+    for (Iterator iter = this._clientProxies.values().iterator(); iter.hasNext();) {
       CacheClientProxy proxy = (CacheClientProxy) iter.next();
       queueSizes.put(proxy.getProxyID(), Integer.valueOf(proxy.getQueueSize()));
     }
@@ -1610,6 +1636,7 @@ public class CacheClientNotifier {
 
   /**
    * Removes an existing <code>CacheClientProxy</code> from the list of known client proxies
+   *
    * @param proxy The <code>CacheClientProxy</code> to remove
    */
   protected void removeClientProxy(CacheClientProxy proxy) {
@@ -1638,6 +1665,7 @@ public class CacheClientNotifier {
   /**
    * Returns an unmodifiable Collection of known <code>CacheClientProxy</code> instances. The
    * collection is not static so its contents may change.
+   *
    * @return the collection of known <code>CacheClientProxy</code> instances
    */
   public Collection<CacheClientProxy> getClientProxies() {
@@ -1682,18 +1710,18 @@ public class CacheClientNotifier {
         logger.debug("Cannot close running durable client: {}", durableClientId);
       }
... 4334 lines suppressed ...