You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ozone.apache.org by ad...@apache.org on 2021/01/30 11:05:52 UTC
[ozone] branch master updated: HDDS-4737. Add ModifierOrder to
checkstyle rules (#1839)
This is an automated email from the ASF dual-hosted git repository.
adoroszlai pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/ozone.git
The following commit(s) were added to refs/heads/master by this push:
new c426595 HDDS-4737. Add ModifierOrder to checkstyle rules (#1839)
c426595 is described below
commit c42659531cd70595bbf4feaa0a410cc0d231754e
Author: lamber-ken <la...@163.com>
AuthorDate: Sat Jan 30 19:05:36 2021 +0800
HDDS-4737. Add ModifierOrder to checkstyle rules (#1839)
---
.../apache/hadoop/hdds/scm/net/NetConstants.java | 26 +++++++++++-----------
.../hadoop/hdds/scm/net/NodeSchemaLoader.java | 2 +-
.../hadoop/hdds/scm/net/NodeSchemaManager.java | 2 +-
.../java/org/apache/hadoop/ozone/OzoneConsts.java | 2 +-
.../org/apache/hadoop/ozone/common/Storage.java | 2 +-
.../org/apache/hadoop/hdds/scm/VersionInfo.java | 2 +-
.../container/common/DataNodeLayoutVersion.java | 2 +-
.../common/statemachine/StateContext.java | 12 +++++-----
.../container/common/utils/ContainerCache.java | 2 +-
.../background/BlockDeletingService.java | 4 ++--
.../hadoop/ozone/TestHddsSecureDatanodeInit.java | 2 +-
.../container/keyvalue/TestKeyValueHandler.java | 2 +-
.../TestKeyValueHandlerWithUnhealthyContainer.java | 2 +-
hadoop-hdds/dev-support/checkstyle/checkstyle.xml | 2 ++
.../hdds/security/token/BlockTokenVerifier.java | 2 +-
.../hadoop/hdds/security/x509/keys/KeyCodec.java | 8 +++----
.../hadoop/hdds/utils/RocksDBStoreMBean.java | 2 +-
.../apache/hadoop/hdds/utils/db/TypedTable.java | 2 +-
.../client/TestCertificateClientInit.java | 4 ++--
.../client/TestDefaultCertificateClient.java | 4 ++--
.../hadoop/hdds/utils/TestMetadataStore.java | 2 +-
.../hadoop/hdds/utils/db/TestDBConfigFromFile.java | 4 ++--
.../hdds/scm/block/SCMBlockDeletingService.java | 2 +-
.../hdds/scm/container/ContainerReplica.java | 12 +++++-----
.../scm/container/states/ContainerStateMap.java | 2 +-
.../hadoop/hdds/scm/block/TestBlockManager.java | 2 +-
.../hadoop/hdds/scm/container/MockNodeManager.java | 4 ++--
.../hdds/scm/node/TestSCMNodeStorageStatMap.java | 2 +-
.../scm/node/states/TestNode2ContainerMap.java | 4 ++--
.../scm/pipeline/TestPipelinePlacementPolicy.java | 4 ++--
.../hadoop/ozone/protocolPB/OzonePBHelper.java | 2 +-
.../ozone/security/OzoneTokenIdentifier.java | 2 +-
.../org/apache/hadoop/ozone/util/RadixTree.java | 2 +-
.../ozone/om/ha/TestOMFailoverProxyProvider.java | 6 ++---
.../apache/hadoop/ozone/util/TestRadixTree.java | 2 +-
.../loadgenerators/ReadOnlyLoadGenerator.java | 2 +-
.../apache/hadoop/ozone/TestMiniOzoneCluster.java | 6 ++---
.../ozoneimpl/TestOzoneContainerWithTLS.java | 2 +-
.../container/server/TestContainerServer.java | 2 +-
.../server/TestSecureContainerServer.java | 2 +-
.../apache/hadoop/ozone/om/TestScmSafeMode.java | 2 +-
.../apache/hadoop/ozone/om/KeyDeletingService.java | 2 +-
.../hadoop/ozone/om/OpenKeyCleanupService.java | 2 +-
.../hadoop/ozone/om/TrashOzoneFileSystem.java | 4 ++--
.../metrics/OzoneManagerDoubleBufferMetrics.java | 2 +-
.../hadoop/ozone/security/AWSV4AuthValidator.java | 2 +-
.../security/TestOMGetDelegationTokenRequest.java | 2 +-
.../security/TestOzoneBlockTokenSecretManager.java | 2 +-
.../TestOzoneDelegationTokenSecretManager.java | 2 +-
.../ozone/security/TestOzoneTokenIdentifier.java | 2 +-
.../fs/ozone/BasicRootedOzoneFileSystem.java | 2 +-
.../org/hadoop/ozone/recon/codegen/SqlDbUtils.java | 6 ++---
.../hadoop/ozone/recon/ReconRestServletModule.java | 2 +-
.../org/apache/hadoop/ozone/recon/ReconUtils.java | 2 +-
.../ozone/recon/fsck/ContainerHealthTask.java | 24 ++++++++++----------
.../hadoop/ozone/recon/scm/ReconNodeManager.java | 2 +-
.../recon/spi/impl/ContainerKeyPrefixCodec.java | 2 +-
.../hadoop/ozone/s3/AWSSignatureProcessor.java | 2 +-
.../hadoop/ozone/s3/OzoneClientProducer.java | 2 +-
.../hadoop/ozone/s3/exception/S3ErrorTable.java | 2 +-
.../ozone/s3/header/AuthorizationHeaderV2.java | 2 +-
.../ozone/s3/header/AuthorizationHeaderV4.java | 8 +++----
.../hadoop/ozone/s3/endpoint/TestListParts.java | 2 +-
.../s3/endpoint/TestMultipartUploadComplete.java | 4 ++--
.../s3/endpoint/TestMultipartUploadWithCopy.java | 8 +++----
.../hadoop/ozone/s3/endpoint/TestPartUpload.java | 2 +-
66 files changed, 124 insertions(+), 122 deletions(-)
diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NetConstants.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NetConstants.java
index 0e1b076..633ffba 100644
--- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NetConstants.java
+++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NetConstants.java
@@ -26,21 +26,21 @@ public final class NetConstants {
private NetConstants() {
// Prevent instantiation
}
- public final static char PATH_SEPARATOR = '/';
+ public static final char PATH_SEPARATOR = '/';
/** Path separator as a string. */
- public final static String PATH_SEPARATOR_STR = "/";
- public final static String SCOPE_REVERSE_STR = "~";
+ public static final String PATH_SEPARATOR_STR = "/";
+ public static final String SCOPE_REVERSE_STR = "~";
/** string representation of root. */
- public final static String ROOT = "";
- public final static int INNER_NODE_COST_DEFAULT = 1;
- public final static int NODE_COST_DEFAULT = 0;
- public final static int ANCESTOR_GENERATION_DEFAULT = 0;
- public final static int ROOT_LEVEL = 1;
- public final static String NODE_COST_PREFIX = "$";
- public final static String DEFAULT_RACK = "/default-rack";
- public final static String DEFAULT_NODEGROUP = "/default-nodegroup";
- public final static String DEFAULT_DATACENTER = "/default-datacenter";
- public final static String DEFAULT_REGION = "/default-dataregion";
+ public static final String ROOT = "";
+ public static final int INNER_NODE_COST_DEFAULT = 1;
+ public static final int NODE_COST_DEFAULT = 0;
+ public static final int ANCESTOR_GENERATION_DEFAULT = 0;
+ public static final int ROOT_LEVEL = 1;
+ public static final String NODE_COST_PREFIX = "$";
+ public static final String DEFAULT_RACK = "/default-rack";
+ public static final String DEFAULT_NODEGROUP = "/default-nodegroup";
+ public static final String DEFAULT_DATACENTER = "/default-datacenter";
+ public static final String DEFAULT_REGION = "/default-dataregion";
// Build-in network topology node schema
public static final NodeSchema ROOT_SCHEMA =
diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NodeSchemaLoader.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NodeSchemaLoader.java
index aa68724..25f4618 100644
--- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NodeSchemaLoader.java
+++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NodeSchemaLoader.java
@@ -65,7 +65,7 @@ public final class NodeSchemaLoader {
private static final String LAYER_DEFAULT_NAME = "default";
private static final int LAYOUT_VERSION = 1;
- private volatile static NodeSchemaLoader instance = null;
+ private static volatile NodeSchemaLoader instance = null;
private NodeSchemaLoader() {}
public static NodeSchemaLoader getInstance() {
diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NodeSchemaManager.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NodeSchemaManager.java
index 698b9da..eecd798 100644
--- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NodeSchemaManager.java
+++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NodeSchemaManager.java
@@ -43,7 +43,7 @@ public final class NodeSchemaManager {
// max level, includes ROOT level
private int maxLevel = -1;
- private volatile static NodeSchemaManager instance = null;
+ private static volatile NodeSchemaManager instance = null;
private NodeSchemaManager() {
}
diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/OzoneConsts.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/OzoneConsts.java
index 527b6dd..a1f490b 100644
--- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/OzoneConsts.java
+++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/OzoneConsts.java
@@ -108,7 +108,7 @@ public final class OzoneConsts {
public static final String FILE_HASH = "SHA-256";
public static final String MD5_HASH = "MD5";
- public final static String CHUNK_OVERWRITE = "OverWriteRequested";
+ public static final String CHUNK_OVERWRITE = "OverWriteRequested";
public static final int CHUNK_SIZE = 1 * 1024 * 1024; // 1 MB
public static final long KB = 1024L;
diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/common/Storage.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/common/Storage.java
index d5e1348..f4b7185 100644
--- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/common/Storage.java
+++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/common/Storage.java
@@ -130,7 +130,7 @@ public abstract class Storage {
return storageInfo;
}
- abstract protected Properties getNodeProperties();
+ protected abstract Properties getNodeProperties();
/**
* Sets the Node properties specific to OM/SCM.
diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/hdds/scm/VersionInfo.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/hdds/scm/VersionInfo.java
index 4e52046..62f5207 100644
--- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/hdds/scm/VersionInfo.java
+++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/hdds/scm/VersionInfo.java
@@ -23,7 +23,7 @@ package org.apache.hadoop.hdds.scm;
public final class VersionInfo {
// We will just be normal and use positive counting numbers for versions.
- private final static VersionInfo[] VERSION_INFOS =
+ private static final VersionInfo[] VERSION_INFOS =
{new VersionInfo("First version of SCM", 1)};
diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/DataNodeLayoutVersion.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/DataNodeLayoutVersion.java
index 2d58c39..b914659 100644
--- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/DataNodeLayoutVersion.java
+++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/DataNodeLayoutVersion.java
@@ -24,7 +24,7 @@ package org.apache.hadoop.ozone.container.common;
public final class DataNodeLayoutVersion {
// We will just be normal and use positive counting numbers for versions.
- private final static DataNodeLayoutVersion[] VERSION_INFOS =
+ private static final DataNodeLayoutVersion[] VERSION_INFOS =
{new DataNodeLayoutVersion(1, "HDDS Datanode LayOut Version 1")};
private final String description;
diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/StateContext.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/StateContext.java
index f87561a..28b7331 100644
--- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/StateContext.java
+++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/StateContext.java
@@ -73,22 +73,22 @@ import org.slf4j.LoggerFactory;
public class StateContext {
@VisibleForTesting
- final static String CONTAINER_REPORTS_PROTO_NAME =
+ static final String CONTAINER_REPORTS_PROTO_NAME =
ContainerReportsProto.getDescriptor().getFullName();
@VisibleForTesting
- final static String NODE_REPORT_PROTO_NAME =
+ static final String NODE_REPORT_PROTO_NAME =
NodeReportProto.getDescriptor().getFullName();
@VisibleForTesting
- final static String PIPELINE_REPORTS_PROTO_NAME =
+ static final String PIPELINE_REPORTS_PROTO_NAME =
PipelineReportsProto.getDescriptor().getFullName();
@VisibleForTesting
- final static String COMMAND_STATUS_REPORTS_PROTO_NAME =
+ static final String COMMAND_STATUS_REPORTS_PROTO_NAME =
CommandStatusReportsProto.getDescriptor().getFullName();
@VisibleForTesting
- final static String INCREMENTAL_CONTAINER_REPORT_PROTO_NAME =
+ static final String INCREMENTAL_CONTAINER_REPORT_PROTO_NAME =
IncrementalContainerReportProto.getDescriptor().getFullName();
// Accepted types of reports that can be queued to incrementalReportsQueue
- private final static Set<String> ACCEPTED_INCREMENTAL_REPORT_TYPE_SET =
+ private static final Set<String> ACCEPTED_INCREMENTAL_REPORT_TYPE_SET =
Sets.newHashSet(COMMAND_STATUS_REPORTS_PROTO_NAME,
INCREMENTAL_CONTAINER_REPORT_PROTO_NAME);
diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/utils/ContainerCache.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/utils/ContainerCache.java
index c56c743..af0958a 100644
--- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/utils/ContainerCache.java
+++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/utils/ContainerCache.java
@@ -68,7 +68,7 @@ public final class ContainerCache extends LRUMap {
* @param conf - Configuration.
* @return A instance of {@link ContainerCache}.
*/
- public synchronized static ContainerCache getInstance(
+ public static synchronized ContainerCache getInstance(
ConfigurationSource conf) {
if (cache == null) {
int cacheSize = conf.getInt(OzoneConfigKeys.OZONE_CONTAINER_CACHE_SIZE,
diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/statemachine/background/BlockDeletingService.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/statemachine/background/BlockDeletingService.java
index 3dab1fa..1fedbb6 100644
--- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/statemachine/background/BlockDeletingService.java
+++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/statemachine/background/BlockDeletingService.java
@@ -94,9 +94,9 @@ public class BlockDeletingService extends BackgroundService {
private final int containerLimitPerInterval;
// Task priority is useful when a to-delete block has weight.
- private final static int TASK_PRIORITY_DEFAULT = 1;
+ private static final int TASK_PRIORITY_DEFAULT = 1;
// Core pool size for container tasks
- private final static int BLOCK_DELETING_SERVICE_CORE_POOL_SIZE = 10;
+ private static final int BLOCK_DELETING_SERVICE_CORE_POOL_SIZE = 10;
public BlockDeletingService(OzoneContainer ozoneContainer,
long serviceInterval, long serviceTimeout, TimeUnit timeUnit,
diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/TestHddsSecureDatanodeInit.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/TestHddsSecureDatanodeInit.java
index a6cd862..f58ecf5 100644
--- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/TestHddsSecureDatanodeInit.java
+++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/TestHddsSecureDatanodeInit.java
@@ -66,7 +66,7 @@ public class TestHddsSecureDatanodeInit {
private static KeyCodec keyCodec;
private static CertificateCodec certCodec;
private static X509CertificateHolder certHolder;
- private final static String DN_COMPONENT = DNCertificateClient.COMPONENT_NAME;
+ private static final String DN_COMPONENT = DNCertificateClient.COMPONENT_NAME;
@BeforeClass
public static void setUp() throws Exception {
diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueHandler.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueHandler.java
index ad26958..dc4c3e5 100644
--- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueHandler.java
+++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueHandler.java
@@ -79,7 +79,7 @@ public class TestKeyValueHandler {
private static HddsDispatcher dispatcher;
private static KeyValueHandler handler;
- private final static String DATANODE_UUID = UUID.randomUUID().toString();
+ private static final String DATANODE_UUID = UUID.randomUUID().toString();
private static final long DUMMY_CONTAINER_ID = 9999;
diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueHandlerWithUnhealthyContainer.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueHandlerWithUnhealthyContainer.java
index d18d145..651df89 100644
--- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueHandlerWithUnhealthyContainer.java
+++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueHandlerWithUnhealthyContainer.java
@@ -51,7 +51,7 @@ public class TestKeyValueHandlerWithUnhealthyContainer {
public static final Logger LOG = LoggerFactory.getLogger(
TestKeyValueHandlerWithUnhealthyContainer.class);
- private final static String DATANODE_UUID = UUID.randomUUID().toString();
+ private static final String DATANODE_UUID = UUID.randomUUID().toString();
private static final long DUMMY_CONTAINER_ID = 9999;
@Test
diff --git a/hadoop-hdds/dev-support/checkstyle/checkstyle.xml b/hadoop-hdds/dev-support/checkstyle/checkstyle.xml
index 12fd04b..9ff858c 100644
--- a/hadoop-hdds/dev-support/checkstyle/checkstyle.xml
+++ b/hadoop-hdds/dev-support/checkstyle/checkstyle.xml
@@ -200,5 +200,7 @@
<!--<module name="TodoComment"/>-->
<module name="UpperEll"/>
+ <module name="ModifierOrder"/>
+
</module>
</module>
diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/token/BlockTokenVerifier.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/token/BlockTokenVerifier.java
index 46dec5a..94aae0d 100644
--- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/token/BlockTokenVerifier.java
+++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/token/BlockTokenVerifier.java
@@ -54,7 +54,7 @@ public class BlockTokenVerifier implements TokenVerifier {
private final CertificateClient caClient;
private final SecurityConfig conf;
private static boolean testStub = false;
- private final static Logger LOGGER =
+ private static final Logger LOGGER =
LoggerFactory.getLogger(BlockTokenVerifier.class);
public BlockTokenVerifier(SecurityConfig conf, CertificateClient caClient) {
diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/x509/keys/KeyCodec.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/x509/keys/KeyCodec.java
index 82873b0..654381d 100644
--- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/x509/keys/KeyCodec.java
+++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/x509/keys/KeyCodec.java
@@ -62,10 +62,10 @@ import static java.nio.file.attribute.PosixFilePermission.OWNER_WRITE;
* tools like OpenSSL can be used to read and decode these files.
*/
public class KeyCodec {
- public final static String PRIVATE_KEY = "PRIVATE KEY";
- public final static String PUBLIC_KEY = "PUBLIC KEY";
- public final static Charset DEFAULT_CHARSET = StandardCharsets.UTF_8;
- private final static Logger LOG =
+ public static final String PRIVATE_KEY = "PRIVATE KEY";
+ public static final String PUBLIC_KEY = "PUBLIC KEY";
+ public static final Charset DEFAULT_CHARSET = StandardCharsets.UTF_8;
+ private static final Logger LOG =
LoggerFactory.getLogger(KeyCodec.class);
private final Path location;
private final SecurityConfig securityConfig;
diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/RocksDBStoreMBean.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/RocksDBStoreMBean.java
index 8b9ef76..ba9a144 100644
--- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/RocksDBStoreMBean.java
+++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/RocksDBStoreMBean.java
@@ -60,7 +60,7 @@ public class RocksDBStoreMBean implements DynamicMBean, MetricsSource {
private static final Logger LOG =
LoggerFactory.getLogger(RocksDBStoreMBean.class);
- public final static String ROCKSDB_CONTEXT_PREFIX = "Rocksdb_";
+ public static final String ROCKSDB_CONTEXT_PREFIX = "Rocksdb_";
public RocksDBStoreMBean(Statistics statistics, String dbName) {
this.contextName = ROCKSDB_CONTEXT_PREFIX + dbName;
diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/TypedTable.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/TypedTable.java
index 5e44384..b7eb096 100644
--- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/TypedTable.java
+++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/TypedTable.java
@@ -58,7 +58,7 @@ public class TypedTable<KEY, VALUE> implements Table<KEY, VALUE> {
private final TableCache<CacheKey<KEY>, CacheValue<VALUE>> cache;
- private final static long EPOCH_DEFAULT = -1L;
+ private static final long EPOCH_DEFAULT = -1L;
/**
* Create an TypedTable from the raw table.
diff --git a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/x509/certificate/client/TestCertificateClientInit.java b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/x509/certificate/client/TestCertificateClientInit.java
index dcd9898..d0ece83 100644
--- a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/x509/certificate/client/TestCertificateClientInit.java
+++ b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/x509/certificate/client/TestCertificateClientInit.java
@@ -69,8 +69,8 @@ public class TestCertificateClientInit {
private KeyCodec dnKeyCodec;
private KeyCodec omKeyCodec;
private X509Certificate x509Certificate;
- private final static String DN_COMPONENT = DNCertificateClient.COMPONENT_NAME;
- private final static String OM_COMPONENT = OMCertificateClient.COMPONENT_NAME;
+ private static final String DN_COMPONENT = DNCertificateClient.COMPONENT_NAME;
+ private static final String OM_COMPONENT = OMCertificateClient.COMPONENT_NAME;
@Parameter
public boolean pvtKeyPresent;
diff --git a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/x509/certificate/client/TestDefaultCertificateClient.java b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/x509/certificate/client/TestDefaultCertificateClient.java
index 053520a..0534c8e 100644
--- a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/x509/certificate/client/TestDefaultCertificateClient.java
+++ b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/x509/certificate/client/TestDefaultCertificateClient.java
@@ -76,8 +76,8 @@ public class TestDefaultCertificateClient {
private Path dnMetaDirPath;
private SecurityConfig omSecurityConfig;
private SecurityConfig dnSecurityConfig;
- private final static String DN_COMPONENT = DNCertificateClient.COMPONENT_NAME;
- private final static String OM_COMPONENT = OMCertificateClient.COMPONENT_NAME;
+ private static final String DN_COMPONENT = DNCertificateClient.COMPONENT_NAME;
+ private static final String OM_COMPONENT = OMCertificateClient.COMPONENT_NAME;
private KeyCodec omKeyCodec;
private KeyCodec dnKeyCodec;
diff --git a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/TestMetadataStore.java b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/TestMetadataStore.java
index ed9bfb3..97603d4 100644
--- a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/TestMetadataStore.java
+++ b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/TestMetadataStore.java
@@ -62,7 +62,7 @@ import static java.nio.charset.StandardCharsets.UTF_8;
@RunWith(Parameterized.class)
public class TestMetadataStore {
- private final static int MAX_GETRANGE_LENGTH = 100;
+ private static final int MAX_GETRANGE_LENGTH = 100;
private final String storeImpl;
@Rule
public ExpectedException expectedException = ExpectedException.none();
diff --git a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/db/TestDBConfigFromFile.java b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/db/TestDBConfigFromFile.java
index 238b1d1..bce88c1 100644
--- a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/db/TestDBConfigFromFile.java
+++ b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/db/TestDBConfigFromFile.java
@@ -47,8 +47,8 @@ import static org.apache.hadoop.hdds.utils.db.DBConfigFromFile.getOptionsFileNam
* DBConf tests.
*/
public class TestDBConfigFromFile {
- private final static String DB_FILE = "test.db";
- private final static String INI_FILE = getOptionsFileNameFromDB(DB_FILE);
+ private static final String DB_FILE = "test.db";
+ private static final String INI_FILE = getOptionsFileNameFromDB(DB_FILE);
@Rule
public TemporaryFolder folder = new TemporaryFolder();
diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/block/SCMBlockDeletingService.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/block/SCMBlockDeletingService.java
index ceeaa10..bc18078 100644
--- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/block/SCMBlockDeletingService.java
+++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/block/SCMBlockDeletingService.java
@@ -57,7 +57,7 @@ public class SCMBlockDeletingService extends BackgroundService {
public static final Logger LOG =
LoggerFactory.getLogger(SCMBlockDeletingService.class);
- private final static int BLOCK_DELETING_SERVICE_CORE_POOL_SIZE = 1;
+ private static final int BLOCK_DELETING_SERVICE_CORE_POOL_SIZE = 1;
private final DeletedBlockLog deletedBlockLog;
private final ContainerManager containerManager;
private final NodeManager nodeManager;
diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/ContainerReplica.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/ContainerReplica.java
index 5118cc8..d0dbecd 100644
--- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/ContainerReplica.java
+++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/ContainerReplica.java
@@ -33,14 +33,14 @@ import java.util.UUID;
*/
public final class ContainerReplica implements Comparable<ContainerReplica> {
- final private ContainerID containerID;
- final private ContainerReplicaProto.State state;
- final private DatanodeDetails datanodeDetails;
- final private UUID placeOfBirth;
+ private final ContainerID containerID;
+ private final ContainerReplicaProto.State state;
+ private final DatanodeDetails datanodeDetails;
+ private final UUID placeOfBirth;
private Long sequenceId;
- final private long keyCount;
- final private long bytesUsed;
+ private final long keyCount;
+ private final long bytesUsed;
private ContainerReplica(final ContainerID containerID,
diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/states/ContainerStateMap.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/states/ContainerStateMap.java
index bafab56..fc20a51 100644
--- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/states/ContainerStateMap.java
+++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/states/ContainerStateMap.java
@@ -81,7 +81,7 @@ public class ContainerStateMap {
private static final Logger LOG =
LoggerFactory.getLogger(ContainerStateMap.class);
- private final static NavigableSet<ContainerID> EMPTY_SET =
+ private static final NavigableSet<ContainerID> EMPTY_SET =
Collections.unmodifiableNavigableSet(new TreeSet<>());
private final ContainerAttribute<LifeCycleState> lifeCycleStateMap;
diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/block/TestBlockManager.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/block/TestBlockManager.java
index 82bdd60..e0016f8 100644
--- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/block/TestBlockManager.java
+++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/block/TestBlockManager.java
@@ -80,7 +80,7 @@ public class TestBlockManager {
private MockNodeManager nodeManager;
private SCMPipelineManager pipelineManager;
private BlockManagerImpl blockManager;
- private final static long DEFAULT_BLOCK_SIZE = 128 * MB;
+ private static final long DEFAULT_BLOCK_SIZE = 128 * MB;
private static HddsProtos.ReplicationFactor factor;
private static HddsProtos.ReplicationType type;
private EventQueue eventQueue;
diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/MockNodeManager.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/MockNodeManager.java
index 513b60f..6a7e929 100644
--- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/MockNodeManager.java
+++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/MockNodeManager.java
@@ -70,8 +70,8 @@ import static org.apache.hadoop.hdds.protocol.proto.HddsProtos.NodeState.STALE;
* Test Helper for testing container Mapping.
*/
public class MockNodeManager implements NodeManager {
- public final static int NUM_PIPELINE_PER_METADATA_DISK = 2;
- private final static NodeData[] NODES = {
+ public static final int NUM_PIPELINE_PER_METADATA_DISK = 2;
+ private static final NodeData[] NODES = {
new NodeData(10L * OzoneConsts.TB, OzoneConsts.GB),
new NodeData(64L * OzoneConsts.TB, 100 * OzoneConsts.GB),
new NodeData(128L * OzoneConsts.TB, 256 * OzoneConsts.GB),
diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/node/TestSCMNodeStorageStatMap.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/node/TestSCMNodeStorageStatMap.java
index e12c643..bc3f623 100644
--- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/node/TestSCMNodeStorageStatMap.java
+++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/node/TestSCMNodeStorageStatMap.java
@@ -46,7 +46,7 @@ import java.util.concurrent.ConcurrentHashMap;
* Test Node Storage Map.
*/
public class TestSCMNodeStorageStatMap {
- private final static int DATANODE_COUNT = 100;
+ private static final int DATANODE_COUNT = 100;
private final long capacity = 10L * OzoneConsts.GB;
private final long used = 2L * OzoneConsts.GB;
private final long remaining = capacity - used;
diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/node/states/TestNode2ContainerMap.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/node/states/TestNode2ContainerMap.java
index 77ed907..6adc6c0 100644
--- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/node/states/TestNode2ContainerMap.java
+++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/node/states/TestNode2ContainerMap.java
@@ -39,8 +39,8 @@ import java.util.concurrent.ConcurrentHashMap;
* Test classes for Node2ContainerMap.
*/
public class TestNode2ContainerMap {
- private final static int DATANODE_COUNT = 300;
- private final static int CONTAINER_COUNT = 1000;
+ private static final int DATANODE_COUNT = 300;
+ private static final int CONTAINER_COUNT = 1000;
private final Map<UUID, TreeSet<ContainerID>> testData = new
ConcurrentHashMap<>();
diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestPipelinePlacementPolicy.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestPipelinePlacementPolicy.java
index 8a2b9c9..4afe587 100644
--- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestPipelinePlacementPolicy.java
+++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestPipelinePlacementPolicy.java
@@ -285,7 +285,7 @@ public class TestPipelinePlacementPolicy {
results.get(2).getNetworkLocation());
}
- private final static Node[] NODES = new NodeImpl[] {
+ private static final Node[] NODES = new NodeImpl[] {
new NodeImpl("h1", "/r1", NetConstants.NODE_COST_DEFAULT),
new NodeImpl("h2", "/r1", NetConstants.NODE_COST_DEFAULT),
new NodeImpl("h3", "/r2", NetConstants.NODE_COST_DEFAULT),
@@ -297,7 +297,7 @@ public class TestPipelinePlacementPolicy {
};
// 3 racks with single node.
- private final static Node[] SINGLE_NODE_RACK = new NodeImpl[] {
+ private static final Node[] SINGLE_NODE_RACK = new NodeImpl[] {
new NodeImpl("h1", "/r1", NetConstants.NODE_COST_DEFAULT),
new NodeImpl("h2", "/r2", NetConstants.NODE_COST_DEFAULT),
new NodeImpl("h3", "/r3", NetConstants.NODE_COST_DEFAULT)
diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/protocolPB/OzonePBHelper.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/protocolPB/OzonePBHelper.java
index 5939ca6..af8742d 100644
--- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/protocolPB/OzonePBHelper.java
+++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/protocolPB/OzonePBHelper.java
@@ -43,7 +43,7 @@ public final class OzonePBHelper {
* This map should not be accessed directly. Used the getFixedByteString
* methods instead.
*/
- private final static ConcurrentHashMap<Object, ByteString>
+ private static final ConcurrentHashMap<Object, ByteString>
FIXED_BYTESTRING_CACHE = new ConcurrentHashMap<>();
/**
diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/security/OzoneTokenIdentifier.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/security/OzoneTokenIdentifier.java
index 54eb7b9..01fcaf8 100644
--- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/security/OzoneTokenIdentifier.java
+++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/security/OzoneTokenIdentifier.java
@@ -46,7 +46,7 @@ import static org.apache.hadoop.ozone.protocol.proto.OzoneManagerProtocolProtos.
public class OzoneTokenIdentifier extends
AbstractDelegationTokenIdentifier {
- public final static Text KIND_NAME = new Text("OzoneToken");
+ public static final Text KIND_NAME = new Text("OzoneToken");
private String omCertSerialId;
private Type tokenType;
private String awsAccessId;
diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/util/RadixTree.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/util/RadixTree.java
index 597f58d..d7794db 100644
--- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/util/RadixTree.java
+++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/util/RadixTree.java
@@ -216,5 +216,5 @@ public class RadixTree<T> {
// root of a radix tree has a name of "/" and may optionally has it value.
private RadixNode root;
- private final static String PATH_DELIMITER = OzoneConsts.OZONE_URI_DELIMITER;
+ private static final String PATH_DELIMITER = OzoneConsts.OZONE_URI_DELIMITER;
}
diff --git a/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/om/ha/TestOMFailoverProxyProvider.java b/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/om/ha/TestOMFailoverProxyProvider.java
index e6c6af0..d5e5618 100644
--- a/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/om/ha/TestOMFailoverProxyProvider.java
+++ b/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/om/ha/TestOMFailoverProxyProvider.java
@@ -44,9 +44,9 @@ import static org.apache.hadoop.ozone.OzoneConfigKeys.
* Tests OMFailoverProxyProvider failover behaviour.
*/
public class TestOMFailoverProxyProvider {
- private final static String OM_SERVICE_ID = "om-service-test1";
- private final static String NODE_ID_BASE_STR = "omNode-";
- private final static String DUMMY_NODE_ADDR = "0.0.0.0:8080";
+ private static final String OM_SERVICE_ID = "om-service-test1";
+ private static final String NODE_ID_BASE_STR = "omNode-";
+ private static final String DUMMY_NODE_ADDR = "0.0.0.0:8080";
private OMFailoverProxyProvider provider;
private long waitBetweenRetries;
private int numNodes = 3;
diff --git a/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/util/TestRadixTree.java b/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/util/TestRadixTree.java
index 57b0268..b97b844 100644
--- a/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/util/TestRadixTree.java
+++ b/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/util/TestRadixTree.java
@@ -30,7 +30,7 @@ import static org.junit.Assert.assertEquals;
*/
public class TestRadixTree {
- final static RadixTree<Integer> ROOT = new RadixTree<>();
+ static final RadixTree<Integer> ROOT = new RadixTree<>();
@BeforeClass
public static void setupRadixTree() {
diff --git a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/ReadOnlyLoadGenerator.java b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/ReadOnlyLoadGenerator.java
index 45fffff..e8e6257 100644
--- a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/ReadOnlyLoadGenerator.java
+++ b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/ReadOnlyLoadGenerator.java
@@ -27,7 +27,7 @@ import java.nio.ByteBuffer;
public class ReadOnlyLoadGenerator extends LoadGenerator {
private final LoadBucket replBucket;
private final DataBuffer dataBuffer;
- private final static int NUM_KEYS = 10;
+ private static final int NUM_KEYS = 10;
public ReadOnlyLoadGenerator(DataBuffer dataBuffer, LoadBucket replBucket) {
this.dataBuffer = dataBuffer;
diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestMiniOzoneCluster.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestMiniOzoneCluster.java
index 0b68d4a..7090ec9 100644
--- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestMiniOzoneCluster.java
+++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestMiniOzoneCluster.java
@@ -77,9 +77,9 @@ public class TestMiniOzoneCluster {
private MiniOzoneCluster cluster;
private static OzoneConfiguration conf;
- private final static File TEST_ROOT = TestGenericTestUtils.getTestDir();
- private final static File WRITE_TMP = new File(TEST_ROOT, "write");
- private final static File READ_TMP = new File(TEST_ROOT, "read");
+ private static final File TEST_ROOT = TestGenericTestUtils.getTestDir();
+ private static final File WRITE_TMP = new File(TEST_ROOT, "write");
+ private static final File READ_TMP = new File(TEST_ROOT, "read");
@BeforeClass
public static void setup() {
diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/ozoneimpl/TestOzoneContainerWithTLS.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/ozoneimpl/TestOzoneContainerWithTLS.java
index 9a43bd4..da99caf 100644
--- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/ozoneimpl/TestOzoneContainerWithTLS.java
+++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/ozoneimpl/TestOzoneContainerWithTLS.java
@@ -68,7 +68,7 @@ import static org.apache.hadoop.ozone.OzoneConfigKeys.OZONE_SECURITY_ENABLED_KEY
@RunWith(Parameterized.class)
@Ignore("TODO:HDDS-1157")
public class TestOzoneContainerWithTLS {
- private final static Logger LOG = LoggerFactory.getLogger(
+ private static final Logger LOG = LoggerFactory.getLogger(
TestOzoneContainerWithTLS.class);
/**
* Set the timeout for every test.
diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/server/TestContainerServer.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/server/TestContainerServer.java
index a29453d..bbb8232 100644
--- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/server/TestContainerServer.java
+++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/server/TestContainerServer.java
@@ -85,7 +85,7 @@ public class TestContainerServer {
private static CertificateClient caClient;
@BeforeClass
- static public void setup() {
+ public static void setup() {
CONF.set(HddsConfigKeys.HDDS_METADATA_DIR_NAME, TEST_DIR);
caClient = new DNCertificateClient(new SecurityConfig(CONF));
}
diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/server/TestSecureContainerServer.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/server/TestSecureContainerServer.java
index f050e2a..acede34 100644
--- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/server/TestSecureContainerServer.java
+++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/server/TestSecureContainerServer.java
@@ -115,7 +115,7 @@ public class TestSecureContainerServer {
}
@BeforeClass
- static public void setup() throws Exception {
+ public static void setup() throws Exception {
DefaultMetricsSystem.setMiniClusterMode(true);
CONF.set(HddsConfigKeys.HDDS_METADATA_DIR_NAME, TEST_DIR);
CONF.setBoolean(OZONE_SECURITY_ENABLED_KEY, true);
diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestScmSafeMode.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestScmSafeMode.java
index d3e228f..8f9235d 100644
--- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestScmSafeMode.java
+++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestScmSafeMode.java
@@ -71,7 +71,7 @@ import static org.junit.Assert.fail;
@Ignore
public class TestScmSafeMode {
- private final static Logger LOG = LoggerFactory
+ private static final Logger LOG = LoggerFactory
.getLogger(TestScmSafeMode.class);
private static MiniOzoneCluster cluster = null;
private static MiniOzoneCluster.Builder builder = null;
diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/KeyDeletingService.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/KeyDeletingService.java
index 7c9dabd..8be6af6 100644
--- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/KeyDeletingService.java
+++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/KeyDeletingService.java
@@ -72,7 +72,7 @@ public class KeyDeletingService extends BackgroundService {
// Use only a single thread for KeyDeletion. Multiple threads would read
// from the same table and can send deletion requests for same key multiple
// times.
- private final static int KEY_DELETING_CORE_POOL_SIZE = 1;
+ private static final int KEY_DELETING_CORE_POOL_SIZE = 1;
private final OzoneManager ozoneManager;
private final ScmBlockLocationProtocol scmClient;
diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OpenKeyCleanupService.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OpenKeyCleanupService.java
index d1ab806..b0f19fc 100644
--- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OpenKeyCleanupService.java
+++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OpenKeyCleanupService.java
@@ -41,7 +41,7 @@ public class OpenKeyCleanupService extends BackgroundService {
private static final Logger LOG =
LoggerFactory.getLogger(OpenKeyCleanupService.class);
- private final static int OPEN_KEY_DELETING_CORE_POOL_SIZE = 2;
+ private static final int OPEN_KEY_DELETING_CORE_POOL_SIZE = 2;
private final KeyManager keyManager;
private final ScmBlockLocationProtocol scmClient;
diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/TrashOzoneFileSystem.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/TrashOzoneFileSystem.java
index 920194e..b30c54a 100644
--- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/TrashOzoneFileSystem.java
+++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/TrashOzoneFileSystem.java
@@ -78,7 +78,7 @@ public class TrashOzoneFileSystem extends FileSystem {
private final AtomicLong runCount;
- private final static ClientId CLIENT_ID = ClientId.randomId();
+ private static final ClientId CLIENT_ID = ClientId.randomId();
private static final Logger LOG =
LoggerFactory.getLogger(TrashOzoneFileSystem.class);
@@ -438,7 +438,7 @@ public class TrashOzoneFileSystem extends FileSystem {
}
private class DeleteIterator extends OzoneListingIterator {
- final private boolean recursive;
+ private final boolean recursive;
private List<String> keysList;
diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/ratis/metrics/OzoneManagerDoubleBufferMetrics.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/ratis/metrics/OzoneManagerDoubleBufferMetrics.java
index 0adbf37..3a59b8c 100644
--- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/ratis/metrics/OzoneManagerDoubleBufferMetrics.java
+++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/ratis/metrics/OzoneManagerDoubleBufferMetrics.java
@@ -58,7 +58,7 @@ public class OzoneManagerDoubleBufferMetrics {
"iteration")
private MutableGaugeFloat avgFlushTransactionsInOneIteration;
- public synchronized static OzoneManagerDoubleBufferMetrics create() {
+ public static synchronized OzoneManagerDoubleBufferMetrics create() {
if (instance != null) {
return instance;
} else {
diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/security/AWSV4AuthValidator.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/security/AWSV4AuthValidator.java
index 0a0e947..ca0bc18 100644
--- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/security/AWSV4AuthValidator.java
+++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/security/AWSV4AuthValidator.java
@@ -38,7 +38,7 @@ import java.security.NoSuchAlgorithmException;
**/
final class AWSV4AuthValidator {
- private final static Logger LOG =
+ private static final Logger LOG =
LoggerFactory.getLogger(AWSV4AuthValidator.class);
private static final String HMAC_SHA256_ALGORITHM = "HmacSHA256";
diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/security/TestOMGetDelegationTokenRequest.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/security/TestOMGetDelegationTokenRequest.java
index dfbb044..ca0ebc8 100644
--- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/security/TestOMGetDelegationTokenRequest.java
+++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/security/TestOMGetDelegationTokenRequest.java
@@ -56,7 +56,7 @@ public class TestOMGetDelegationTokenRequest extends
private OMRequest originalRequest;
private OMRequest modifiedRequest;
private OMGetDelegationTokenRequest omGetDelegationTokenRequest;
- final private String checkResponse = "";
+ private final String checkResponse = "";
@Before
public void setupGetDelegationToken() {
diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/TestOzoneBlockTokenSecretManager.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/TestOzoneBlockTokenSecretManager.java
index 9f33dd6..ad4fb27 100644
--- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/TestOzoneBlockTokenSecretManager.java
+++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/TestOzoneBlockTokenSecretManager.java
@@ -80,7 +80,7 @@ public class TestOzoneBlockTokenSecretManager {
private static final String BASEDIR = GenericTestUtils
.getTempPath(TestOzoneBlockTokenSecretManager.class.getSimpleName());
private BlockTokenVerifier tokenVerifier;
- private final static String ALGORITHM = "SHA256withRSA";
+ private static final String ALGORITHM = "SHA256withRSA";
@Rule
public ExpectedException exception = ExpectedException.none();
diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/TestOzoneDelegationTokenSecretManager.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/TestOzoneDelegationTokenSecretManager.java
index 2971ca0..2154974 100644
--- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/TestOzoneDelegationTokenSecretManager.java
+++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/TestOzoneDelegationTokenSecretManager.java
@@ -67,7 +67,7 @@ public class TestOzoneDelegationTokenSecretManager {
private long expiryTime;
private Text serviceRpcAdd;
private OzoneConfiguration conf;
- private final static Text TEST_USER = new Text("testUser");
+ private static final Text TEST_USER = new Text("testUser");
private long tokenMaxLifetime = 1000 * 20;
private long tokenRemoverScanInterval = 1000 * 20;
private S3SecretManager s3SecretManager;
diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/TestOzoneTokenIdentifier.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/TestOzoneTokenIdentifier.java
index 391759a..87bfe52 100644
--- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/TestOzoneTokenIdentifier.java
+++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/TestOzoneTokenIdentifier.java
@@ -103,7 +103,7 @@ public class TestOzoneTokenIdentifier {
}
@AfterClass
- static public void cleanUp() throws Exception {
+ public static void cleanUp() throws Exception {
FileUtil.fullyDelete(base);
KeyStoreTestUtil.cleanupSSLConfig(KEYSTORES_DIR, sslConfsDir);
}
diff --git a/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/BasicRootedOzoneFileSystem.java b/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/BasicRootedOzoneFileSystem.java
index 6fb1284..3f45c3f 100644
--- a/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/BasicRootedOzoneFileSystem.java
+++ b/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/BasicRootedOzoneFileSystem.java
@@ -409,7 +409,7 @@ public class BasicRootedOzoneFileSystem extends FileSystem {
}
private class DeleteIterator extends OzoneListingIterator {
- final private boolean recursive;
+ private final boolean recursive;
private final OzoneBucket bucket;
private final BasicRootedOzoneClientAdapterImpl adapterImpl;
diff --git a/hadoop-ozone/recon-codegen/src/main/java/org/hadoop/ozone/recon/codegen/SqlDbUtils.java b/hadoop-ozone/recon-codegen/src/main/java/org/hadoop/ozone/recon/codegen/SqlDbUtils.java
index 7e68541..2db54e6 100644
--- a/hadoop-ozone/recon-codegen/src/main/java/org/hadoop/ozone/recon/codegen/SqlDbUtils.java
+++ b/hadoop-ozone/recon-codegen/src/main/java/org/hadoop/ozone/recon/codegen/SqlDbUtils.java
@@ -37,10 +37,10 @@ import org.slf4j.LoggerFactory;
*/
public final class SqlDbUtils {
- public final static String DERBY_DRIVER_CLASS =
+ public static final String DERBY_DRIVER_CLASS =
"org.apache.derby.jdbc.EmbeddedDriver";
- public final static String SQLITE_DRIVER_CLASS = "org.sqlite.JDBC";
- public final static String DERBY_DISABLE_LOG_METHOD =
+ public static final String SQLITE_DRIVER_CLASS = "org.sqlite.JDBC";
+ public static final String DERBY_DISABLE_LOG_METHOD =
SqlDbUtils.class.getName() + ".disableDerbyLogFile";
private static final Logger LOG =
diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/ReconRestServletModule.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/ReconRestServletModule.java
index b657f79..584cb89 100644
--- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/ReconRestServletModule.java
+++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/ReconRestServletModule.java
@@ -48,7 +48,7 @@ public abstract class ReconRestServletModule extends ServletModule {
LoggerFactory.getLogger(ReconRestServletModule.class);
@Override
- abstract protected void configureServlets();
+ protected abstract void configureServlets();
/**
* Interface to provide packages for scanning.
diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/ReconUtils.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/ReconUtils.java
index 752cbce..9ec1136 100644
--- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/ReconUtils.java
+++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/ReconUtils.java
@@ -62,7 +62,7 @@ import org.slf4j.LoggerFactory;
@Singleton
public class ReconUtils {
- private final static int WRITE_BUFFER = 1048576; //1MB
+ private static final int WRITE_BUFFER = 1048576; //1MB
public ReconUtils() {
}
diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/fsck/ContainerHealthTask.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/fsck/ContainerHealthTask.java
index 9ceb5dd..a4ca664 100644
--- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/fsck/ContainerHealthTask.java
+++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/fsck/ContainerHealthTask.java
@@ -188,7 +188,7 @@ public class ContainerHealthTask extends ReconScmTask {
* Helper methods to generate and update the required database records for
* unhealthy containers.
*/
- static public class ContainerHealthRecords {
+ public static class ContainerHealthRecords {
/**
* Given an existing database record and a ContainerHealthStatus object,
@@ -205,7 +205,7 @@ public class ContainerHealthTask extends ReconScmTask {
* @param rec Existing database record from the UnhealthyContainers table.
* @return
*/
- static public boolean retainOrUpdateRecord(
+ public static boolean retainOrUpdateRecord(
ContainerHealthStatus container, UnhealthyContainersRecord rec) {
boolean returnValue = false;
switch(UnHealthyContainerStates.valueOf(rec.getContainerState())) {
@@ -227,7 +227,7 @@ public class ContainerHealthTask extends ReconScmTask {
return returnValue;
}
- static public List<UnhealthyContainers> generateUnhealthyRecords(
+ public static List<UnhealthyContainers> generateUnhealthyRecords(
ContainerHealthStatus container, long time) {
return generateUnhealthyRecords(container, new HashSet<>(), time);
}
@@ -240,7 +240,7 @@ public class ContainerHealthTask extends ReconScmTask {
* missing records which have not been seen already.
* @return List of UnhealthyContainer records to be stored in the DB
*/
- static public List<UnhealthyContainers> generateUnhealthyRecords(
+ public static List<UnhealthyContainers> generateUnhealthyRecords(
ContainerHealthStatus container, Set<String> recordForStateExists,
long time) {
List<UnhealthyContainers> records = new ArrayList<>();
@@ -280,7 +280,7 @@ public class ContainerHealthTask extends ReconScmTask {
return records;
}
- static private UnhealthyContainers recordForState(
+ private static UnhealthyContainers recordForState(
ContainerHealthStatus container, UnHealthyContainerStates state,
long time) {
UnhealthyContainers rec = new UnhealthyContainers();
@@ -300,7 +300,7 @@ public class ContainerHealthTask extends ReconScmTask {
return rec;
}
- static private boolean keepOverReplicatedRecord(
+ private static boolean keepOverReplicatedRecord(
ContainerHealthStatus container, UnhealthyContainersRecord rec) {
if (container.isOverReplicated()) {
updateExpectedReplicaCount(rec, container.getReplicationFactor());
@@ -311,7 +311,7 @@ public class ContainerHealthTask extends ReconScmTask {
return false;
}
- static private boolean keepUnderReplicatedRecord(
+ private static boolean keepUnderReplicatedRecord(
ContainerHealthStatus container, UnhealthyContainersRecord rec) {
if (container.isUnderReplicated()) {
updateExpectedReplicaCount(rec, container.getReplicationFactor());
@@ -322,7 +322,7 @@ public class ContainerHealthTask extends ReconScmTask {
return false;
}
- static private boolean keepMisReplicatedRecord(
+ private static boolean keepMisReplicatedRecord(
ContainerHealthStatus container, UnhealthyContainersRecord rec) {
if (container.isMisReplicated()) {
updateExpectedReplicaCount(rec, container.expectedPlacementCount());
@@ -341,28 +341,28 @@ public class ContainerHealthTask extends ReconScmTask {
* has really changed. The methods below ensure we do not update the Jooq
* record unless the values have changed and hence save a DB execution
*/
- static private void updateExpectedReplicaCount(
+ private static void updateExpectedReplicaCount(
UnhealthyContainersRecord rec, int expectedCount) {
if (rec.getExpectedReplicaCount() != expectedCount) {
rec.setExpectedReplicaCount(expectedCount);
}
}
- static private void updateActualReplicaCount(
+ private static void updateActualReplicaCount(
UnhealthyContainersRecord rec, int actualCount) {
if (rec.getActualReplicaCount() != actualCount) {
rec.setActualReplicaCount(actualCount);
}
}
- static private void updateReplicaDelta(
+ private static void updateReplicaDelta(
UnhealthyContainersRecord rec, int delta) {
if (rec.getReplicaDelta() != delta) {
rec.setReplicaDelta(delta);
}
}
- static private void updateReason(
+ private static void updateReason(
UnhealthyContainersRecord rec, String reason) {
if (!rec.getReason().equals(reason)) {
rec.setReason(reason);
diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/scm/ReconNodeManager.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/scm/ReconNodeManager.java
index d7a6104..bf848fe 100644
--- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/scm/ReconNodeManager.java
+++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/scm/ReconNodeManager.java
@@ -52,7 +52,7 @@ public class ReconNodeManager extends SCMNodeManager {
.getLogger(ReconNodeManager.class);
private Table<UUID, DatanodeDetails> nodeDB;
- private final static Set<Type> ALLOWED_COMMANDS =
+ private static final Set<Type> ALLOWED_COMMANDS =
ImmutableSet.of(reregisterCommand);
/**
diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/spi/impl/ContainerKeyPrefixCodec.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/spi/impl/ContainerKeyPrefixCodec.java
index 3c4b790..9625db6 100644
--- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/spi/impl/ContainerKeyPrefixCodec.java
+++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/spi/impl/ContainerKeyPrefixCodec.java
@@ -36,7 +36,7 @@ import com.google.common.primitives.Longs;
*/
public class ContainerKeyPrefixCodec implements Codec<ContainerKeyPrefix>{
- private final static String KEY_DELIMITER = "_";
+ private static final String KEY_DELIMITER = "_";
@Override
public byte[] toPersistedFormat(ContainerKeyPrefix containerKeyPrefix)
diff --git a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/AWSSignatureProcessor.java b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/AWSSignatureProcessor.java
index 26c1a3e..38e6641 100644
--- a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/AWSSignatureProcessor.java
+++ b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/AWSSignatureProcessor.java
@@ -64,7 +64,7 @@ import org.slf4j.LoggerFactory;
@RequestScoped
public class AWSSignatureProcessor implements SignatureProcessor {
- private final static Logger LOG =
+ private static final Logger LOG =
LoggerFactory.getLogger(AWSSignatureProcessor.class);
@Context
diff --git a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/OzoneClientProducer.java b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/OzoneClientProducer.java
index bee6e65..4d3b498 100644
--- a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/OzoneClientProducer.java
+++ b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/OzoneClientProducer.java
@@ -50,7 +50,7 @@ import org.slf4j.LoggerFactory;
@RequestScoped
public class OzoneClientProducer {
- private final static Logger LOG =
+ private static final Logger LOG =
LoggerFactory.getLogger(OzoneClientProducer.class);
private OzoneClient client;
diff --git a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/exception/S3ErrorTable.java b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/exception/S3ErrorTable.java
index a2c9f17..13eac13 100644
--- a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/exception/S3ErrorTable.java
+++ b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/exception/S3ErrorTable.java
@@ -33,7 +33,7 @@ import static org.apache.hadoop.ozone.s3.util.S3Consts.RANGE_NOT_SATISFIABLE;
*/
public final class S3ErrorTable {
- private final static Logger LOG = LoggerFactory.getLogger(
+ private static final Logger LOG = LoggerFactory.getLogger(
S3ErrorTable.class);
private S3ErrorTable() {
diff --git a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/header/AuthorizationHeaderV2.java b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/header/AuthorizationHeaderV2.java
index fe096ce..d29d7df 100644
--- a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/header/AuthorizationHeaderV2.java
+++ b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/header/AuthorizationHeaderV2.java
@@ -29,7 +29,7 @@ import static org.apache.commons.lang3.StringUtils.isBlank;
*/
public class AuthorizationHeaderV2 {
- public final static String IDENTIFIER = "AWS";
+ public static final String IDENTIFIER = "AWS";
private String authHeader;
private String identifier;
private String accessKeyID;
diff --git a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/header/AuthorizationHeaderV4.java b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/header/AuthorizationHeaderV4.java
index 1e48689..4f4bff3 100644
--- a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/header/AuthorizationHeaderV4.java
+++ b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/header/AuthorizationHeaderV4.java
@@ -45,12 +45,12 @@ import static org.apache.hadoop.ozone.s3.AWSSignatureProcessor.DATE_FORMATTER;
*/
public class AuthorizationHeaderV4 {
- private final static Logger LOG = LoggerFactory.getLogger(
+ private static final Logger LOG = LoggerFactory.getLogger(
AuthorizationHeaderV4.class);
- private final static String CREDENTIAL = "Credential=";
- private final static String SIGNEDHEADERS = "SignedHeaders=";
- private final static String SIGNATURE = "Signature=";
+ private static final String CREDENTIAL = "Credential=";
+ private static final String SIGNEDHEADERS = "SignedHeaders=";
+ private static final String SIGNATURE = "Signature=";
private String authHeader;
private String algorithm;
diff --git a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestListParts.java b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestListParts.java
index 337bfde..92fe20c 100644
--- a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestListParts.java
+++ b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestListParts.java
@@ -44,7 +44,7 @@ import static org.mockito.Mockito.when;
public class TestListParts {
- private final static ObjectEndpoint REST = new ObjectEndpoint();
+ private static final ObjectEndpoint REST = new ObjectEndpoint();
private static String uploadID;
@BeforeClass
diff --git a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestMultipartUploadComplete.java b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestMultipartUploadComplete.java
index 80b5742..fc9c1b4 100644
--- a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestMultipartUploadComplete.java
+++ b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestMultipartUploadComplete.java
@@ -51,8 +51,8 @@ import static org.mockito.Mockito.when;
public class TestMultipartUploadComplete {
- private final static ObjectEndpoint REST = new ObjectEndpoint();
- private final static OzoneClient CLIENT = new OzoneClientStub();
+ private static final ObjectEndpoint REST = new ObjectEndpoint();
+ private static final OzoneClient CLIENT = new OzoneClientStub();
@BeforeClass
public static void setUp() throws Exception {
diff --git a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestMultipartUploadWithCopy.java b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestMultipartUploadWithCopy.java
index 764a03e..95c8fef 100644
--- a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestMultipartUploadWithCopy.java
+++ b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestMultipartUploadWithCopy.java
@@ -58,12 +58,12 @@ import static org.mockito.Mockito.when;
public class TestMultipartUploadWithCopy {
- private final static ObjectEndpoint REST = new ObjectEndpoint();
+ private static final ObjectEndpoint REST = new ObjectEndpoint();
- private final static String KEY = "key2";
- private final static String EXISTING_KEY = "key1";
+ private static final String KEY = "key2";
+ private static final String EXISTING_KEY = "key1";
private static final String EXISTING_KEY_CONTENT = "testkey";
- private final static OzoneClient CLIENT = new OzoneClientStub();
+ private static final OzoneClient CLIENT = new OzoneClientStub();
private static final int RANGE_FROM = 2;
private static final int RANGE_TO = 4;
diff --git a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestPartUpload.java b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestPartUpload.java
index 8834f9d..33fdb45 100644
--- a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestPartUpload.java
+++ b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestPartUpload.java
@@ -46,7 +46,7 @@ import static org.mockito.Mockito.when;
*/
public class TestPartUpload {
- private final static ObjectEndpoint REST = new ObjectEndpoint();
+ private static final ObjectEndpoint REST = new ObjectEndpoint();
@BeforeClass
public static void setUp() throws Exception {
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@ozone.apache.org
For additional commands, e-mail: commits-help@ozone.apache.org