You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by oz...@apache.org on 2015/02/24 16:32:22 UTC
[2/2] hadoop git commit: Revert "HADOOP-11602. Fix
toUpperCase/toLowerCase to use Locale.ENGLISH. (ozawa)"
Revert "HADOOP-11602. Fix toUpperCase/toLowerCase to use Locale.ENGLISH. (ozawa)"
This reverts commit 946456c6d88780abe0251b098dd771e9e1e93ab3.
Conflicts:
hadoop-common-project/hadoop-common/CHANGES.txt
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/QuotaByStorageTypeEntry.java
Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo
Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/9cedad11
Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/9cedad11
Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/9cedad11
Branch: refs/heads/trunk
Commit: 9cedad11d8d2197a54732667a15344983de5c437
Parents: b53fd71
Author: Tsuyoshi Ozawa <oz...@apache.org>
Authored: Wed Feb 25 00:32:04 2015 +0900
Committer: Tsuyoshi Ozawa <oz...@apache.org>
Committed: Wed Feb 25 00:32:04 2015 +0900
----------------------------------------------------------------------
.../classification/tools/StabilityOptions.java | 5 ++-
.../AltKerberosAuthenticationHandler.java | 6 ++--
.../authentication/util/TestKerberosUtil.java | 14 ++++----
hadoop-common-project/hadoop-common/CHANGES.txt | 2 --
.../org/apache/hadoop/conf/Configuration.java | 3 +-
.../org/apache/hadoop/crypto/CipherSuite.java | 3 +-
.../hadoop/crypto/key/JavaKeyStoreProvider.java | 3 +-
.../java/org/apache/hadoop/fs/FileSystem.java | 7 ++--
.../java/org/apache/hadoop/fs/StorageType.java | 3 +-
.../apache/hadoop/fs/permission/AclEntry.java | 6 ++--
.../org/apache/hadoop/fs/shell/find/Name.java | 5 ++-
.../io/compress/CompressionCodecFactory.java | 6 ++--
.../hadoop/metrics2/impl/MetricsConfig.java | 7 ++--
.../hadoop/metrics2/impl/MetricsSystemImpl.java | 2 +-
.../apache/hadoop/security/SecurityUtil.java | 5 ++-
.../hadoop/security/WhitelistBasedResolver.java | 4 +--
.../security/ssl/FileBasedKeyStoresFactory.java | 4 +--
.../apache/hadoop/security/ssl/SSLFactory.java | 3 +-
.../security/ssl/SSLHostnameVerifier.java | 10 +++---
.../DelegationTokenAuthenticationHandler.java | 3 +-
.../web/DelegationTokenAuthenticator.java | 5 ++-
.../org/apache/hadoop/util/StringUtils.java | 2 +-
.../java/org/apache/hadoop/ipc/TestIPC.java | 3 +-
.../java/org/apache/hadoop/ipc/TestSaslRPC.java | 3 +-
.../hadoop/security/TestSecurityUtil.java | 6 ++--
.../security/TestUserGroupInformation.java | 5 ++-
.../hadoop/test/TimedOutTestsListener.java | 4 +--
.../org/apache/hadoop/util/TestWinUtils.java | 7 ++--
.../java/org/apache/hadoop/nfs/NfsExports.java | 5 ++-
.../server/CheckUploadContentTypeFilter.java | 4 +--
.../hadoop/fs/http/server/FSOperations.java | 4 +--
.../http/server/HttpFSParametersProvider.java | 4 +--
.../org/apache/hadoop/lib/server/Server.java | 3 +-
.../service/hadoop/FileSystemAccessService.java | 5 ++-
.../org/apache/hadoop/lib/wsrs/EnumParam.java | 3 +-
.../apache/hadoop/lib/wsrs/EnumSetParam.java | 3 +-
.../hadoop/lib/wsrs/ParametersProvider.java | 3 +-
.../org/apache/hadoop/hdfs/XAttrHelper.java | 21 ++++--------
.../hadoop/hdfs/protocol/HdfsConstants.java | 3 +-
.../BlockStoragePolicySuite.java | 4 +--
.../hdfs/server/common/HdfsServerConstants.java | 5 ++-
.../hdfs/server/datanode/StorageLocation.java | 4 +--
.../hdfs/server/namenode/FSEditLogOp.java | 3 +-
.../namenode/QuotaByStorageTypeEntry.java | 5 ++-
.../hdfs/server/namenode/SecondaryNameNode.java | 2 +-
.../org/apache/hadoop/hdfs/tools/GetConf.java | 17 +++++-----
.../OfflineEditsVisitorFactory.java | 7 ++--
.../offlineImageViewer/FSImageHandler.java | 3 +-
.../org/apache/hadoop/hdfs/web/AuthFilter.java | 3 +-
.../org/apache/hadoop/hdfs/web/ParamFilter.java | 3 +-
.../hadoop/hdfs/web/WebHdfsFileSystem.java | 5 ++-
.../hadoop/hdfs/web/resources/EnumParam.java | 3 +-
.../hadoop/hdfs/web/resources/EnumSetParam.java | 4 +--
.../namenode/snapshot/TestSnapshotManager.java | 5 ++-
.../jobhistory/JobHistoryEventHandler.java | 4 +--
.../mapreduce/v2/app/webapp/AppController.java | 2 +-
.../apache/hadoop/mapreduce/TypeConverter.java | 3 +-
.../apache/hadoop/mapreduce/v2/util/MRApps.java | 5 ++-
.../hadoop/mapreduce/TestTypeConverter.java | 4 +--
.../java/org/apache/hadoop/mapred/Task.java | 3 +-
.../counters/FileSystemCounterGroup.java | 2 +-
.../mapreduce/filecache/DistributedCache.java | 4 +--
.../hadoop/mapreduce/lib/db/DBInputFormat.java | 4 +--
.../org/apache/hadoop/mapreduce/tools/CLI.java | 8 ++---
.../java/org/apache/hadoop/fs/TestDFSIO.java | 20 ++++++------
.../org/apache/hadoop/fs/TestFileSystem.java | 6 +---
.../org/apache/hadoop/fs/slive/Constants.java | 6 ++--
.../apache/hadoop/fs/slive/OperationData.java | 4 +--
.../apache/hadoop/fs/slive/OperationOutput.java | 4 +--
.../org/apache/hadoop/fs/slive/SliveTest.java | 3 +-
.../java/org/apache/hadoop/io/FileBench.java | 17 ++++------
.../org/apache/hadoop/mapred/TestMapRed.java | 3 +-
.../apache/hadoop/examples/DBCountPageView.java | 3 +-
.../plugin/versioninfo/VersionInfoMojo.java | 4 +--
.../fs/azure/AzureNativeFileSystemStore.java | 4 +--
.../apache/hadoop/tools/util/DistCpUtils.java | 11 +++----
.../java/org/apache/hadoop/tools/DistCpV1.java | 5 +--
.../gridmix/GridmixJobSubmissionPolicy.java | 3 +-
.../hadoop/tools/rumen/HadoopLogsAnalyzer.java | 34 +++++++++-----------
.../apache/hadoop/tools/rumen/JobBuilder.java | 3 +-
.../apache/hadoop/tools/rumen/LoggedTask.java | 3 +-
.../hadoop/tools/rumen/LoggedTaskAttempt.java | 3 +-
.../apache/hadoop/streaming/Environment.java | 2 +-
.../hadoop/yarn/client/cli/ApplicationCLI.java | 5 ++-
.../apache/hadoop/yarn/client/cli/NodeCLI.java | 4 +--
.../impl/pb/GetApplicationsRequestPBImpl.java | 6 ++--
.../pb/ApplicationSubmissionContextPBImpl.java | 3 +-
.../hadoop/yarn/webapp/hamlet/HamletGen.java | 5 ++-
.../webapp/AHSWebServices.java | 4 +--
.../timeline/webapp/TimelineWebServices.java | 3 +-
.../hadoop/yarn/server/webapp/WebServices.java | 10 +++---
.../server/resourcemanager/ClientRMService.java | 3 +-
.../resource/ResourceWeights.java | 3 +-
.../CapacitySchedulerConfiguration.java | 6 ++--
.../fair/FairSchedulerConfiguration.java | 3 +-
.../scheduler/fair/SchedulingPolicy.java | 3 +-
.../resourcemanager/webapp/NodesPage.java | 3 +-
.../resourcemanager/webapp/RMWebServices.java | 19 ++++-------
98 files changed, 192 insertions(+), 332 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-annotations/src/main/java/org/apache/hadoop/classification/tools/StabilityOptions.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-annotations/src/main/java/org/apache/hadoop/classification/tools/StabilityOptions.java b/hadoop-common-project/hadoop-annotations/src/main/java/org/apache/hadoop/classification/tools/StabilityOptions.java
index 657dbce..dbce31e 100644
--- a/hadoop-common-project/hadoop-annotations/src/main/java/org/apache/hadoop/classification/tools/StabilityOptions.java
+++ b/hadoop-common-project/hadoop-annotations/src/main/java/org/apache/hadoop/classification/tools/StabilityOptions.java
@@ -21,7 +21,6 @@ import com.sun.javadoc.DocErrorReporter;
import java.util.ArrayList;
import java.util.List;
-import java.util.Locale;
class StabilityOptions {
public static final String STABLE_OPTION = "-stable";
@@ -29,7 +28,7 @@ class StabilityOptions {
public static final String UNSTABLE_OPTION = "-unstable";
public static Integer optionLength(String option) {
- String opt = option.toLowerCase(Locale.ENGLISH);
+ String opt = option.toLowerCase();
if (opt.equals(UNSTABLE_OPTION)) return 1;
if (opt.equals(EVOLVING_OPTION)) return 1;
if (opt.equals(STABLE_OPTION)) return 1;
@@ -39,7 +38,7 @@ class StabilityOptions {
public static void validOptions(String[][] options,
DocErrorReporter reporter) {
for (int i = 0; i < options.length; i++) {
- String opt = options[i][0].toLowerCase(Locale.ENGLISH);
+ String opt = options[i][0].toLowerCase();
if (opt.equals(UNSTABLE_OPTION)) {
RootDocProcessor.stability = UNSTABLE_OPTION;
} else if (opt.equals(EVOLVING_OPTION)) {
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-auth/src/main/java/org/apache/hadoop/security/authentication/server/AltKerberosAuthenticationHandler.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-auth/src/main/java/org/apache/hadoop/security/authentication/server/AltKerberosAuthenticationHandler.java b/hadoop-common-project/hadoop-auth/src/main/java/org/apache/hadoop/security/authentication/server/AltKerberosAuthenticationHandler.java
index dae3b50..987330f 100644
--- a/hadoop-common-project/hadoop-auth/src/main/java/org/apache/hadoop/security/authentication/server/AltKerberosAuthenticationHandler.java
+++ b/hadoop-common-project/hadoop-auth/src/main/java/org/apache/hadoop/security/authentication/server/AltKerberosAuthenticationHandler.java
@@ -14,7 +14,6 @@
package org.apache.hadoop.security.authentication.server;
import java.io.IOException;
-import java.util.Locale;
import java.util.Properties;
import javax.servlet.ServletException;
import javax.servlet.http.HttpServletRequest;
@@ -69,8 +68,7 @@ public abstract class AltKerberosAuthenticationHandler
NON_BROWSER_USER_AGENTS, NON_BROWSER_USER_AGENTS_DEFAULT)
.split("\\W*,\\W*");
for (int i = 0; i < nonBrowserUserAgents.length; i++) {
- nonBrowserUserAgents[i] =
- nonBrowserUserAgents[i].toLowerCase(Locale.ENGLISH);
+ nonBrowserUserAgents[i] = nonBrowserUserAgents[i].toLowerCase();
}
}
@@ -122,7 +120,7 @@ public abstract class AltKerberosAuthenticationHandler
if (userAgent == null) {
return false;
}
- userAgent = userAgent.toLowerCase(Locale.ENGLISH);
+ userAgent = userAgent.toLowerCase();
boolean isBrowser = true;
for (String nonBrowserUserAgent : nonBrowserUserAgents) {
if (userAgent.contains(nonBrowserUserAgent)) {
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-auth/src/test/java/org/apache/hadoop/security/authentication/util/TestKerberosUtil.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-auth/src/test/java/org/apache/hadoop/security/authentication/util/TestKerberosUtil.java b/hadoop-common-project/hadoop-auth/src/test/java/org/apache/hadoop/security/authentication/util/TestKerberosUtil.java
index 89e07d1..b0e8f04 100644
--- a/hadoop-common-project/hadoop-auth/src/test/java/org/apache/hadoop/security/authentication/util/TestKerberosUtil.java
+++ b/hadoop-common-project/hadoop-auth/src/test/java/org/apache/hadoop/security/authentication/util/TestKerberosUtil.java
@@ -21,7 +21,6 @@ import java.io.IOException;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.List;
-import java.util.Locale;
import java.util.regex.Pattern;
import org.apache.directory.server.kerberos.shared.keytab.Keytab;
@@ -59,25 +58,24 @@ public class TestKerberosUtil {
// send null hostname
Assert.assertEquals("When no hostname is sent",
- service + "/" + localHostname.toLowerCase(Locale.ENGLISH),
+ service + "/" + localHostname.toLowerCase(),
KerberosUtil.getServicePrincipal(service, null));
// send empty hostname
Assert.assertEquals("When empty hostname is sent",
- service + "/" + localHostname.toLowerCase(Locale.ENGLISH),
+ service + "/" + localHostname.toLowerCase(),
KerberosUtil.getServicePrincipal(service, ""));
// send 0.0.0.0 hostname
Assert.assertEquals("When 0.0.0.0 hostname is sent",
- service + "/" + localHostname.toLowerCase(Locale.ENGLISH),
+ service + "/" + localHostname.toLowerCase(),
KerberosUtil.getServicePrincipal(service, "0.0.0.0"));
// send uppercase hostname
Assert.assertEquals("When uppercase hostname is sent",
- service + "/" + testHost.toLowerCase(Locale.ENGLISH),
+ service + "/" + testHost.toLowerCase(),
KerberosUtil.getServicePrincipal(service, testHost));
// send lowercase hostname
Assert.assertEquals("When lowercase hostname is sent",
- service + "/" + testHost.toLowerCase(Locale.ENGLISH),
- KerberosUtil.getServicePrincipal(
- service, testHost.toLowerCase(Locale.ENGLISH)));
+ service + "/" + testHost.toLowerCase(),
+ KerberosUtil.getServicePrincipal(service, testHost.toLowerCase()));
}
@Test
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/CHANGES.txt
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt
index 68d686e..e9ddb7d 100644
--- a/hadoop-common-project/hadoop-common/CHANGES.txt
+++ b/hadoop-common-project/hadoop-common/CHANGES.txt
@@ -405,8 +405,6 @@ Trunk (Unreleased)
HADOOP-11585. Fix formatting in Tracing.md (Masatake Iwasaki via aw)
- HADOOP-11602. Fix toUpperCase/toLowerCase to use Locale.ENGLISH. (ozawa)
-
HADOOP-11625. Minor fixes to command manual & SLA doc (aw)
OPTIMIZATIONS
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java
index 5909e62..02654b7 100644
--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java
+++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java
@@ -46,7 +46,6 @@ import java.util.Iterator;
import java.util.LinkedList;
import java.util.List;
import java.util.ListIterator;
-import java.util.Locale;
import java.util.Map;
import java.util.Map.Entry;
import java.util.Properties;
@@ -1452,7 +1451,7 @@ public class Configuration implements Iterable<Map.Entry<String,String>>,
return defaultValue;
}
- valueString = valueString.toLowerCase(Locale.ENGLISH);
+ valueString = valueString.toLowerCase();
if ("true".equals(valueString))
return true;
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/CipherSuite.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/CipherSuite.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/CipherSuite.java
index c5601eb..c9355d7 100644
--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/CipherSuite.java
+++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/CipherSuite.java
@@ -18,7 +18,6 @@
package org.apache.hadoop.crypto;
-import java.util.Locale;
import org.apache.hadoop.classification.InterfaceAudience;
/**
@@ -98,7 +97,7 @@ public enum CipherSuite {
String[] parts = name.split("/");
StringBuilder suffix = new StringBuilder();
for (String part : parts) {
- suffix.append(".").append(part.toLowerCase(Locale.ENGLISH));
+ suffix.append(".").append(part.toLowerCase());
}
return suffix.toString();
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/key/JavaKeyStoreProvider.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/key/JavaKeyStoreProvider.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/key/JavaKeyStoreProvider.java
index 9e09b6e..bfec1ef 100644
--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/key/JavaKeyStoreProvider.java
+++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/key/JavaKeyStoreProvider.java
@@ -53,7 +53,6 @@ import java.util.Date;
import java.util.Enumeration;
import java.util.HashMap;
import java.util.List;
-import java.util.Locale;
import java.util.Map;
import java.util.concurrent.locks.Lock;
import java.util.concurrent.locks.ReadWriteLock;
@@ -423,7 +422,7 @@ public class JavaKeyStoreProvider extends KeyProvider {
@Override
public KeyVersion createKey(String name, byte[] material,
Options options) throws IOException {
- Preconditions.checkArgument(name.equals(name.toLowerCase(Locale.ENGLISH)),
+ Preconditions.checkArgument(name.equals(name.toLowerCase()),
"Uppercase key names are unsupported: %s", name);
writeLock.lock();
try {
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileSystem.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileSystem.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileSystem.java
index f52ecad..cfa5198 100644
--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileSystem.java
+++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileSystem.java
@@ -33,7 +33,6 @@ import java.util.IdentityHashMap;
import java.util.Iterator;
import java.util.LinkedList;
import java.util.List;
-import java.util.Locale;
import java.util.Map;
import java.util.NoSuchElementException;
import java.util.ServiceLoader;
@@ -2796,10 +2795,8 @@ public abstract class FileSystem extends Configured implements Closeable {
}
Key(URI uri, Configuration conf, long unique) throws IOException {
- scheme = uri.getScheme() == null ?
- "" : uri.getScheme().toLowerCase(Locale.ENGLISH);
- authority = uri.getAuthority() == null ?
- "" : uri.getAuthority().toLowerCase(Locale.ENGLISH);
+ scheme = uri.getScheme()==null?"":uri.getScheme().toLowerCase();
+ authority = uri.getAuthority()==null?"":uri.getAuthority().toLowerCase();
this.unique = unique;
this.ugi = UserGroupInformation.getCurrentUser();
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/StorageType.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/StorageType.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/StorageType.java
index 24055d7..e306502 100644
--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/StorageType.java
+++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/StorageType.java
@@ -21,7 +21,6 @@ package org.apache.hadoop.fs;
import java.util.Arrays;
import java.util.ArrayList;
import java.util.List;
-import java.util.Locale;
import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.classification.InterfaceStability;
@@ -79,7 +78,7 @@ public enum StorageType {
}
public static StorageType parseStorageType(String s) {
- return StorageType.valueOf(s.toUpperCase(Locale.ENGLISH));
+ return StorageType.valueOf(s.toUpperCase());
}
private static List<StorageType> getNonTransientTypes() {
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/permission/AclEntry.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/permission/AclEntry.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/permission/AclEntry.java
index 6397564..b9def64 100644
--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/permission/AclEntry.java
+++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/permission/AclEntry.java
@@ -20,7 +20,6 @@ package org.apache.hadoop.fs.permission;
import java.util.ArrayList;
import java.util.Collection;
import java.util.List;
-import java.util.Locale;
import com.google.common.base.Objects;
@@ -107,7 +106,7 @@ public class AclEntry {
sb.append("default:");
}
if (type != null) {
- sb.append(type.toString().toLowerCase(Locale.ENGLISH));
+ sb.append(type.toString().toLowerCase());
}
sb.append(':');
if (name != null) {
@@ -264,8 +263,7 @@ public class AclEntry {
AclEntryType aclType = null;
try {
- aclType = Enum.valueOf(
- AclEntryType.class, split[index].toUpperCase(Locale.ENGLISH));
+ aclType = Enum.valueOf(AclEntryType.class, split[index].toUpperCase());
builder.setType(aclType);
index++;
} catch (IllegalArgumentException iae) {
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/shell/find/Name.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/shell/find/Name.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/shell/find/Name.java
index 4c937ef..88314c6 100644
--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/shell/find/Name.java
+++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/shell/find/Name.java
@@ -19,7 +19,6 @@ package org.apache.hadoop.fs.shell.find;
import java.io.IOException;
import java.util.Deque;
-import java.util.Locale;
import org.apache.hadoop.fs.GlobPattern;
import org.apache.hadoop.fs.shell.PathData;
@@ -74,7 +73,7 @@ final class Name extends BaseExpression {
public void prepare() throws IOException {
String argPattern = getArgument(1);
if (!caseSensitive) {
- argPattern = argPattern.toLowerCase(Locale.ENGLISH);
+ argPattern = argPattern.toLowerCase();
}
globPattern = new GlobPattern(argPattern);
}
@@ -83,7 +82,7 @@ final class Name extends BaseExpression {
public Result apply(PathData item, int depth) throws IOException {
String name = getPath(item).getName();
if (!caseSensitive) {
- name = name.toLowerCase(Locale.ENGLISH);
+ name = name.toLowerCase();
}
if (globPattern.matches(name)) {
return Result.PASS;
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/compress/CompressionCodecFactory.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/compress/CompressionCodecFactory.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/compress/CompressionCodecFactory.java
index 6d16823..eb35759 100644
--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/compress/CompressionCodecFactory.java
+++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/compress/CompressionCodecFactory.java
@@ -65,10 +65,10 @@ public class CompressionCodecFactory {
codecsByClassName.put(codec.getClass().getCanonicalName(), codec);
String codecName = codec.getClass().getSimpleName();
- codecsByName.put(codecName.toLowerCase(Locale.ENGLISH), codec);
+ codecsByName.put(codecName.toLowerCase(), codec);
if (codecName.endsWith("Codec")) {
codecName = codecName.substring(0, codecName.length() - "Codec".length());
- codecsByName.put(codecName.toLowerCase(Locale.ENGLISH), codec);
+ codecsByName.put(codecName.toLowerCase(), codec);
}
}
@@ -240,7 +240,7 @@ public class CompressionCodecFactory {
CompressionCodec codec = getCodecByClassName(codecName);
if (codec == null) {
// trying to get the codec by name in case the name was specified instead a class
- codec = codecsByName.get(codecName.toLowerCase(Locale.ENGLISH));
+ codec = codecsByName.get(codecName.toLowerCase());
}
return codec;
}
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsConfig.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsConfig.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsConfig.java
index f87ef4d..167205e 100644
--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsConfig.java
+++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsConfig.java
@@ -85,13 +85,12 @@ class MetricsConfig extends SubsetConfiguration {
private ClassLoader pluginLoader;
MetricsConfig(Configuration c, String prefix) {
- super(c, prefix.toLowerCase(Locale.ENGLISH), ".");
+ super(c, prefix.toLowerCase(Locale.US), ".");
}
static MetricsConfig create(String prefix) {
- return loadFirst(prefix, "hadoop-metrics2" + "-"
- + prefix.toLowerCase(Locale.ENGLISH)
- +".properties", DEFAULT_FILE_NAME);
+ return loadFirst(prefix, "hadoop-metrics2-"+ prefix.toLowerCase(Locale.US)
+ +".properties", DEFAULT_FILE_NAME);
}
static MetricsConfig create(String prefix, String... fileNames) {
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java
index 8964934..32b00f3 100644
--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java
+++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java
@@ -617,6 +617,6 @@ public class MetricsSystemImpl extends MetricsSystem implements MetricsSource {
String m = System.getProperty(MS_INIT_MODE_KEY);
String m2 = m == null ? System.getenv(MS_INIT_MODE_KEY) : m;
return InitMode.valueOf((m2 == null ? InitMode.NORMAL.name() : m2)
- .toUpperCase(Locale.ENGLISH));
+ .toUpperCase(Locale.US));
}
}
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SecurityUtil.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SecurityUtil.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SecurityUtil.java
index 355ea91..7cbee26 100644
--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SecurityUtil.java
+++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SecurityUtil.java
@@ -182,8 +182,7 @@ public class SecurityUtil {
if (fqdn == null || fqdn.isEmpty() || fqdn.equals("0.0.0.0")) {
fqdn = getLocalHostName();
}
- return components[0] + "/" + fqdn.toLowerCase(Locale.ENGLISH) + "@"
- + components[2];
+ return components[0] + "/" + fqdn.toLowerCase(Locale.US) + "@" + components[2];
}
static String getLocalHostName() throws UnknownHostException {
@@ -380,7 +379,7 @@ public class SecurityUtil {
}
host = addr.getAddress().getHostAddress();
} else {
- host = addr.getHostName().toLowerCase(Locale.ENGLISH);
+ host = addr.getHostName().toLowerCase();
}
return new Text(host + ":" + addr.getPort());
}
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/WhitelistBasedResolver.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/WhitelistBasedResolver.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/WhitelistBasedResolver.java
index 59d1492..dc0815e 100644
--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/WhitelistBasedResolver.java
+++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/WhitelistBasedResolver.java
@@ -19,7 +19,6 @@ package org.apache.hadoop.security;
import java.net.InetAddress;
import java.net.UnknownHostException;
-import java.util.Locale;
import java.util.Map;
import java.util.TreeMap;
@@ -139,8 +138,7 @@ public class WhitelistBasedResolver extends SaslPropertiesResolver {
QualityOfProtection.PRIVACY.toString());
for (int i=0; i < qop.length; i++) {
- qop[i] = QualityOfProtection.valueOf(
- qop[i].toUpperCase(Locale.ENGLISH)).getSaslQop();
+ qop[i] = QualityOfProtection.valueOf(qop[i].toUpperCase()).getSaslQop();
}
saslProps.put(Sasl.QOP, StringUtils.join(",", qop));
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/ssl/FileBasedKeyStoresFactory.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/ssl/FileBasedKeyStoresFactory.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/ssl/FileBasedKeyStoresFactory.java
index 50f96d6..4b81e17 100644
--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/ssl/FileBasedKeyStoresFactory.java
+++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/ssl/FileBasedKeyStoresFactory.java
@@ -33,7 +33,6 @@ import java.io.InputStream;
import java.security.GeneralSecurityException;
import java.security.KeyStore;
import java.text.MessageFormat;
-import java.util.Locale;
/**
* {@link KeyStoresFactory} implementation that reads the certificates from
@@ -95,8 +94,7 @@ public class FileBasedKeyStoresFactory implements KeyStoresFactory {
@VisibleForTesting
public static String resolvePropertyName(SSLFactory.Mode mode,
String template) {
- return MessageFormat.format(
- template, mode.toString().toLowerCase(Locale.ENGLISH));
+ return MessageFormat.format(template, mode.toString().toLowerCase());
}
/**
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/ssl/SSLFactory.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/ssl/SSLFactory.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/ssl/SSLFactory.java
index 370f09f..bbea33b 100644
--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/ssl/SSLFactory.java
+++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/ssl/SSLFactory.java
@@ -33,7 +33,6 @@ import javax.net.ssl.SSLSocketFactory;
import java.io.IOException;
import java.net.HttpURLConnection;
import java.security.GeneralSecurityException;
-import java.util.Locale;
/**
* Factory that creates SSLEngine and SSLSocketFactory instances using
@@ -139,7 +138,7 @@ public class SSLFactory implements ConnectionConfigurator {
private HostnameVerifier getHostnameVerifier(Configuration conf)
throws GeneralSecurityException, IOException {
return getHostnameVerifier(conf.get(SSL_HOSTNAME_VERIFIER_KEY, "DEFAULT").
- trim().toUpperCase(Locale.ENGLISH));
+ trim().toUpperCase());
}
public static HostnameVerifier getHostnameVerifier(String verifier)
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/ssl/SSLHostnameVerifier.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/ssl/SSLHostnameVerifier.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/ssl/SSLHostnameVerifier.java
index 7a905f1..dd5e67b 100644
--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/ssl/SSLHostnameVerifier.java
+++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/ssl/SSLHostnameVerifier.java
@@ -41,7 +41,6 @@ import java.util.Collection;
import java.util.Iterator;
import java.util.LinkedList;
import java.util.List;
-import java.util.Locale;
import java.util.Set;
import java.util.StringTokenizer;
import java.util.TreeSet;
@@ -366,7 +365,7 @@ public interface SSLHostnameVerifier extends javax.net.ssl.HostnameVerifier {
buf.append('<');
for (int i = 0; i < hosts.length; i++) {
String h = hosts[i];
- h = h != null ? h.trim().toLowerCase(Locale.ENGLISH) : "";
+ h = h != null ? h.trim().toLowerCase() : "";
hosts[i] = h;
if (i > 0) {
buf.append('/');
@@ -407,7 +406,7 @@ public interface SSLHostnameVerifier extends javax.net.ssl.HostnameVerifier {
out:
for (Iterator<String> it = names.iterator(); it.hasNext();) {
// Don't trim the CN, though!
- final String cn = it.next().toLowerCase(Locale.ENGLISH);
+ final String cn = it.next().toLowerCase();
// Store CN in StringBuffer in case we need to report an error.
buf.append(" <");
buf.append(cn);
@@ -425,8 +424,7 @@ public interface SSLHostnameVerifier extends javax.net.ssl.HostnameVerifier {
acceptableCountryWildcard(cn);
for (int i = 0; i < hosts.length; i++) {
- final String hostName =
- hosts[i].trim().toLowerCase(Locale.ENGLISH);
+ final String hostName = hosts[i].trim().toLowerCase();
if (doWildcard) {
match = hostName.endsWith(cn.substring(1));
if (match && strictWithSubDomains) {
@@ -481,7 +479,7 @@ public interface SSLHostnameVerifier extends javax.net.ssl.HostnameVerifier {
}
public static boolean isLocalhost(String host) {
- host = host != null ? host.trim().toLowerCase(Locale.ENGLISH) : "";
+ host = host != null ? host.trim().toLowerCase() : "";
if (host.startsWith("::1")) {
int x = host.lastIndexOf('%');
if (x >= 0) {
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/token/delegation/web/DelegationTokenAuthenticationHandler.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/token/delegation/web/DelegationTokenAuthenticationHandler.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/token/delegation/web/DelegationTokenAuthenticationHandler.java
index f1bcd5d..c18b5d3 100644
--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/token/delegation/web/DelegationTokenAuthenticationHandler.java
+++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/token/delegation/web/DelegationTokenAuthenticationHandler.java
@@ -23,7 +23,6 @@ import java.text.MessageFormat;
import java.util.HashMap;
import java.util.HashSet;
import java.util.LinkedHashMap;
-import java.util.Locale;
import java.util.Map;
import java.util.Properties;
import java.util.Set;
@@ -170,7 +169,7 @@ public abstract class DelegationTokenAuthenticationHandler
boolean requestContinues = true;
String op = ServletUtils.getParameter(request,
KerberosDelegationTokenAuthenticator.OP_PARAM);
- op = (op != null) ? op.toUpperCase(Locale.ENGLISH) : null;
+ op = (op != null) ? op.toUpperCase() : null;
if (DELEGATION_TOKEN_OPS.contains(op) &&
!request.getMethod().equals("OPTIONS")) {
KerberosDelegationTokenAuthenticator.DelegationTokenOperation dtOp =
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/token/delegation/web/DelegationTokenAuthenticator.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/token/delegation/web/DelegationTokenAuthenticator.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/token/delegation/web/DelegationTokenAuthenticator.java
index 5d826b7..d93f7ac 100644
--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/token/delegation/web/DelegationTokenAuthenticator.java
+++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/token/delegation/web/DelegationTokenAuthenticator.java
@@ -37,7 +37,6 @@ import java.net.InetSocketAddress;
import java.net.URL;
import java.net.URLEncoder;
import java.util.HashMap;
-import java.util.Locale;
import java.util.Map;
/**
@@ -287,8 +286,8 @@ public abstract class DelegationTokenAuthenticator implements Authenticator {
HttpExceptionUtils.validateResponse(conn, HttpURLConnection.HTTP_OK);
if (hasResponse) {
String contentType = conn.getHeaderField(CONTENT_TYPE);
- contentType = (contentType != null) ?
- contentType.toLowerCase(Locale.ENGLISH) : null;
+ contentType = (contentType != null) ? contentType.toLowerCase()
+ : null;
if (contentType != null &&
contentType.contains(APPLICATION_JSON_MIME)) {
try {
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/StringUtils.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/StringUtils.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/StringUtils.java
index c1acc7e..ff8edc3 100644
--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/StringUtils.java
+++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/StringUtils.java
@@ -901,7 +901,7 @@ public class StringUtils {
*/
public static String camelize(String s) {
StringBuilder sb = new StringBuilder();
- String[] words = split(s.toLowerCase(Locale.ENGLISH), ESCAPE_CHAR, '_');
+ String[] words = split(s.toLowerCase(Locale.US), ESCAPE_CHAR, '_');
for (String word : words)
sb.append(org.apache.commons.lang.StringUtils.capitalize(word));
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestIPC.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestIPC.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestIPC.java
index e52cb26..eb19f48 100644
--- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestIPC.java
+++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestIPC.java
@@ -43,7 +43,6 @@ import java.net.SocketTimeoutException;
import java.util.ArrayList;
import java.util.Collections;
import java.util.List;
-import java.util.Locale;
import java.util.Random;
import java.util.concurrent.BrokenBarrierException;
import java.util.concurrent.CountDownLatch;
@@ -1297,7 +1296,7 @@ public class TestIPC {
StringBuilder hexString = new StringBuilder();
- for (String line : hexdump.toUpperCase(Locale.ENGLISH).split("\n")) {
+ for (String line : hexdump.toUpperCase().split("\n")) {
hexString.append(line.substring(0, LAST_HEX_COL).replace(" ", ""));
}
return StringUtils.hexStringToByte(hexString.toString());
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestSaslRPC.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestSaslRPC.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestSaslRPC.java
index 9b25b77..903990b 100644
--- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestSaslRPC.java
+++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestSaslRPC.java
@@ -41,7 +41,6 @@ import java.security.Security;
import java.util.ArrayList;
import java.util.Collection;
import java.util.HashMap;
-import java.util.Locale;
import java.util.Map;
import java.util.Set;
import java.util.regex.Pattern;
@@ -182,7 +181,7 @@ public class TestSaslRPC {
StringBuilder sb = new StringBuilder();
int i = 0;
for (QualityOfProtection qop:qops){
- sb.append(qop.name().toLowerCase(Locale.ENGLISH));
+ sb.append(qop.name().toLowerCase());
if (++i < qops.length){
sb.append(",");
}
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/security/TestSecurityUtil.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/security/TestSecurityUtil.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/security/TestSecurityUtil.java
index 1d68f8a..4616c90 100644
--- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/security/TestSecurityUtil.java
+++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/security/TestSecurityUtil.java
@@ -103,13 +103,13 @@ public class TestSecurityUtil {
String realm = "@REALM";
String principalInConf = service + SecurityUtil.HOSTNAME_PATTERN + realm;
String hostname = "FooHost";
- String principal = service + hostname.toLowerCase(Locale.ENGLISH) + realm;
+ String principal = service + hostname.toLowerCase() + realm;
verify(principalInConf, hostname, principal);
}
@Test
public void testLocalHostNameForNullOrWild() throws Exception {
- String local = SecurityUtil.getLocalHostName().toLowerCase(Locale.ENGLISH);
+ String local = SecurityUtil.getLocalHostName().toLowerCase(Locale.US);
assertEquals("hdfs/" + local + "@REALM",
SecurityUtil.getServerPrincipal("hdfs/_HOST@REALM", (String)null));
assertEquals("hdfs/" + local + "@REALM",
@@ -260,7 +260,7 @@ public class TestSecurityUtil {
//LOG.info("address:"+addr+" host:"+host+" ip:"+ip+" port:"+port);
SecurityUtil.setTokenServiceUseIp(useIp);
- String serviceHost = useIp ? ip : host.toLowerCase(Locale.ENGLISH);
+ String serviceHost = useIp ? ip : host.toLowerCase();
Token<?> token = new Token<TokenIdentifier>();
Text service = new Text(serviceHost+":"+port);
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/security/TestUserGroupInformation.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/security/TestUserGroupInformation.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/security/TestUserGroupInformation.java
index d77d9b5..48b9b99 100644
--- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/security/TestUserGroupInformation.java
+++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/security/TestUserGroupInformation.java
@@ -41,7 +41,6 @@ import java.security.PrivilegedExceptionAction;
import java.util.Collection;
import java.util.ConcurrentModificationException;
import java.util.LinkedHashSet;
-import java.util.Locale;
import java.util.Set;
import static org.apache.hadoop.fs.CommonConfigurationKeys.HADOOP_USER_GROUP_METRICS_PERCENTILES_INTERVALS;
@@ -214,7 +213,7 @@ public class TestUserGroupInformation {
userName = userName.substring(sp + 1);
}
// user names are case insensitive on Windows. Make consistent
- userName = userName.toLowerCase(Locale.ENGLISH);
+ userName = userName.toLowerCase();
}
// get the groups
pp = Runtime.getRuntime().exec(Shell.WINDOWS ?
@@ -234,7 +233,7 @@ public class TestUserGroupInformation {
String loginUserName = login.getShortUserName();
if(Shell.WINDOWS) {
// user names are case insensitive on Windows. Make consistent
- loginUserName = loginUserName.toLowerCase(Locale.ENGLISH);
+ loginUserName = loginUserName.toLowerCase();
}
assertEquals(userName, loginUserName);
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/test/TimedOutTestsListener.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/test/TimedOutTestsListener.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/test/TimedOutTestsListener.java
index 4ebf29a..220ab1d 100644
--- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/test/TimedOutTestsListener.java
+++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/test/TimedOutTestsListener.java
@@ -27,7 +27,6 @@ import java.lang.management.ThreadMXBean;
import java.text.DateFormat;
import java.text.SimpleDateFormat;
import java.util.Date;
-import java.util.Locale;
import java.util.Map;
import org.junit.runner.notification.Failure;
@@ -94,8 +93,7 @@ public class TimedOutTestsListener extends RunListener {
thread.getPriority(),
thread.getId(),
Thread.State.WAITING.equals(thread.getState()) ?
- "in Object.wait()" :
- thread.getState().name().toLowerCase(Locale.ENGLISH),
+ "in Object.wait()" : thread.getState().name().toLowerCase(),
Thread.State.WAITING.equals(thread.getState()) ?
"WAITING (on object monitor)" : thread.getState()));
for (StackTraceElement stackTraceElement : e.getValue()) {
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/util/TestWinUtils.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/util/TestWinUtils.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/util/TestWinUtils.java
index 708fc4c..2d4e442 100644
--- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/util/TestWinUtils.java
+++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/util/TestWinUtils.java
@@ -27,7 +27,6 @@ import java.io.FileInputStream;
import java.io.FileOutputStream;
import java.io.FileWriter;
import java.io.IOException;
-import java.util.Locale;
import org.apache.commons.io.FileUtils;
import org.apache.commons.logging.Log;
@@ -383,10 +382,8 @@ public class TestWinUtils {
private void assertOwners(File file, String expectedUser,
String expectedGroup) throws IOException {
String [] args = lsF(file).trim().split("[\\|]");
- assertEquals(expectedUser.toLowerCase(Locale.ENGLISH),
- args[2].toLowerCase(Locale.ENGLISH));
- assertEquals(expectedGroup.toLowerCase(Locale.ENGLISH),
- args[3].toLowerCase(Locale.ENGLISH));
+ assertEquals(expectedUser.toLowerCase(), args[2].toLowerCase());
+ assertEquals(expectedGroup.toLowerCase(), args[3].toLowerCase());
}
@Test (timeout = 30000)
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/NfsExports.java
----------------------------------------------------------------------
diff --git a/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/NfsExports.java b/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/NfsExports.java
index cef8f99..b617ae5 100644
--- a/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/NfsExports.java
+++ b/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/NfsExports.java
@@ -20,7 +20,6 @@ package org.apache.hadoop.nfs;
import java.net.InetAddress;
import java.util.ArrayList;
import java.util.List;
-import java.util.Locale;
import java.util.regex.Pattern;
import org.apache.commons.logging.Log;
@@ -360,10 +359,10 @@ public class NfsExports {
AccessPrivilege privilege = AccessPrivilege.READ_ONLY;
switch (parts.length) {
case 1:
- host = parts[0].toLowerCase(Locale.ENGLISH).trim();
+ host = parts[0].toLowerCase().trim();
break;
case 2:
- host = parts[0].toLowerCase(Locale.ENGLISH).trim();
+ host = parts[0].toLowerCase().trim();
String option = parts[1].trim();
if ("rw".equalsIgnoreCase(option)) {
privilege = AccessPrivilege.READ_WRITE;
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/CheckUploadContentTypeFilter.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/CheckUploadContentTypeFilter.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/CheckUploadContentTypeFilter.java
index 7074ba2..836b4ce 100644
--- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/CheckUploadContentTypeFilter.java
+++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/CheckUploadContentTypeFilter.java
@@ -32,7 +32,6 @@ import javax.servlet.http.HttpServletRequest;
import javax.servlet.http.HttpServletResponse;
import java.io.IOException;
import java.util.HashSet;
-import java.util.Locale;
import java.util.Set;
/**
@@ -83,8 +82,7 @@ public class CheckUploadContentTypeFilter implements Filter {
String method = httpReq.getMethod();
if (method.equals("PUT") || method.equals("POST")) {
String op = httpReq.getParameter(HttpFSFileSystem.OP_PARAM);
- if (op != null &&
- UPLOAD_OPERATIONS.contains(op.toUpperCase(Locale.ENGLISH))) {
+ if (op != null && UPLOAD_OPERATIONS.contains(op.toUpperCase())) {
if ("true".equalsIgnoreCase(httpReq.getParameter(HttpFSParametersProvider.DataParam.NAME))) {
String contentType = httpReq.getContentType();
contentTypeOK =
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/FSOperations.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/FSOperations.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/FSOperations.java
index bc290a2..633589c 100644
--- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/FSOperations.java
+++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/FSOperations.java
@@ -43,7 +43,6 @@ import java.io.OutputStream;
import java.util.EnumSet;
import java.util.LinkedHashMap;
import java.util.List;
-import java.util.Locale;
import java.util.Map;
import java.util.Map.Entry;
@@ -569,8 +568,7 @@ public class FSOperations {
@Override
public JSONObject execute(FileSystem fs) throws IOException {
boolean deleted = fs.delete(path, recursive);
- return toJSON(
- HttpFSFileSystem.DELETE_JSON.toLowerCase(Locale.ENGLISH), deleted);
+ return toJSON(HttpFSFileSystem.DELETE_JSON.toLowerCase(), deleted);
}
}
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSParametersProvider.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSParametersProvider.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSParametersProvider.java
index 73853c4..271f3d9 100644
--- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSParametersProvider.java
+++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSParametersProvider.java
@@ -33,7 +33,6 @@ import org.apache.hadoop.lib.wsrs.StringParam;
import javax.ws.rs.ext.Provider;
import java.util.HashMap;
-import java.util.Locale;
import java.util.Map;
import java.util.regex.Pattern;
@@ -169,8 +168,7 @@ public class HttpFSParametersProvider extends ParametersProvider {
*/
public OperationParam(String operation) {
super(NAME, HttpFSFileSystem.Operation.class,
- HttpFSFileSystem.Operation.valueOf(
- operation.toUpperCase(Locale.ENGLISH)));
+ HttpFSFileSystem.Operation.valueOf(operation.toUpperCase()));
}
}
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/server/Server.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/server/Server.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/server/Server.java
index e2ef7a1..5c1bb4f 100644
--- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/server/Server.java
+++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/server/Server.java
@@ -36,7 +36,6 @@ import java.util.ArrayList;
import java.util.Collections;
import java.util.LinkedHashMap;
import java.util.List;
-import java.util.Locale;
import java.util.Map;
import java.util.Properties;
@@ -203,7 +202,7 @@ public class Server {
* @param config server configuration.
*/
public Server(String name, String homeDir, String configDir, String logDir, String tempDir, Configuration config) {
- this.name = Check.notEmpty(name, "name").trim().toLowerCase(Locale.ENGLISH);
+ this.name = Check.notEmpty(name, "name").trim().toLowerCase();
this.homeDir = Check.notEmpty(homeDir, "homeDir");
this.configDir = Check.notEmpty(configDir, "configDir");
this.logDir = Check.notEmpty(logDir, "logDir");
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/service/hadoop/FileSystemAccessService.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/service/hadoop/FileSystemAccessService.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/service/hadoop/FileSystemAccessService.java
index fd2e822..ccb15a3 100644
--- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/service/hadoop/FileSystemAccessService.java
+++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/service/hadoop/FileSystemAccessService.java
@@ -43,7 +43,6 @@ import java.net.URI;
import java.security.PrivilegedExceptionAction;
import java.util.Collection;
import java.util.HashSet;
-import java.util.Locale;
import java.util.Map;
import java.util.Set;
import java.util.concurrent.ConcurrentHashMap;
@@ -255,7 +254,7 @@ public class FileSystemAccessService extends BaseService implements FileSystemAc
private Set<String> toLowerCase(Collection<String> collection) {
Set<String> set = new HashSet<String>();
for (String value : collection) {
- set.add(value.toLowerCase(Locale.ENGLISH));
+ set.add(value.toLowerCase());
}
return set;
}
@@ -301,7 +300,7 @@ public class FileSystemAccessService extends BaseService implements FileSystemAc
protected void validateNamenode(String namenode) throws FileSystemAccessException {
if (nameNodeWhitelist.size() > 0 && !nameNodeWhitelist.contains("*")) {
- if (!nameNodeWhitelist.contains(namenode.toLowerCase(Locale.ENGLISH))) {
+ if (!nameNodeWhitelist.contains(namenode.toLowerCase())) {
throw new FileSystemAccessException(FileSystemAccessException.ERROR.H05, namenode, "not in whitelist");
}
}
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/wsrs/EnumParam.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/wsrs/EnumParam.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/wsrs/EnumParam.java
index 08eb60d..8baef67 100644
--- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/wsrs/EnumParam.java
+++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/wsrs/EnumParam.java
@@ -22,7 +22,6 @@ import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.util.StringUtils;
import java.util.Arrays;
-import java.util.Locale;
@InterfaceAudience.Private
public abstract class EnumParam<E extends Enum<E>> extends Param<E> {
@@ -35,7 +34,7 @@ public abstract class EnumParam<E extends Enum<E>> extends Param<E> {
@Override
protected E parse(String str) throws Exception {
- return Enum.valueOf(klass, str.toUpperCase(Locale.ENGLISH));
+ return Enum.valueOf(klass, str.toUpperCase());
}
@Override
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/wsrs/EnumSetParam.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/wsrs/EnumSetParam.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/wsrs/EnumSetParam.java
index 25158fd..8d79b71 100644
--- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/wsrs/EnumSetParam.java
+++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/wsrs/EnumSetParam.java
@@ -20,7 +20,6 @@ package org.apache.hadoop.lib.wsrs;
import java.util.Arrays;
import java.util.EnumSet;
import java.util.Iterator;
-import java.util.Locale;
import org.apache.hadoop.classification.InterfaceAudience;
@@ -38,7 +37,7 @@ public abstract class EnumSetParam<E extends Enum<E>> extends Param<EnumSet<E>>
final EnumSet<E> set = EnumSet.noneOf(klass);
if (!str.isEmpty()) {
for (String sub : str.split(",")) {
- set.add(Enum.valueOf(klass, sub.trim().toUpperCase(Locale.ENGLISH)));
+ set.add(Enum.valueOf(klass, sub.trim().toUpperCase()));
}
}
return set;
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/wsrs/ParametersProvider.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/wsrs/ParametersProvider.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/wsrs/ParametersProvider.java
index 9857244..4703a90 100644
--- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/wsrs/ParametersProvider.java
+++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/wsrs/ParametersProvider.java
@@ -33,7 +33,6 @@ import java.lang.reflect.Type;
import java.text.MessageFormat;
import java.util.HashMap;
import java.util.List;
-import java.util.Locale;
import java.util.Map;
/**
@@ -71,7 +70,7 @@ public class ParametersProvider
}
Enum op;
try {
- op = Enum.valueOf(enumClass, str.toUpperCase(Locale.ENGLISH));
+ op = Enum.valueOf(enumClass, str.toUpperCase());
} catch (IllegalArgumentException ex) {
throw new IllegalArgumentException(
MessageFormat.format("Invalid Operation [{0}]", str));
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/XAttrHelper.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/XAttrHelper.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/XAttrHelper.java
index 1b5b8eb..04364ccf 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/XAttrHelper.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/XAttrHelper.java
@@ -18,7 +18,6 @@
package org.apache.hadoop.hdfs;
import java.util.List;
-import java.util.Locale;
import java.util.Map;
import org.apache.hadoop.HadoopIllegalArgumentException;
@@ -58,22 +57,16 @@ public class XAttrHelper {
}
NameSpace ns;
- final String prefix = name.substring(0, prefixIndex)
- .toLowerCase(Locale.ENGLISH);
- if (prefix.equals(
- NameSpace.USER.toString().toLowerCase(Locale.ENGLISH))) {
+ final String prefix = name.substring(0, prefixIndex).toLowerCase();
+ if (prefix.equals(NameSpace.USER.toString().toLowerCase())) {
ns = NameSpace.USER;
- } else if (prefix.equals(
- NameSpace.TRUSTED.toString().toLowerCase(Locale.ENGLISH))) {
+ } else if (prefix.equals(NameSpace.TRUSTED.toString().toLowerCase())) {
ns = NameSpace.TRUSTED;
- } else if (prefix.equals(
- NameSpace.SYSTEM.toString().toLowerCase(Locale.ENGLISH))) {
+ } else if (prefix.equals(NameSpace.SYSTEM.toString().toLowerCase())) {
ns = NameSpace.SYSTEM;
- } else if (prefix.equals(
- NameSpace.SECURITY.toString().toLowerCase(Locale.ENGLISH))) {
+ } else if (prefix.equals(NameSpace.SECURITY.toString().toLowerCase())) {
ns = NameSpace.SECURITY;
- } else if (prefix.equals(
- NameSpace.RAW.toString().toLowerCase(Locale.ENGLISH))) {
+ } else if (prefix.equals(NameSpace.RAW.toString().toLowerCase())) {
ns = NameSpace.RAW;
} else {
throw new HadoopIllegalArgumentException("An XAttr name must be " +
@@ -152,7 +145,7 @@ public class XAttrHelper {
}
String namespace = xAttr.getNameSpace().toString();
- return namespace.toLowerCase(Locale.ENGLISH) + "." + xAttr.getName();
+ return namespace.toLowerCase() + "." + xAttr.getName();
}
/**
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocol/HdfsConstants.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocol/HdfsConstants.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocol/HdfsConstants.java
index 1769794..54da8eb 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocol/HdfsConstants.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocol/HdfsConstants.java
@@ -18,7 +18,6 @@
package org.apache.hadoop.hdfs.protocol;
import java.util.HashMap;
-import java.util.Locale;
import java.util.Map;
import org.apache.hadoop.classification.InterfaceAudience;
@@ -99,7 +98,7 @@ public class HdfsConstants {
/** Covert the given String to a RollingUpgradeAction. */
public static RollingUpgradeAction fromString(String s) {
- return MAP.get(s.toUpperCase(Locale.ENGLISH));
+ return MAP.get(s.toUpperCase());
}
}
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockStoragePolicySuite.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockStoragePolicySuite.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockStoragePolicySuite.java
index 9d8bdb5..0c03a42 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockStoragePolicySuite.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockStoragePolicySuite.java
@@ -30,7 +30,6 @@ import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import java.util.List;
-import java.util.Locale;
/** A collection of block storage policies. */
public class BlockStoragePolicySuite {
@@ -132,8 +131,7 @@ public class BlockStoragePolicySuite {
}
public static String buildXAttrName() {
- return XAttrNS.toString().toLowerCase(Locale.ENGLISH) + "."
- + STORAGE_POLICY_XATTR_NAME;
+ return XAttrNS.toString().toLowerCase() + "." + STORAGE_POLICY_XATTR_NAME;
}
public static XAttr buildXAttr(byte policyId) {
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/common/HdfsServerConstants.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/common/HdfsServerConstants.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/common/HdfsServerConstants.java
index 3674d2c..9bba2c9 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/common/HdfsServerConstants.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/common/HdfsServerConstants.java
@@ -20,7 +20,6 @@ package org.apache.hadoop.hdfs.server.common;
import java.io.DataInput;
import java.io.DataOutput;
import java.io.IOException;
-import java.util.Locale;
import java.util.regex.Matcher;
import java.util.regex.Pattern;
@@ -54,7 +53,7 @@ public final class HdfsServerConstants {
public String getOptionString() {
return StartupOption.ROLLINGUPGRADE.getName() + " "
- + name().toLowerCase(Locale.ENGLISH);
+ + name().toLowerCase();
}
public boolean matches(StartupOption option) {
@@ -77,7 +76,7 @@ public final class HdfsServerConstants {
public static String getAllOptionString() {
final StringBuilder b = new StringBuilder("<");
for(RollingUpgradeStartupOption opt : VALUES) {
- b.append(opt.name().toLowerCase(Locale.ENGLISH)).append("|");
+ b.append(opt.name().toLowerCase()).append("|");
}
b.setCharAt(b.length() - 1, '>');
return b.toString();
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/StorageLocation.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/StorageLocation.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/StorageLocation.java
index 1a553c9..7cda670 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/StorageLocation.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/StorageLocation.java
@@ -18,7 +18,6 @@
package org.apache.hadoop.hdfs.server.datanode;
-import java.util.Locale;
import java.util.regex.Pattern;
import java.io.File;
@@ -89,8 +88,7 @@ public class StorageLocation {
String classString = matcher.group(1);
location = matcher.group(2);
if (!classString.isEmpty()) {
- storageType = StorageType.valueOf(
- classString.toUpperCase(Locale.ENGLISH));
+ storageType = StorageType.valueOf(classString.toUpperCase());
}
}
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogOp.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogOp.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogOp.java
index d9ec543..c41a46a 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogOp.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogOp.java
@@ -75,7 +75,6 @@ import java.io.IOException;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.EnumMap;
-import java.util.Locale;
import java.util.List;
import java.util.zip.CheckedInputStream;
import java.util.zip.Checksum;
@@ -4349,7 +4348,7 @@ public abstract class FSEditLogOp {
public RollingUpgradeOp(FSEditLogOpCodes code, String name) {
super(code);
- this.name = name.toUpperCase(Locale.ENGLISH);
+ this.name = name.toUpperCase();
}
static RollingUpgradeOp getStartInstance(OpInstanceCache cache) {
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/QuotaByStorageTypeEntry.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/QuotaByStorageTypeEntry.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/QuotaByStorageTypeEntry.java
index 7059c64..d115acc 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/QuotaByStorageTypeEntry.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/QuotaByStorageTypeEntry.java
@@ -18,8 +18,7 @@
package org.apache.hadoop.hdfs.server.namenode;
import com.google.common.base.Objects;
-import org.apache.hadoop.fs.StorageType;
-import java.util.Locale;
+import org.apache.hadoop.hdfs.StorageType;
public class QuotaByStorageTypeEntry {
private StorageType type;
@@ -54,7 +53,7 @@ import java.util.Locale;
public String toString() {
StringBuilder sb = new StringBuilder();
assert (type != null);
- sb.append(type.toString().toLowerCase(Locale.ENGLISH));
+ sb.append(type.toString().toLowerCase());
sb.append(':');
sb.append(quota);
return sb.toString();
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/SecondaryNameNode.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/SecondaryNameNode.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/SecondaryNameNode.java
index 1157bb8..83e6426 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/SecondaryNameNode.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/SecondaryNameNode.java
@@ -587,7 +587,7 @@ public class SecondaryNameNode implements Runnable,
return 0;
}
- String cmd = opts.getCommand().toString().toLowerCase(Locale.ENGLISH);
+ String cmd = opts.getCommand().toString().toLowerCase();
int exitCode = 0;
try {
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/GetConf.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/GetConf.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/GetConf.java
index c380901..92a16cd 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/GetConf.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/GetConf.java
@@ -24,7 +24,6 @@ import java.security.PrivilegedExceptionAction;
import java.util.Arrays;
import java.util.HashMap;
import java.util.List;
-import java.util.Locale;
import java.util.Map;
import org.apache.hadoop.HadoopIllegalArgumentException;
@@ -80,19 +79,19 @@ public class GetConf extends Configured implements Tool {
private static final Map<String, CommandHandler> map;
static {
map = new HashMap<String, CommandHandler>();
- map.put(NAMENODE.getName().toLowerCase(Locale.ENGLISH),
+ map.put(NAMENODE.getName().toLowerCase(),
new NameNodesCommandHandler());
- map.put(SECONDARY.getName().toLowerCase(Locale.ENGLISH),
+ map.put(SECONDARY.getName().toLowerCase(),
new SecondaryNameNodesCommandHandler());
- map.put(BACKUP.getName().toLowerCase(Locale.ENGLISH),
+ map.put(BACKUP.getName().toLowerCase(),
new BackupNodesCommandHandler());
- map.put(INCLUDE_FILE.getName().toLowerCase(Locale.ENGLISH),
+ map.put(INCLUDE_FILE.getName().toLowerCase(),
new CommandHandler(DFSConfigKeys.DFS_HOSTS));
- map.put(EXCLUDE_FILE.getName().toLowerCase(Locale.ENGLISH),
+ map.put(EXCLUDE_FILE.getName().toLowerCase(),
new CommandHandler(DFSConfigKeys.DFS_HOSTS_EXCLUDE));
- map.put(NNRPCADDRESSES.getName().toLowerCase(Locale.ENGLISH),
+ map.put(NNRPCADDRESSES.getName().toLowerCase(),
new NNRpcAddressesCommandHandler());
- map.put(CONFKEY.getName().toLowerCase(Locale.ENGLISH),
+ map.put(CONFKEY.getName().toLowerCase(),
new PrintConfKeyCommandHandler());
}
@@ -117,7 +116,7 @@ public class GetConf extends Configured implements Tool {
}
public static CommandHandler getHandler(String cmd) {
- return map.get(cmd.toLowerCase(Locale.ENGLISH));
+ return map.get(cmd.toLowerCase());
}
}
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineEditsViewer/OfflineEditsVisitorFactory.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineEditsViewer/OfflineEditsVisitorFactory.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineEditsViewer/OfflineEditsVisitorFactory.java
index aa542d3..c4b8424 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineEditsViewer/OfflineEditsVisitorFactory.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineEditsViewer/OfflineEditsVisitorFactory.java
@@ -20,7 +20,6 @@ package org.apache.hadoop.hdfs.tools.offlineEditsViewer;
import java.io.FileOutputStream;
import java.io.IOException;
import java.io.OutputStream;
-import java.util.Locale;
import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.classification.InterfaceStability;
@@ -44,7 +43,7 @@ public class OfflineEditsVisitorFactory {
*/
static public OfflineEditsVisitor getEditsVisitor(String filename,
String processor, boolean printToScreen) throws IOException {
- if(processor.toLowerCase(Locale.ENGLISH).equals("binary")) {
+ if(processor.toLowerCase().equals("binary")) {
return new BinaryEditsVisitor(filename);
}
OfflineEditsVisitor vis;
@@ -60,9 +59,9 @@ public class OfflineEditsVisitorFactory {
outs[1] = System.out;
out = new TeeOutputStream(outs);
}
- if(processor.toLowerCase(Locale.ENGLISH).equals("xml")) {
+ if(processor.toLowerCase().equals("xml")) {
vis = new XmlEditsVisitor(out);
- } else if(processor.toLowerCase(Locale.ENGLISH).equals("stats")) {
+ } else if(processor.toLowerCase().equals("stats")) {
vis = new StatisticsEditsVisitor(out);
} else {
throw new IOException("Unknown proccesor " + processor +
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/FSImageHandler.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/FSImageHandler.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/FSImageHandler.java
index aa36516..43fcd69 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/FSImageHandler.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/FSImageHandler.java
@@ -37,7 +37,6 @@ import org.apache.hadoop.hdfs.web.JsonUtil;
import java.io.FileNotFoundException;
import java.io.IOException;
import java.util.List;
-import java.util.Locale;
import java.util.Map;
import static io.netty.handler.codec.http.HttpHeaders.Names.CONNECTION;
@@ -142,7 +141,7 @@ class FSImageHandler extends SimpleChannelInboundHandler<HttpRequest> {
private static String getOp(QueryStringDecoder decoder) {
Map<String, List<String>> parameters = decoder.parameters();
return parameters.containsKey("op")
- ? parameters.get("op").get(0).toUpperCase(Locale.ENGLISH) : null;
+ ? parameters.get("op").get(0).toUpperCase() : null;
}
private static String getPath(QueryStringDecoder decoder)
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/AuthFilter.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/AuthFilter.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/AuthFilter.java
index 1a0aaeb..b6ff4b6 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/AuthFilter.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/AuthFilter.java
@@ -23,7 +23,6 @@ import java.util.Enumeration;
import java.util.HashMap;
import java.util.Iterator;
import java.util.List;
-import java.util.Locale;
import java.util.Map;
import java.util.Properties;
@@ -97,7 +96,7 @@ public class AuthFilter extends AuthenticationFilter {
final Map<String, List<String>> m = new HashMap<String, List<String>>();
for(Map.Entry<String, String[]> entry : original.entrySet()) {
- final String key = entry.getKey().toLowerCase(Locale.ENGLISH);
+ final String key = entry.getKey().toLowerCase();
List<String> strings = m.get(key);
if (strings == null) {
strings = new ArrayList<String>();
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/ParamFilter.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/ParamFilter.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/ParamFilter.java
index 6a18377..2ae3445 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/ParamFilter.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/ParamFilter.java
@@ -19,7 +19,6 @@ package org.apache.hadoop.hdfs.web;
import java.net.URI;
import java.util.List;
-import java.util.Locale;
import java.util.Map;
import javax.ws.rs.core.MultivaluedMap;
@@ -76,7 +75,7 @@ public class ParamFilter implements ResourceFilter {
final MultivaluedMap<String, String> parameters) {
UriBuilder b = UriBuilder.fromUri(uri).replaceQuery("");
for(Map.Entry<String, List<String>> e : parameters.entrySet()) {
- final String key = e.getKey().toLowerCase(Locale.ENGLISH);
+ final String key = e.getKey().toLowerCase();
for(String v : e.getValue()) {
b = b.queryParam(key, v);
}
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/WebHdfsFileSystem.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/WebHdfsFileSystem.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/WebHdfsFileSystem.java
index b1026c0..938f7c7 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/WebHdfsFileSystem.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/WebHdfsFileSystem.java
@@ -32,7 +32,6 @@ import java.security.PrivilegedExceptionAction;
import java.util.ArrayList;
import java.util.EnumSet;
import java.util.List;
-import java.util.Locale;
import java.util.Map;
import java.util.StringTokenizer;
@@ -1243,7 +1242,7 @@ public class WebHdfsFileSystem extends FileSystem
if (query == null) {
return url;
}
- final String lower = query.toLowerCase(Locale.ENGLISH);
+ final String lower = query.toLowerCase();
if (!lower.startsWith(OFFSET_PARAM_PREFIX)
&& !lower.contains("&" + OFFSET_PARAM_PREFIX)) {
return url;
@@ -1254,7 +1253,7 @@ public class WebHdfsFileSystem extends FileSystem
for(final StringTokenizer st = new StringTokenizer(query, "&");
st.hasMoreTokens();) {
final String token = st.nextToken();
- if (!token.toLowerCase(Locale.ENGLISH).startsWith(OFFSET_PARAM_PREFIX)) {
+ if (!token.toLowerCase().startsWith(OFFSET_PARAM_PREFIX)) {
if (b == null) {
b = new StringBuilder("?").append(token);
} else {
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/resources/EnumParam.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/resources/EnumParam.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/resources/EnumParam.java
index 6b4ec2c..1703e3b 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/resources/EnumParam.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/resources/EnumParam.java
@@ -18,7 +18,6 @@
package org.apache.hadoop.hdfs.web.resources;
import java.util.Arrays;
-import java.util.Locale;
abstract class EnumParam<E extends Enum<E>> extends Param<E, EnumParam.Domain<E>> {
EnumParam(final Domain<E> domain, final E value) {
@@ -41,7 +40,7 @@ abstract class EnumParam<E extends Enum<E>> extends Param<E, EnumParam.Domain<E>
@Override
final E parse(final String str) {
- return Enum.valueOf(enumClass, str.toUpperCase(Locale.ENGLISH));
+ return Enum.valueOf(enumClass, str.toUpperCase());
}
}
}
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/resources/EnumSetParam.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/resources/EnumSetParam.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/resources/EnumSetParam.java
index 23ba16c..5adb5a6 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/resources/EnumSetParam.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/resources/EnumSetParam.java
@@ -20,7 +20,6 @@ package org.apache.hadoop.hdfs.web.resources;
import java.util.Arrays;
import java.util.EnumSet;
import java.util.Iterator;
-import java.util.Locale;
abstract class EnumSetParam<E extends Enum<E>> extends Param<EnumSet<E>, EnumSetParam.Domain<E>> {
/** Convert an EnumSet to a string of comma separated values. */
@@ -83,8 +82,7 @@ abstract class EnumSetParam<E extends Enum<E>> extends Param<EnumSet<E>, EnumSet
i = j > 0 ? j + 1 : 0;
j = str.indexOf(',', i);
final String sub = j >= 0? str.substring(i, j): str.substring(i);
- set.add(
- Enum.valueOf(enumClass, sub.trim().toUpperCase(Locale.ENGLISH)));
+ set.add(Enum.valueOf(enumClass, sub.trim().toUpperCase()));
}
}
return set;
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestSnapshotManager.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestSnapshotManager.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestSnapshotManager.java
index db8f290..ac6acf9 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestSnapshotManager.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestSnapshotManager.java
@@ -25,7 +25,6 @@ import static org.mockito.Mockito.mock;
import static org.mockito.Mockito.spy;
import java.util.ArrayList;
-import java.util.Locale;
import org.apache.hadoop.hdfs.protocol.SnapshotException;
import org.apache.hadoop.hdfs.server.namenode.FSDirectory;
@@ -71,7 +70,7 @@ public class TestSnapshotManager {
Assert.fail("Expected SnapshotException not thrown");
} catch (SnapshotException se) {
Assert.assertTrue(
- se.getMessage().toLowerCase(Locale.ENGLISH).contains("rollover"));
+ se.getMessage().toLowerCase().contains("rollover"));
}
// Delete a snapshot to free up a slot.
@@ -87,7 +86,7 @@ public class TestSnapshotManager {
Assert.fail("Expected SnapshotException not thrown");
} catch (SnapshotException se) {
Assert.assertTrue(
- se.getMessage().toLowerCase(Locale.ENGLISH).contains("rollover"));
+ se.getMessage().toLowerCase().contains("rollover"));
}
}
}