You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by aa...@apache.org on 2016/02/10 15:47:03 UTC

hadoop git commit: Revert "HDFS-9686. Remove useless boxing/unboxing code. Contributed by Kousuke Saruta."

Repository: hadoop
Updated Branches:
  refs/heads/trunk e9a622606 -> 5e565ce7a


Revert "HDFS-9686. Remove useless boxing/unboxing code. Contributed by Kousuke Saruta."

This reverts commit fe124da5ffc16e4795c3dd5542accd58361e1b08.


Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo
Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/5e565ce7
Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/5e565ce7
Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/5e565ce7

Branch: refs/heads/trunk
Commit: 5e565ce7a7e194f82e935b0e9808744e6bd64287
Parents: e9a6226
Author: Akira Ajisaka <aa...@apache.org>
Authored: Wed Feb 10 23:43:33 2016 +0900
Committer: Akira Ajisaka <aa...@apache.org>
Committed: Wed Feb 10 23:43:33 2016 +0900

----------------------------------------------------------------------
 .../apache/hadoop/lib/servlet/ServerWebApp.java | 23 +++----
 hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt     |  3 -
 .../hdfs/server/namenode/FSEditLogOp.java       | 41 +++++-------
 .../top/window/RollingWindowManager.java        |  8 +--
 .../hdfs/ErasureCodeBenchmarkThroughput.java    | 17 +++--
 .../shortcircuit/TestShortCircuitLocalRead.java | 66 ++++++++------------
 .../mapreduce/v2/app/TestStagingCleanup.java    | 57 -----------------
 7 files changed, 60 insertions(+), 155 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/hadoop/blob/5e565ce7/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/ServerWebApp.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/ServerWebApp.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/ServerWebApp.java
index 5e855de..c745f45 100644
--- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/ServerWebApp.java
+++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/ServerWebApp.java
@@ -36,9 +36,7 @@ import java.text.MessageFormat;
  * and uses its lifecycle to start and stop the server.
  */
 @InterfaceAudience.Private
-public abstract class ServerWebApp
-       extends Server
-       implements ServletContextListener {
+public abstract class ServerWebApp extends Server implements ServletContextListener {
 
   private static final String HOME_DIR = ".home.dir";
   private static final String CONFIG_DIR = ".config.dir";
@@ -63,8 +61,8 @@ public abstract class ServerWebApp
   /**
    * Constructor for testing purposes.
    */
-  protected ServerWebApp(String name, String homeDir, String configDir,
-                         String logDir, String tempDir, Configuration config) {
+  protected ServerWebApp(String name, String homeDir, String configDir, String logDir, String tempDir,
+                         Configuration config) {
     super(name, homeDir, configDir, logDir, tempDir, config);
   }
 
@@ -122,8 +120,7 @@ public abstract class ServerWebApp
       String sysProp = name + HOME_DIR;
       homeDir = System.getProperty(sysProp);
       if (homeDir == null) {
-        throw new IllegalArgumentException(MessageFormat.format(
-            "System property [{0}] not defined", sysProp));
+        throw new IllegalArgumentException(MessageFormat.format("System property [{0}] not defined", sysProp));
       }
     }
     return homeDir;
@@ -163,8 +160,7 @@ public abstract class ServerWebApp
   }
 
   /**
-   * Resolves the host and port InetSocketAddress the
-   * web server is listening to.
+   * Resolves the host and port InetSocketAddress the web server is listening to.
    * <p>
    * This implementation looks for the following 2 properties:
    * <ul>
@@ -172,10 +168,8 @@ public abstract class ServerWebApp
    *   <li>#SERVER_NAME#.http.port</li>
    * </ul>
    *
-   * @return the host and port InetSocketAddress the
-   *         web server is listening to.
-   * @throws ServerException thrown
-   *         if any of the above 2 properties is not defined.
+   * @return the host and port InetSocketAddress the web server is listening to.
+   * @throws ServerException thrown if any of the above 2 properties is not defined.
    */
   protected InetSocketAddress resolveAuthority() throws ServerException {
     String hostnameKey = getName() + HTTP_HOSTNAME;
@@ -239,7 +233,6 @@ public abstract class ServerWebApp
    *
    */
   public boolean isSslEnabled() {
-    return Boolean.parseBoolean(
-      System.getProperty(getName() + SSL_ENABLED, "false"));
+    return Boolean.valueOf(System.getProperty(getName() + SSL_ENABLED, "false"));
   }
 }

http://git-wip-us.apache.org/repos/asf/hadoop/blob/5e565ce7/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
index 4b8bdcb..2901fd8 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
+++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
@@ -1962,9 +1962,6 @@ Release 2.8.0 - UNRELEASED
 
     HDFS-9726. Refactor IBR code to a new class.  (szetszwo)
 
-    HDFS-9686. Remove useless boxing/unboxing code.
-    (Kousuke Saruta via aajisaka)
-
   BUG FIXES
 
     HDFS-7501. TransactionsSinceLastCheckpoint can be negative on SBNs.

http://git-wip-us.apache.org/repos/asf/hadoop/blob/5e565ce7/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogOp.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogOp.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogOp.java
index a8389f0..93d1b33 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogOp.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogOp.java
@@ -414,9 +414,7 @@ public abstract class FSEditLogOp {
   }
 
   @SuppressWarnings("unchecked")
-  static abstract class AddCloseOp
-         extends FSEditLogOp
-          implements BlockListUpdatingOp {
+  static abstract class AddCloseOp extends FSEditLogOp implements BlockListUpdatingOp {
     int length;
     long inodeId;
     String path;
@@ -637,8 +635,7 @@ public abstract class FSEditLogOp {
             NameNodeLayoutVersion.Feature.BLOCK_STORAGE_POLICY, logVersion)) {
           this.storagePolicyId = FSImageSerialization.readByte(in);
         } else {
-          this.storagePolicyId =
-              HdfsConstants.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED;
+          this.storagePolicyId = HdfsConstants.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED;
         }
         // read clientId and callId
         readRpcIds(in, logVersion);
@@ -718,7 +715,7 @@ public abstract class FSEditLogOp {
           Long.toString(inodeId));
       XMLUtils.addSaxString(contentHandler, "PATH", path);
       XMLUtils.addSaxString(contentHandler, "REPLICATION",
-          Short.toString(replication));
+          Short.valueOf(replication).toString());
       XMLUtils.addSaxString(contentHandler, "MTIME",
           Long.toString(mtime));
       XMLUtils.addSaxString(contentHandler, "ATIME",
@@ -746,7 +743,7 @@ public abstract class FSEditLogOp {
       this.length = Integer.parseInt(st.getValue("LENGTH"));
       this.inodeId = Long.parseLong(st.getValue("INODEID"));
       this.path = st.getValue("PATH");
-      this.replication = Short.parseShort(st.getValue("REPLICATION"));
+      this.replication = Short.valueOf(st.getValue("REPLICATION"));
       this.mtime = Long.parseLong(st.getValue("MTIME"));
       this.atime = Long.parseLong(st.getValue("ATIME"));
       this.blockSize = Long.parseLong(st.getValue("BLOCKSIZE"));
@@ -1187,12 +1184,12 @@ public abstract class FSEditLogOp {
     protected void toXml(ContentHandler contentHandler) throws SAXException {
       XMLUtils.addSaxString(contentHandler, "PATH", path);
       XMLUtils.addSaxString(contentHandler, "REPLICATION",
-          Short.toString(replication));
+          Short.valueOf(replication).toString());
     }
     
     @Override void fromXml(Stanza st) throws InvalidXmlException {
       this.path = st.getValue("PATH");
-      this.replication = Short.parseShort(st.getValue("REPLICATION"));
+      this.replication = Short.valueOf(st.getValue("REPLICATION"));
     }
   }
 
@@ -1980,13 +1977,13 @@ public abstract class FSEditLogOp {
     protected void toXml(ContentHandler contentHandler) throws SAXException {
       XMLUtils.addSaxString(contentHandler, "SRC", src);
       XMLUtils.addSaxString(contentHandler, "MODE",
-          Short.toString(permissions.toShort()));
+          Short.valueOf(permissions.toShort()).toString());
     }
     
     @Override void fromXml(Stanza st) throws InvalidXmlException {
       this.src = st.getValue("SRC");
       this.permissions = new FsPermission(
-          Short.parseShort(st.getValue("MODE")));
+          Short.valueOf(st.getValue("MODE")));
     }
   }
 
@@ -4470,13 +4467,13 @@ public abstract class FSEditLogOp {
     protected void toXml(ContentHandler contentHandler) throws SAXException {
       XMLUtils.addSaxString(contentHandler, "PATH", path);
       XMLUtils.addSaxString(contentHandler, "POLICYID",
-          Byte.toString(policyId));
+          Byte.valueOf(policyId).toString());
     }
 
     @Override
     void fromXml(Stanza st) throws InvalidXmlException {
       this.path = st.getValue("PATH");
-      this.policyId = Byte.parseByte(st.getValue("POLICYID"));
+      this.policyId = Byte.valueOf(st.getValue("POLICYID"));
     }
   }  
 
@@ -4953,8 +4950,7 @@ public abstract class FSEditLogOp {
 
   public static void delegationTokenToXml(ContentHandler contentHandler,
       DelegationTokenIdentifier token) throws SAXException {
-    contentHandler.startElement(
-        "", "", "DELEGATION_TOKEN_IDENTIFIER", new AttributesImpl());
+    contentHandler.startElement("", "", "DELEGATION_TOKEN_IDENTIFIER", new AttributesImpl());
     XMLUtils.addSaxString(contentHandler, "KIND", token.getKind().toString());
     XMLUtils.addSaxString(contentHandler, "SEQUENCE_NUMBER",
         Integer.toString(token.getSequenceNumber()));
@@ -5000,8 +4996,7 @@ public abstract class FSEditLogOp {
 
   public static void delegationKeyToXml(ContentHandler contentHandler,
       DelegationKey key) throws SAXException {
-    contentHandler.startElement(
-        "", "", "DELEGATION_KEY", new AttributesImpl());
+    contentHandler.startElement("", "", "DELEGATION_KEY", new AttributesImpl());
     XMLUtils.addSaxString(contentHandler, "KEY_ID",
         Integer.toString(key.getKeyId()));
     XMLUtils.addSaxString(contentHandler, "EXPIRY_DATE",
@@ -5029,8 +5024,7 @@ public abstract class FSEditLogOp {
 
   public static void permissionStatusToXml(ContentHandler contentHandler,
       PermissionStatus perm) throws SAXException {
-    contentHandler.startElement(
-        "", "", "PERMISSION_STATUS", new AttributesImpl());
+    contentHandler.startElement("", "", "PERMISSION_STATUS", new AttributesImpl());
     XMLUtils.addSaxString(contentHandler, "USERNAME", perm.getUserName());
     XMLUtils.addSaxString(contentHandler, "GROUPNAME", perm.getGroupName());
     fsPermissionToXml(contentHandler, perm.getPermission());
@@ -5048,13 +5042,13 @@ public abstract class FSEditLogOp {
 
   public static void fsPermissionToXml(ContentHandler contentHandler,
       FsPermission mode) throws SAXException {
-    XMLUtils.addSaxString(contentHandler, "MODE",
-        Short.toString(mode.toShort()));
+    XMLUtils.addSaxString(contentHandler, "MODE", Short.valueOf(mode.toShort())
+        .toString());
   }
 
   public static FsPermission fsPermissionFromXml(Stanza st)
       throws InvalidXmlException {
-    short mode = Short.parseShort(st.getValue("MODE"));
+    short mode = Short.valueOf(st.getValue("MODE"));
     return new FsPermission(mode);
   }
 
@@ -5063,8 +5057,7 @@ public abstract class FSEditLogOp {
     XMLUtils.addSaxString(contentHandler, "PERM", v.SYMBOL);
   }
 
-  private static FsAction fsActionFromXml(Stanza st)
-      throws InvalidXmlException {
+  private static FsAction fsActionFromXml(Stanza st) throws InvalidXmlException {
     FsAction v = FSACTION_SYMBOL_MAP.get(st.getValue("PERM"));
     if (v == null)
       throw new InvalidXmlException("Invalid value for FsAction");

http://git-wip-us.apache.org/repos/asf/hadoop/blob/5e565ce7/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/top/window/RollingWindowManager.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/top/window/RollingWindowManager.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/top/window/RollingWindowManager.java
index 2834ebb..63438ff 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/top/window/RollingWindowManager.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/top/window/RollingWindowManager.java
@@ -173,8 +173,7 @@ public class RollingWindowManager {
    * @param user the user that updated the metric
    * @param delta the amount of change in the metric, e.g., +1
    */
-  public void recordMetric(long time, String command,
-      String user, long delta) {
+  public void recordMetric(long time, String command, String user, long delta) {
     RollingWindow window = getRollingWindow(command, user);
     window.incAt(time, delta);
   }
@@ -209,7 +208,7 @@ public class RollingWindowManager {
       }
       for (int i = 0; i < size; i++) {
         NameValuePair userEntry = reverse.pop();
-        User user = new User(userEntry.name, userEntry.value);
+        User user = new User(userEntry.name, Long.valueOf(userEntry.value));
         op.addUser(user);
       }
     }
@@ -244,8 +243,7 @@ public class RollingWindowManager {
           metricName, userName, windowSum);
       topN.offer(new NameValuePair(userName, windowSum));
     }
-    LOG.debug("topN users size for command {} is: {}",
-        metricName, topN.size());
+    LOG.debug("topN users size for command {} is: {}", metricName, topN.size());
     return topN;
   }
 

http://git-wip-us.apache.org/repos/asf/hadoop/blob/5e565ce7/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/ErasureCodeBenchmarkThroughput.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/ErasureCodeBenchmarkThroughput.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/ErasureCodeBenchmarkThroughput.java
index 13dc997..da4b321 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/ErasureCodeBenchmarkThroughput.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/ErasureCodeBenchmarkThroughput.java
@@ -69,8 +69,7 @@ import java.util.concurrent.TimeUnit;
  * For example, if the user wants to test reading 1024MB data with 10 clients,
  * he/she should firstly generate 1024MB data with 10 (or more) clients.
  */
-public class ErasureCodeBenchmarkThroughput
-       extends Configured implements Tool {
+public class ErasureCodeBenchmarkThroughput extends Configured implements Tool {
 
   private static final int BUFFER_SIZE_MB = 128;
   private static final String DFS_TMP_DIR = System.getProperty(
@@ -115,15 +114,13 @@ public class ErasureCodeBenchmarkThroughput
       System.out.println(msg);
     }
     System.err.println("Usage: ErasureCodeBenchmarkThroughput " +
-        "<read|write|gen|clean> <size in MB> " +
-        "<ec|rep> [num clients] [stf|pos]\n" +
-        "Stateful and positional option is only available for read.");
+        "<read|write|gen|clean> <size in MB> <ec|rep> [num clients] [stf|pos]\n"
+        + "Stateful and positional option is only available for read.");
     System.exit(1);
   }
 
-  private List<Long> doBenchmark(boolean isRead, int dataSizeMB,
-      int numClients, boolean isEc, boolean statefulRead, boolean isGen)
-      throws Exception {
+  private List<Long> doBenchmark(boolean isRead, int dataSizeMB, int numClients,
+      boolean isEc, boolean statefulRead, boolean isGen) throws Exception {
     CompletionService<Long> cs = new ExecutorCompletionService<Long>(
         Executors.newFixedThreadPool(numClients));
     for (int i = 0; i < numClients; i++) {
@@ -220,7 +217,7 @@ public class ErasureCodeBenchmarkThroughput
         printUsage("Unknown operation: " + args[0]);
       }
       try {
-        dataSizeMB = Integer.parseInt(args[1]);
+        dataSizeMB = Integer.valueOf(args[1]);
         if (dataSizeMB <= 0) {
           printUsage("Invalid data size: " + dataSizeMB);
         }
@@ -236,7 +233,7 @@ public class ErasureCodeBenchmarkThroughput
     }
     if (args.length >= 4 && type != OpType.CLEAN) {
       try {
-        numClients = Integer.parseInt(args[3]);
+        numClients = Integer.valueOf(args[3]);
         if (numClients <= 0) {
           printUsage("Invalid num of clients: " + numClients);
         }

http://git-wip-us.apache.org/repos/asf/hadoop/blob/5e565ce7/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/shortcircuit/TestShortCircuitLocalRead.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/shortcircuit/TestShortCircuitLocalRead.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/shortcircuit/TestShortCircuitLocalRead.java
index a069003..e6ee7f3 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/shortcircuit/TestShortCircuitLocalRead.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/shortcircuit/TestShortCircuitLocalRead.java
@@ -72,9 +72,9 @@ import org.junit.Test;
 /**
  * Test for short circuit read functionality using {@link BlockReaderLocal}.
  * When a block is being read by a client is on the local datanode, instead of
- * using {@link DataTransferProtocol} and connect to datanode,
- * the short circuit read allows reading the file directly
- * from the files on the local file system.
+ * using {@link DataTransferProtocol} and connect to datanode, the short circuit
+ * read allows reading the file directly from the files on the local file
+ * system.
  */
 public class TestShortCircuitLocalRead {
   private static TemporarySocketDirectory sockDir;
@@ -195,8 +195,7 @@ public class TestShortCircuitLocalRead {
 
     HdfsDataInputStream stm = (HdfsDataInputStream)fs.open(name);
 
-    ByteBuffer actual =
-        ByteBuffer.allocateDirect(expected.length - readOffset);
+    ByteBuffer actual = ByteBuffer.allocateDirect(expected.length - readOffset);
 
     IOUtils.skipFully(stm, readOffset);
 
@@ -231,8 +230,7 @@ public class TestShortCircuitLocalRead {
 
   public void doTestShortCircuitReadLegacy(boolean ignoreChecksum, int size,
       int readOffset, String shortCircuitUser, String readingUser,
-      boolean legacyShortCircuitFails)
-      throws IOException, InterruptedException {
+      boolean legacyShortCircuitFails) throws IOException, InterruptedException {
     doTestShortCircuitReadImpl(ignoreChecksum, size, readOffset,
         shortCircuitUser, readingUser, legacyShortCircuitFails);
   }
@@ -249,8 +247,7 @@ public class TestShortCircuitLocalRead {
    */
   public void doTestShortCircuitReadImpl(boolean ignoreChecksum, int size,
       int readOffset, String shortCircuitUser, String readingUser,
-      boolean legacyShortCircuitFails)
-      throws IOException, InterruptedException {
+      boolean legacyShortCircuitFails) throws IOException, InterruptedException {
     Configuration conf = new Configuration();
     conf.setBoolean(HdfsClientConfigKeys.Read.ShortCircuit.KEY, true);
     conf.setBoolean(HdfsClientConfigKeys.Read.ShortCircuit.SKIP_CHECKSUM_KEY,
@@ -265,8 +262,7 @@ public class TestShortCircuitLocalRead {
     if (shortCircuitUser != null) {
       conf.set(DFSConfigKeys.DFS_BLOCK_LOCAL_PATH_ACCESS_USER_KEY,
           shortCircuitUser);
-      conf.setBoolean(
-          HdfsClientConfigKeys.DFS_CLIENT_USE_LEGACY_BLOCKREADERLOCAL, true);
+      conf.setBoolean(HdfsClientConfigKeys.DFS_CLIENT_USE_LEGACY_BLOCKREADERLOCAL, true);
     }
     if (simulatedStorage) {
       SimulatedFSDataset.setFactory(conf);
@@ -328,8 +324,7 @@ public class TestShortCircuitLocalRead {
    */
   @Test(timeout=60000)
   public void testLocalReadFallback() throws Exception {
-    doTestShortCircuitReadLegacy(
-        true, 13, 0, getCurrentUser(), "notallowed", true);
+    doTestShortCircuitReadLegacy(true, 13, 0, getCurrentUser(), "notallowed", true);
   }
 
   @Test(timeout=60000)
@@ -371,9 +366,8 @@ public class TestShortCircuitLocalRead {
       ExtendedBlock blk = new ExtendedBlock(lb.get(0).getBlock());
       Token<BlockTokenIdentifier> token = lb.get(0).getBlockToken();
       final DatanodeInfo dnInfo = lb.get(0).getLocations()[0];
-      ClientDatanodeProtocol proxy =
-          DFSUtilClient.createClientDatanodeProtocolProxy(
-              dnInfo, conf, 60000, false);
+      ClientDatanodeProtocol proxy = 
+          DFSUtilClient.createClientDatanodeProtocolProxy(dnInfo, conf, 60000, false);
       try {
         proxy.getBlockLocalPathInfo(blk, token);
         Assert.fail("The call should have failed as this user "
@@ -393,8 +387,7 @@ public class TestShortCircuitLocalRead {
     int size = blockSize;
     Configuration conf = new Configuration();
     conf.setBoolean(HdfsClientConfigKeys.Read.ShortCircuit.KEY, true);
-    conf.setBoolean(
-        HdfsClientConfigKeys.Read.ShortCircuit.SKIP_CHECKSUM_KEY, false);
+    conf.setBoolean(HdfsClientConfigKeys.Read.ShortCircuit.SKIP_CHECKSUM_KEY, false);
     conf.set(DFSConfigKeys.DFS_DOMAIN_SOCKET_PATH_KEY,
         new File(sockDir.getDir(),
             "testSkipWithVerifyChecksum._PORT.sock").getAbsolutePath());
@@ -441,8 +434,7 @@ public class TestShortCircuitLocalRead {
     MiniDFSCluster cluster = null;
     HdfsConfiguration conf = new HdfsConfiguration();
     conf.setBoolean(HdfsClientConfigKeys.Read.ShortCircuit.KEY, true);
-    conf.setBoolean(
-        HdfsClientConfigKeys.Read.ShortCircuit.SKIP_CHECKSUM_KEY, false);
+    conf.setBoolean(HdfsClientConfigKeys.Read.ShortCircuit.SKIP_CHECKSUM_KEY, false);
     conf.set(DFSConfigKeys.DFS_DOMAIN_SOCKET_PATH_KEY,
         new File(sockDir.getDir(),
             "testHandleTruncatedBlockFile._PORT.sock").getAbsolutePath());
@@ -531,8 +523,8 @@ public class TestShortCircuitLocalRead {
       System.out.println("Usage: test shortcircuit checksum threadCount");
       System.exit(1);
     }
-    boolean shortcircuit = Boolean.parseBoolean(args[0]);
-    boolean checksum = Boolean.parseBoolean(args[1]);
+    boolean shortcircuit = Boolean.valueOf(args[0]);
+    boolean checksum = Boolean.valueOf(args[1]);
     int threadCount = Integer.parseInt(args[2]);
 
     // Setup create a file
@@ -543,8 +535,7 @@ public class TestShortCircuitLocalRead {
     conf.setBoolean(HdfsClientConfigKeys.Read.ShortCircuit.SKIP_CHECKSUM_KEY,
         checksum);
 
-    // Override fileSize and DATA_TO_WRITE to
-    // much larger values for benchmark test
+    // Override fileSize and DATA_TO_WRITE to much larger values for benchmark test
     int fileSize = 1000 * blockSize + 100; // File with 1000 blocks
     final byte [] dataToWrite = AppendTestUtil.randomBytes(seed, fileSize);
 
@@ -566,8 +557,7 @@ public class TestShortCircuitLocalRead {
           for (int i = 0; i < iteration; i++) {
             try {
               String user = getCurrentUser();
-              checkFileContent(
-                  fs.getUri(), file1, dataToWrite, 0, user, conf, true);
+              checkFileContent(fs.getUri(), file1, dataToWrite, 0, user, conf, true);
             } catch (IOException e) {
               e.printStackTrace();
             } catch (InterruptedException e) {
@@ -600,13 +590,11 @@ public class TestShortCircuitLocalRead {
    * through RemoteBlockReader
    * @throws IOException
   */
-  public void doTestShortCircuitReadWithRemoteBlockReader(
-      boolean ignoreChecksum, int size, String shortCircuitUser,
-      int readOffset, boolean shortCircuitFails)
-      throws IOException, InterruptedException {
+  public void doTestShortCircuitReadWithRemoteBlockReader(boolean ignoreChecksum,
+      int size, String shortCircuitUser, int readOffset,
+      boolean shortCircuitFails) throws IOException, InterruptedException {
     Configuration conf = new Configuration();
-    conf.setBoolean(
-        HdfsClientConfigKeys.DFS_CLIENT_USE_LEGACY_BLOCKREADER, true);
+    conf.setBoolean(HdfsClientConfigKeys.DFS_CLIENT_USE_LEGACY_BLOCKREADER, true);
     conf.setBoolean(HdfsClientConfigKeys.Read.ShortCircuit.KEY, true);
 
     MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(1)
@@ -615,8 +603,7 @@ public class TestShortCircuitLocalRead {
     // check that / exists
     Path path = new Path("/");
     URI uri = cluster.getURI();
-    assertTrue(
-        "/ should be a directory", fs.getFileStatus(path).isDirectory());
+    assertTrue("/ should be a directory", fs.getFileStatus(path).isDirectory());
 
     byte[] fileData = AppendTestUtil.randomBytes(seed, size);
     Path file1 = new Path("filelocal.dat");
@@ -628,12 +615,10 @@ public class TestShortCircuitLocalRead {
       checkFileContent(uri, file1, fileData, readOffset, shortCircuitUser, 
           conf, shortCircuitFails);
       //RemoteBlockReader have unsupported method read(ByteBuffer bf)
-      assertTrue(
-          "RemoteBlockReader unsupported method read(ByteBuffer bf) error",
-          checkUnsupportedMethod(fs, file1, fileData, readOffset));
+      assertTrue("RemoteBlockReader unsupported method read(ByteBuffer bf) error",
+                    checkUnsupportedMethod(fs, file1, fileData, readOffset));
     } catch(IOException e) {
-      throw new IOException(
-          "doTestShortCircuitReadWithRemoteBlockReader ex error ", e);
+      throw new IOException("doTestShortCircuitReadWithRemoteBlockReader ex error ", e);
     } catch(InterruptedException inEx) {
       throw inEx;
     } finally {
@@ -645,8 +630,7 @@ public class TestShortCircuitLocalRead {
   private boolean checkUnsupportedMethod(FileSystem fs, Path file,
       byte[] expected, int readOffset) throws IOException {
     HdfsDataInputStream stm = (HdfsDataInputStream)fs.open(file);
-    ByteBuffer actual =
-        ByteBuffer.allocateDirect(expected.length - readOffset);
+    ByteBuffer actual = ByteBuffer.allocateDirect(expected.length - readOffset);
     IOUtils.skipFully(stm, readOffset);
     try {
       stm.read(actual);

http://git-wip-us.apache.org/repos/asf/hadoop/blob/5e565ce7/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/test/java/org/apache/hadoop/mapreduce/v2/app/TestStagingCleanup.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/test/java/org/apache/hadoop/mapreduce/v2/app/TestStagingCleanup.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/test/java/org/apache/hadoop/mapreduce/v2/app/TestStagingCleanup.java
index e22a12e..fc64996 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/test/java/org/apache/hadoop/mapreduce/v2/app/TestStagingCleanup.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/test/java/org/apache/hadoop/mapreduce/v2/app/TestStagingCleanup.java
@@ -245,63 +245,6 @@ import org.junit.Test;
      verify(fs).delete(stagingJobPath, true);
    }
 
-  @Test
-  public void testByPreserveFailedStaging() throws IOException {
-    conf.set(MRJobConfig.MAPREDUCE_JOB_DIR, stagingJobDir);
-    // Failed task's staging files should be kept
-    conf.setBoolean(MRJobConfig.PRESERVE_FAILED_TASK_FILES, true);
-    fs = mock(FileSystem.class);
-    when(fs.delete(any(Path.class), anyBoolean())).thenReturn(true);
-    //Staging Dir exists
-    String user = UserGroupInformation.getCurrentUser().getShortUserName();
-    Path stagingDir = MRApps.getStagingAreaDir(conf, user);
-    when(fs.exists(stagingDir)).thenReturn(true);
-    ApplicationId appId = ApplicationId.newInstance(System.currentTimeMillis(),
-            0);
-    ApplicationAttemptId attemptId = ApplicationAttemptId.newInstance(appId, 1);
-    JobId jobid = recordFactory.newRecordInstance(JobId.class);
-    jobid.setAppId(appId);
-    ContainerAllocator mockAlloc = mock(ContainerAllocator.class);
-    Assert.assertTrue(MRJobConfig.DEFAULT_MR_AM_MAX_ATTEMPTS > 1);
-    MRAppMaster appMaster = new TestMRApp(attemptId, mockAlloc,
-            JobStateInternal.FAILED, MRJobConfig.DEFAULT_MR_AM_MAX_ATTEMPTS);
-    appMaster.init(conf);
-    appMaster.start();
-    appMaster.shutDownJob();
-    //test whether notifyIsLastAMRetry called
-    Assert.assertEquals(true, ((TestMRApp)appMaster).getTestIsLastAMRetry());
-    verify(fs, times(0)).delete(stagingJobPath, true);
-  }
-
-  @Test
-  public void testPreservePatternMatchedStaging() throws IOException {
-    conf.set(MRJobConfig.MAPREDUCE_JOB_DIR, stagingJobDir);
-    // The staging files that are matched to the pattern
-    // should not be deleted
-    conf.set(MRJobConfig.PRESERVE_FILES_PATTERN, "JobDir");
-    fs = mock(FileSystem.class);
-    when(fs.delete(any(Path.class), anyBoolean())).thenReturn(true);
-    //Staging Dir exists
-    String user = UserGroupInformation.getCurrentUser().getShortUserName();
-    Path stagingDir = MRApps.getStagingAreaDir(conf, user);
-    when(fs.exists(stagingDir)).thenReturn(true);
-    ApplicationId appId = ApplicationId.newInstance(System.currentTimeMillis(),
-            0);
-    ApplicationAttemptId attemptId = ApplicationAttemptId.newInstance(appId, 1);
-    JobId jobid = recordFactory.newRecordInstance(JobId.class);
-    jobid.setAppId(appId);
-    ContainerAllocator mockAlloc = mock(ContainerAllocator.class);
-    Assert.assertTrue(MRJobConfig.DEFAULT_MR_AM_MAX_ATTEMPTS > 1);
-    MRAppMaster appMaster = new TestMRApp(attemptId, mockAlloc,
-            JobStateInternal.RUNNING, MRJobConfig.DEFAULT_MR_AM_MAX_ATTEMPTS);
-    appMaster.init(conf);
-    appMaster.start();
-    appMaster.shutDownJob();
-    //test whether notifyIsLastAMRetry called
-    Assert.assertEquals(true, ((TestMRApp)appMaster).getTestIsLastAMRetry());
-    verify(fs, times(0)).delete(stagingJobPath, true);
-  }
-
    private class TestMRApp extends MRAppMaster {
      ContainerAllocator allocator;
      boolean testIsLastAMRetry = false;