You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by zh...@apache.org on 2018/01/24 09:50:59 UTC

[17/34] hbase git commit: HBASE-19811 Fix findbugs and error-prone warnings in hbase-server (branch-2)

http://git-wip-us.apache.org/repos/asf/hbase/blob/93a182f2/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedAction.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedAction.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedAction.java
index 5098e0b..410dd0c 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedAction.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedAction.java
@@ -101,7 +101,7 @@ public abstract class MultiThreadedAction {
 
     @Override
     public byte[] getDeterministicUniqueKey(long keyBase) {
-      return LoadTestKVGenerator.md5PrefixedKey(keyBase).getBytes();
+      return Bytes.toBytes(LoadTestKVGenerator.md5PrefixedKey(keyBase));
     }
 
     @Override
@@ -114,7 +114,7 @@ public abstract class MultiThreadedAction {
       int numColumns = minColumnsPerKey + random.nextInt(maxColumnsPerKey - minColumnsPerKey + 1);
       byte[][] columns = new byte[numColumns][];
       for (int i = 0; i < numColumns; ++i) {
-        columns[i] = Integer.toString(i).getBytes();
+        columns[i] = Bytes.toBytes(Integer.toString(i));
       }
       return columns;
     }

http://git-wip-us.apache.org/repos/asf/hbase/blob/93a182f2/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedReader.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedReader.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedReader.java
index 447cca8..6864366 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedReader.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedReader.java
@@ -201,8 +201,7 @@ public class MultiThreadedReader extends MultiThreadedAction
                 "to read " + k + " is out of range (startKey=" + startKey +
                 ", endKey=" + endKey + ")");
           }
-          if (k % numThreads != readerId ||
-              writer != null && writer.failedToWriteKey(k)) {
+          if (k % numThreads != readerId || (writer != null && writer.failedToWriteKey(k))) {
             // Skip keys that this thread should not read, as well as the keys
             // that we know the writer failed to write.
             continue;

http://git-wip-us.apache.org/repos/asf/hbase/blob/93a182f2/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestBoundedPriorityBlockingQueue.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestBoundedPriorityBlockingQueue.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestBoundedPriorityBlockingQueue.java
index 7112d50..7746bea 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestBoundedPriorityBlockingQueue.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestBoundedPriorityBlockingQueue.java
@@ -42,7 +42,7 @@ import org.junit.experimental.categories.Category;
 public class TestBoundedPriorityBlockingQueue {
   private final static int CAPACITY = 16;
 
-  class TestObject {
+  static class TestObject {
     private final int priority;
     private final int seqId;
 
@@ -60,7 +60,7 @@ public class TestBoundedPriorityBlockingQueue {
     }
   }
 
-  class TestObjectComparator implements Comparator<TestObject> {
+  static class TestObjectComparator implements Comparator<TestObject> {
     public TestObjectComparator() {}
 
     @Override
@@ -208,6 +208,7 @@ public class TestBoundedPriorityBlockingQueue {
     final CyclicBarrier threadsStarted = new CyclicBarrier(2);
     ExecutorService executor = Executors.newFixedThreadPool(2);
     executor.execute(new Runnable() {
+      @Override
       public void run() {
         try {
           assertNull(queue.poll(1000, TimeUnit.MILLISECONDS));
@@ -221,6 +222,7 @@ public class TestBoundedPriorityBlockingQueue {
     });
 
     executor.execute(new Runnable() {
+      @Override
       public void run() {
         try {
             threadsStarted.await();

http://git-wip-us.apache.org/repos/asf/hbase/blob/93a182f2/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestByteBuffUtils.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestByteBuffUtils.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestByteBuffUtils.java
index 4c6990e..ecc6611 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestByteBuffUtils.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestByteBuffUtils.java
@@ -36,7 +36,7 @@ public class TestByteBuffUtils {
     ByteBuffer bb2 = ByteBuffer.allocate(50);
     MultiByteBuff src = new MultiByteBuff(bb1, bb2);
     for (int i = 0; i < 7; i++) {
-      src.putLong(8l);
+      src.putLong(8L);
     }
     src.put((byte) 1);
     src.put((byte) 1);
@@ -58,7 +58,7 @@ public class TestByteBuffUtils {
     bb3 = ByteBuffer.allocate(100);
     SingleByteBuff sbb = new SingleByteBuff(bb3);
     for (int i = 0; i < 7; i++) {
-      sbb.putLong(8l);
+      sbb.putLong(8L);
     }
     sbb.put((byte) 1);
     sbb.put((byte) 1);

http://git-wip-us.apache.org/repos/asf/hbase/blob/93a182f2/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSHDFSUtils.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSHDFSUtils.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSHDFSUtils.java
index 333d050..22f34de 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSHDFSUtils.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSHDFSUtils.java
@@ -161,11 +161,12 @@ public class TestFSHDFSUtils {
   /**
    * Version of DFS that has HDFS-4525 in it.
    */
-  class IsFileClosedDistributedFileSystem extends DistributedFileSystem {
+  static class IsFileClosedDistributedFileSystem extends DistributedFileSystem {
     /**
      * Close status of a file. Copied over from HDFS-4525
      * @return true if file is already closed
      **/
+    @Override
     public boolean isFileClosed(Path f) throws IOException{
       return false;
     }

http://git-wip-us.apache.org/repos/asf/hbase/blob/93a182f2/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSVisitor.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSVisitor.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSVisitor.java
index cb23a0b..be302d5 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSVisitor.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSVisitor.java
@@ -77,6 +77,7 @@ public class TestFSVisitor {
     final Set<String> families = new HashSet<>();
     final Set<String> hfiles = new HashSet<>();
     FSVisitor.visitTableStoreFiles(fs, tableDir, new FSVisitor.StoreFileVisitor() {
+      @Override
       public void storeFile(final String region, final String family, final String hfileName)
           throws IOException {
         regions.add(region);
@@ -84,9 +85,9 @@ public class TestFSVisitor {
         hfiles.add(hfileName);
       }
     });
-    assertEquals(tableRegions, regions);
-    assertEquals(tableFamilies, families);
-    assertEquals(tableHFiles, hfiles);
+    assertEquals(regions, tableRegions);
+    assertEquals(families, tableFamilies);
+    assertEquals(hfiles, tableHFiles);
   }
 
   /*

http://git-wip-us.apache.org/repos/asf/hbase/blob/93a182f2/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckEncryption.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckEncryption.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckEncryption.java
index 641f66e..31f1909 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckEncryption.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckEncryption.java
@@ -131,12 +131,12 @@ public class TestHBaseFsckEncryption {
 
     // Insure HBck doesn't consider them corrupt
     HBaseFsck res = HbckTestingUtil.doHFileQuarantine(conf, htd.getTableName());
-    assertEquals(res.getRetCode(), 0);
+    assertEquals(0, res.getRetCode());
     HFileCorruptionChecker hfcc = res.getHFilecorruptionChecker();
-    assertEquals(hfcc.getCorrupted().size(), 0);
-    assertEquals(hfcc.getFailures().size(), 0);
-    assertEquals(hfcc.getQuarantined().size(), 0);
-    assertEquals(hfcc.getMissing().size(), 0);
+    assertEquals(0, hfcc.getCorrupted().size());
+    assertEquals(0, hfcc.getFailures().size());
+    assertEquals(0, hfcc.getQuarantined().size());
+    assertEquals(0, hfcc.getMissing().size());
   }
 
   private List<Path> findStorefilePaths(TableName tableName) throws Exception {

http://git-wip-us.apache.org/repos/asf/hbase/blob/93a182f2/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckMOB.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckMOB.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckMOB.java
index 3661207..2548ecf 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckMOB.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckMOB.java
@@ -118,18 +118,18 @@ public class TestHBaseFsckMOB extends BaseTestHBaseFsck {
       // A corrupt mob file doesn't abort the start of regions, so we can enable the table.
       admin.enableTable(table);
       HBaseFsck res = HbckTestingUtil.doHFileQuarantine(conf, table);
-      assertEquals(res.getRetCode(), 0);
+      assertEquals(0, res.getRetCode());
       HFileCorruptionChecker hfcc = res.getHFilecorruptionChecker();
-      assertEquals(hfcc.getHFilesChecked(), 4);
-      assertEquals(hfcc.getCorrupted().size(), 0);
-      assertEquals(hfcc.getFailures().size(), 0);
-      assertEquals(hfcc.getQuarantined().size(), 0);
-      assertEquals(hfcc.getMissing().size(), 0);
-      assertEquals(hfcc.getMobFilesChecked(), 5);
-      assertEquals(hfcc.getCorruptedMobFiles().size(), 1);
-      assertEquals(hfcc.getFailureMobFiles().size(), 0);
-      assertEquals(hfcc.getQuarantinedMobFiles().size(), 1);
-      assertEquals(hfcc.getMissedMobFiles().size(), 0);
+      assertEquals(4, hfcc.getHFilesChecked());
+      assertEquals(0, hfcc.getCorrupted().size());
+      assertEquals(0, hfcc.getFailures().size());
+      assertEquals(0, hfcc.getQuarantined().size());
+      assertEquals(0, hfcc.getMissing().size());
+      assertEquals(5, hfcc.getMobFilesChecked());
+      assertEquals(1, hfcc.getCorruptedMobFiles().size());
+      assertEquals(0, hfcc.getFailureMobFiles().size());
+      assertEquals(1, hfcc.getQuarantinedMobFiles().size());
+      assertEquals(0, hfcc.getMissedMobFiles().size());
       String quarantinedMobFile = hfcc.getQuarantinedMobFiles().iterator().next().getName();
       assertEquals(corruptMobFile, quarantinedMobFile);
     } finally {

http://git-wip-us.apache.org/repos/asf/hbase/blob/93a182f2/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestJSONMetricUtil.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestJSONMetricUtil.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestJSONMetricUtil.java
index 13c6df5..bb68898 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestJSONMetricUtil.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestJSONMetricUtil.java
@@ -86,7 +86,7 @@ public class TestJSONMetricUtil {
     Hashtable<String, String> properties = JSONMetricUtil.buldKeyValueTable(keys, values);
     ObjectName testObject = JSONMetricUtil.buildObjectName(JSONMetricUtil.JAVA_LANG_DOMAIN,
       properties);
-    assertEquals(testObject.getDomain(), JSONMetricUtil.JAVA_LANG_DOMAIN);
+    assertEquals(JSONMetricUtil.JAVA_LANG_DOMAIN, testObject.getDomain());
     assertEquals(testObject.getKeyPropertyList(), properties);
   }
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/93a182f2/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMiniClusterLoadParallel.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMiniClusterLoadParallel.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMiniClusterLoadParallel.java
index 7b1cd2d..e6b05e9 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMiniClusterLoadParallel.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMiniClusterLoadParallel.java
@@ -40,6 +40,7 @@ public class TestMiniClusterLoadParallel
     super(isMultiPut, encoding);
   }
 
+  @Override
   @Test(timeout=TIMEOUT_MS)
   public void loadTest() throws Exception {
     prepareForLoadTest();

http://git-wip-us.apache.org/repos/asf/hbase/blob/93a182f2/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestRegionSplitCalculator.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestRegionSplitCalculator.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestRegionSplitCalculator.java
index fd86beb..eee3030 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestRegionSplitCalculator.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestRegionSplitCalculator.java
@@ -68,6 +68,7 @@ public class TestRegionSplitCalculator {
       return end;
     }
 
+    @Override
     public String toString() {
       return "[" + Bytes.toString(start) + ", " + Bytes.toString(end) + "]";
     }
@@ -135,8 +136,7 @@ public class TestRegionSplitCalculator {
     LOG.info("Standard");
     String res = dump(sc.getSplits(), regions);
     checkDepths(sc.getSplits(), regions, 1, 1, 1, 0);
-    assertEquals(res, "A:\t[A, B]\t\n" + "B:\t[B, C]\t\n" + "C:\t[C, D]\t\n"
-        + "D:\t\n");
+    assertEquals("A:\t[A, B]\t\n" + "B:\t[B, C]\t\n" + "C:\t[C, D]\t\nD:\t\n", res);
   }
 
   @Test

http://git-wip-us.apache.org/repos/asf/hbase/blob/93a182f2/hbase-server/src/test/java/org/apache/hadoop/hbase/util/test/LoadTestDataGenerator.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/test/LoadTestDataGenerator.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/test/LoadTestDataGenerator.java
index 3650706..1aab1f8 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/test/LoadTestDataGenerator.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/test/LoadTestDataGenerator.java
@@ -20,10 +20,11 @@ import java.io.IOException;
 import java.util.Random;
 import java.util.Set;
 
-import org.apache.yetus.audience.InterfaceAudience;
+import org.apache.hadoop.hbase.util.Bytes;
 import org.apache.hadoop.hbase.client.Get;
 import org.apache.hadoop.hbase.client.Mutation;
 import org.apache.hadoop.hbase.util.LoadTestKVGenerator;
+import org.apache.yetus.audience.InterfaceAudience;
 
 /**
  * A generator of random data (keys/cfs/columns/values) for load testing.
@@ -35,11 +36,11 @@ public abstract class LoadTestDataGenerator {
 
   // The mutate info column stores information
   // about update done to this column family this row.
-  public final static byte[] MUTATE_INFO = "mutate_info".getBytes();
+  public final static byte[] MUTATE_INFO = Bytes.toBytes("mutate_info");
 
   // The increment column always has a long value,
   // which can be incremented later on during updates.
-  public final static byte[] INCREMENT = "increment".getBytes();
+  public final static byte[] INCREMENT = Bytes.toBytes("increment");
 
   protected String[] args;
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/93a182f2/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALFactory.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALFactory.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALFactory.java
index f0f35e7..ceb43d5 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALFactory.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALFactory.java
@@ -476,7 +476,7 @@ public class TestWALFactory {
     reader.close();
 
     // Reset the lease period
-    setLeasePeriod.invoke(cluster, new Object[]{new Long(60000), new Long(3600000)});
+    setLeasePeriod.invoke(cluster, new Object[]{ 60000L, 3600000L });
   }
 
   /**

http://git-wip-us.apache.org/repos/asf/hbase/blob/93a182f2/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplit.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplit.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplit.java
index a1206aa..dded506 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplit.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplit.java
@@ -131,9 +131,9 @@ public class TestWALSplit {
   private static final String FILENAME_BEING_SPLIT = "testfile";
   private static final TableName TABLE_NAME =
       TableName.valueOf("t1");
-  private static final byte[] FAMILY = "f1".getBytes();
-  private static final byte[] QUALIFIER = "q1".getBytes();
-  private static final byte[] VALUE = "v1".getBytes();
+  private static final byte[] FAMILY = Bytes.toBytes("f1");
+  private static final byte[] QUALIFIER = Bytes.toBytes("q1");
+  private static final byte[] VALUE = Bytes.toBytes("v1");
   private static final String WAL_FILE_PREFIX = "wal.dat.";
   private static List<String> REGIONS = new ArrayList<>();
   private static final String HBASE_SKIP_ERRORS = "hbase.hlog.split.skip.errors";
@@ -341,7 +341,7 @@ public class TestWALSplit {
       while (!stop.get()) {
         try {
           long seq = appendEntry(writer, TABLE_NAME, regionBytes,
-              ("r" + editsCount.get()).getBytes(), regionBytes, QUALIFIER, VALUE, 0);
+              Bytes.toBytes("r" + editsCount.get()), regionBytes, QUALIFIER, VALUE, 0);
           long count = editsCount.incrementAndGet();
           LOG.info(getName() + " sync count=" + count + ", seq=" + seq);
           try {
@@ -407,7 +407,7 @@ public class TestWALSplit {
             TableName.META_TABLE_NAME, 1, now, HConstants.DEFAULT_CLUSTER_ID),
             new WALEdit());
     Path parent = WALSplitter.getRegionDirRecoveredEditsDir(regiondir);
-    assertEquals(parent.getName(), HConstants.RECOVERED_EDITS_DIR);
+    assertEquals(HConstants.RECOVERED_EDITS_DIR, parent.getName());
     fs.createNewFile(parent); // create a recovered.edits file
 
     Path p = WALSplitter.getRegionSplitEditsPath(fs, entry, HBASEDIR,
@@ -588,8 +588,8 @@ public class TestWALSplit {
         archivedLogs.add(log.getPath().getName());
       }
       LOG.debug(archived.toString());
-      assertEquals(failureType.name() + ": expected to find all of our wals corrupt.",
-        walDirContents, archivedLogs);
+      assertEquals(failureType.name() + ": expected to find all of our wals corrupt.", archivedLogs,
+          walDirContents);
     }
   }
 
@@ -687,7 +687,7 @@ public class TestWALSplit {
 
     // should not have stored the EOF files as corrupt
     FileStatus[] archivedLogs = fs.listStatus(CORRUPTDIR);
-    assertEquals(archivedLogs.length, 0);
+    assertEquals(0, archivedLogs.length);
 
   }
 
@@ -749,7 +749,7 @@ public class TestWALSplit {
 
     InstrumentedLogWriter.activateFailure = false;
     appendEntry(writer, TABLE_NAME, Bytes.toBytes(region),
-        ("r" + 999).getBytes(), FAMILY, QUALIFIER, VALUE, 0);
+        Bytes.toBytes("r" + 999), FAMILY, QUALIFIER, VALUE, 0);
     writer.close();
 
     try {
@@ -1206,8 +1206,8 @@ public class TestWALSplit {
         int prefix = 0;
         for (String region : REGIONS) {
           String row_key = region + prefix++ + i + j;
-          appendEntry(ws[i], TABLE_NAME, region.getBytes(), row_key.getBytes(), FAMILY, QUALIFIER,
-              VALUE, seq++);
+          appendEntry(ws[i], TABLE_NAME, Bytes.toBytes(region), Bytes.toBytes(row_key), FAMILY,
+              QUALIFIER, VALUE, seq++);
 
           if (numRegionEventsAdded < regionEvents) {
             numRegionEventsAdded ++;
@@ -1233,7 +1233,7 @@ public class TestWALSplit {
     Path tdir = FSUtils.getTableDir(rootdir, table);
     @SuppressWarnings("deprecation")
     Path editsdir = WALSplitter.getRegionDirRecoveredEditsDir(HRegion.getRegionDir(tdir,
-        Bytes.toString(region.getBytes())));
+        Bytes.toString(Bytes.toBytes(region))));
     FileStatus[] files = fs.listStatus(editsdir, new PathFilter() {
       @Override
       public boolean accept(Path p) {
@@ -1260,46 +1260,46 @@ public class TestWALSplit {
     in.close();
 
     switch (corruption) {
-    case APPEND_GARBAGE:
-      fs.delete(path, false);
-      out = fs.create(path);
-      out.write(corrupted_bytes);
-      out.write("-----".getBytes());
-      closeOrFlush(close, out);
-      break;
-
-    case INSERT_GARBAGE_ON_FIRST_LINE:
-      fs.delete(path, false);
-      out = fs.create(path);
-      out.write(0);
-      out.write(corrupted_bytes);
-      closeOrFlush(close, out);
-      break;
-
-    case INSERT_GARBAGE_IN_THE_MIDDLE:
-      fs.delete(path, false);
-      out = fs.create(path);
-      int middle = (int) Math.floor(corrupted_bytes.length / 2);
-      out.write(corrupted_bytes, 0, middle);
-      out.write(0);
-      out.write(corrupted_bytes, middle, corrupted_bytes.length - middle);
-      closeOrFlush(close, out);
-      break;
-
-    case TRUNCATE:
-      fs.delete(path, false);
-      out = fs.create(path);
-      out.write(corrupted_bytes, 0, fileSize
-          - (32 + ProtobufLogReader.PB_WAL_COMPLETE_MAGIC.length + Bytes.SIZEOF_INT));
-      closeOrFlush(close, out);
-      break;
-
-    case TRUNCATE_TRAILER:
-      fs.delete(path, false);
-      out = fs.create(path);
-      out.write(corrupted_bytes, 0, fileSize - Bytes.SIZEOF_INT);// trailer is truncated.
-      closeOrFlush(close, out);
-      break;
+      case APPEND_GARBAGE:
+        fs.delete(path, false);
+        out = fs.create(path);
+        out.write(corrupted_bytes);
+        out.write(Bytes.toBytes("-----"));
+        closeOrFlush(close, out);
+        break;
+
+      case INSERT_GARBAGE_ON_FIRST_LINE:
+        fs.delete(path, false);
+        out = fs.create(path);
+        out.write(0);
+        out.write(corrupted_bytes);
+        closeOrFlush(close, out);
+        break;
+
+      case INSERT_GARBAGE_IN_THE_MIDDLE:
+        fs.delete(path, false);
+        out = fs.create(path);
+        int middle = (int) Math.floor(corrupted_bytes.length / 2);
+        out.write(corrupted_bytes, 0, middle);
+        out.write(0);
+        out.write(corrupted_bytes, middle, corrupted_bytes.length - middle);
+        closeOrFlush(close, out);
+        break;
+
+      case TRUNCATE:
+        fs.delete(path, false);
+        out = fs.create(path);
+        out.write(corrupted_bytes, 0, fileSize
+            - (32 + ProtobufLogReader.PB_WAL_COMPLETE_MAGIC.length + Bytes.SIZEOF_INT));
+        closeOrFlush(close, out);
+        break;
+
+      case TRUNCATE_TRAILER:
+        fs.delete(path, false);
+        out = fs.create(path);
+        out.write(corrupted_bytes, 0, fileSize - Bytes.SIZEOF_INT);// trailer is truncated.
+        closeOrFlush(close, out);
+        break;
     }
   }
 
@@ -1360,14 +1360,14 @@ public class TestWALSplit {
     WALProtos.RegionEventDescriptor regionOpenDesc = ProtobufUtil.toRegionEventDescriptor(
         WALProtos.RegionEventDescriptor.EventType.REGION_OPEN,
         TABLE_NAME.toBytes(),
-        region.getBytes(),
-        String.valueOf(region.hashCode()).getBytes(),
+        Bytes.toBytes(region),
+        Bytes.toBytes(String.valueOf(region.hashCode())),
         1,
         ServerName.parseServerName("ServerName:9099"), ImmutableMap.<byte[], List<Path>>of());
     final long time = EnvironmentEdgeManager.currentTime();
-    KeyValue kv = new KeyValue(region.getBytes(), WALEdit.METAFAMILY, WALEdit.REGION_EVENT,
+    KeyValue kv = new KeyValue(Bytes.toBytes(region), WALEdit.METAFAMILY, WALEdit.REGION_EVENT,
         time, regionOpenDesc.toByteArray());
-    final WALKeyImpl walKey = new WALKeyImpl(region.getBytes(), TABLE_NAME, 1, time,
+    final WALKeyImpl walKey = new WALKeyImpl(Bytes.toBytes(region), TABLE_NAME, 1, time,
         HConstants.DEFAULT_CLUSTER_ID);
     w.append(
         new Entry(walKey, new WALEdit().add(kv)));

http://git-wip-us.apache.org/repos/asf/hbase/blob/93a182f2/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplitBoundedLogWriterCreation.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplitBoundedLogWriterCreation.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplitBoundedLogWriterCreation.java
index 844cb3a..400d12b 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplitBoundedLogWriterCreation.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplitBoundedLogWriterCreation.java
@@ -35,6 +35,7 @@ public class TestWALSplitBoundedLogWriterCreation extends TestWALSplit{
   /**
    * The logic of this test has conflict with the limit writers split logic, skip this test
    */
+  @Override
   @Test(timeout=300000)
   @Ignore
   public void testThreadingSlowWriterSmallBuffer() throws Exception {

http://git-wip-us.apache.org/repos/asf/hbase/blob/93a182f2/hbase-server/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZooKeeperACL.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZooKeeperACL.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZooKeeperACL.java
index 4e67b91..3e9e650 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZooKeeperACL.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZooKeeperACL.java
@@ -129,10 +129,10 @@ public class TestZooKeeperACL {
 
     List<ACL> acls = zkw.getRecoverableZooKeeper().getZooKeeper()
         .getACL("/hbase", new Stat());
-    assertEquals(acls.size(),1);
-    assertEquals(acls.get(0).getId().getScheme(),"sasl");
-    assertEquals(acls.get(0).getId().getId(),"hbase");
-    assertEquals(acls.get(0).getPerms(), ZooDefs.Perms.ALL);
+    assertEquals(1, acls.size());
+    assertEquals("sasl", acls.get(0).getId().getScheme());
+    assertEquals("hbase", acls.get(0).getId().getId());
+    assertEquals(ZooDefs.Perms.ALL, acls.get(0).getPerms());
   }
 
   /**
@@ -148,20 +148,20 @@ public class TestZooKeeperACL {
 
     List<ACL> acls = zkw.getRecoverableZooKeeper().getZooKeeper()
         .getACL("/hbase/root-region-server", new Stat());
-    assertEquals(acls.size(),2);
+    assertEquals(2, acls.size());
 
     boolean foundWorldReadableAcl = false;
     boolean foundHBaseOwnerAcl = false;
     for(int i = 0; i < 2; i++) {
       if (acls.get(i).getId().getScheme().equals("world") == true) {
-        assertEquals(acls.get(0).getId().getId(),"anyone");
-        assertEquals(acls.get(0).getPerms(), ZooDefs.Perms.READ);
+        assertEquals("anyone", acls.get(0).getId().getId());
+        assertEquals(ZooDefs.Perms.READ, acls.get(0).getPerms());
         foundWorldReadableAcl = true;
       }
       else {
         if (acls.get(i).getId().getScheme().equals("sasl") == true) {
-          assertEquals(acls.get(1).getId().getId(),"hbase");
-          assertEquals(acls.get(1).getId().getScheme(),"sasl");
+          assertEquals("hbase", acls.get(1).getId().getId());
+          assertEquals("sasl", acls.get(1).getId().getScheme());
           foundHBaseOwnerAcl = true;
         } else { // error: should not get here: test fails.
           assertTrue(false);
@@ -185,19 +185,19 @@ public class TestZooKeeperACL {
 
     List<ACL> acls = zkw.getRecoverableZooKeeper().getZooKeeper()
         .getACL("/hbase/master", new Stat());
-    assertEquals(acls.size(),2);
+    assertEquals(2, acls.size());
 
     boolean foundWorldReadableAcl = false;
     boolean foundHBaseOwnerAcl = false;
     for(int i = 0; i < 2; i++) {
       if (acls.get(i).getId().getScheme().equals("world") == true) {
-        assertEquals(acls.get(0).getId().getId(),"anyone");
-        assertEquals(acls.get(0).getPerms(), ZooDefs.Perms.READ);
+        assertEquals("anyone", acls.get(0).getId().getId());
+        assertEquals(ZooDefs.Perms.READ, acls.get(0).getPerms());
         foundWorldReadableAcl = true;
       } else {
         if (acls.get(i).getId().getScheme().equals("sasl") == true) {
-          assertEquals(acls.get(1).getId().getId(),"hbase");
-          assertEquals(acls.get(1).getId().getScheme(),"sasl");
+          assertEquals("hbase", acls.get(1).getId().getId());
+          assertEquals("sasl", acls.get(1).getId().getScheme());
           foundHBaseOwnerAcl = true;
         } else { // error: should not get here: test fails.
           assertTrue(false);
@@ -221,19 +221,19 @@ public class TestZooKeeperACL {
 
     List<ACL> acls = zkw.getRecoverableZooKeeper().getZooKeeper()
         .getACL("/hbase/hbaseid", new Stat());
-    assertEquals(acls.size(),2);
+    assertEquals(2, acls.size());
 
     boolean foundWorldReadableAcl = false;
     boolean foundHBaseOwnerAcl = false;
     for(int i = 0; i < 2; i++) {
       if (acls.get(i).getId().getScheme().equals("world") == true) {
-        assertEquals(acls.get(0).getId().getId(),"anyone");
-        assertEquals(acls.get(0).getPerms(), ZooDefs.Perms.READ);
+        assertEquals("anyone", acls.get(0).getId().getId());
+        assertEquals(ZooDefs.Perms.READ, acls.get(0).getPerms());
         foundWorldReadableAcl = true;
       } else {
         if (acls.get(i).getId().getScheme().equals("sasl") == true) {
-          assertEquals(acls.get(1).getId().getId(),"hbase");
-          assertEquals(acls.get(1).getId().getScheme(),"sasl");
+          assertEquals("hbase", acls.get(1).getId().getId());
+          assertEquals("sasl", acls.get(1).getId().getScheme());
           foundHBaseOwnerAcl = true;
         } else { // error: should not get here: test fails.
           assertTrue(false);
@@ -257,10 +257,10 @@ public class TestZooKeeperACL {
     ZKUtil.createWithParents(zkw, "/testACLNode");
     List<ACL> acls = zkw.getRecoverableZooKeeper().getZooKeeper()
         .getACL("/testACLNode", new Stat());
-    assertEquals(acls.size(),1);
-    assertEquals(acls.get(0).getId().getScheme(),"sasl");
-    assertEquals(acls.get(0).getId().getId(),"hbase");
-    assertEquals(acls.get(0).getPerms(), ZooDefs.Perms.ALL);
+    assertEquals(1, acls.size());
+    assertEquals("sasl", acls.get(0).getId().getScheme());
+    assertEquals("hbase", acls.get(0).getId().getId());
+    assertEquals(ZooDefs.Perms.ALL, acls.get(0).getPerms());
   }
 
   /**
@@ -281,7 +281,7 @@ public class TestZooKeeperACL {
         saslConfFile.getAbsolutePath());
 
     testJaasConfig = ZKUtil.isSecureZooKeeper(new Configuration(TEST_UTIL.getConfiguration()));
-    assertEquals(testJaasConfig, false);
+    assertEquals(false, testJaasConfig);
     saslConfFile.delete();
   }
 
@@ -295,13 +295,13 @@ public class TestZooKeeperACL {
 
     Configuration config = new Configuration(HBaseConfiguration.create());
     boolean testJaasConfig = ZKUtil.isSecureZooKeeper(config);
-    assertEquals(testJaasConfig, false);
+    assertEquals(false, testJaasConfig);
 
     // Now set authentication scheme to Kerberos still it should return false
     // because no configuration set
     config.set("hbase.security.authentication", "kerberos");
     testJaasConfig = ZKUtil.isSecureZooKeeper(config);
-    assertEquals(testJaasConfig, false);
+    assertEquals(false, testJaasConfig);
 
     // Now set programmatic options related to security
     config.set(HConstants.ZK_CLIENT_KEYTAB_FILE, "/dummy/file");