You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by zh...@apache.org on 2018/01/23 10:24:39 UTC
[04/50] [abbrv] hbase git commit: HBASE-19811 Fix findbugs and
error-prone warnings in hbase-server (branch-2)
http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedAction.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedAction.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedAction.java
index 5098e0b..410dd0c 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedAction.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedAction.java
@@ -101,7 +101,7 @@ public abstract class MultiThreadedAction {
@Override
public byte[] getDeterministicUniqueKey(long keyBase) {
- return LoadTestKVGenerator.md5PrefixedKey(keyBase).getBytes();
+ return Bytes.toBytes(LoadTestKVGenerator.md5PrefixedKey(keyBase));
}
@Override
@@ -114,7 +114,7 @@ public abstract class MultiThreadedAction {
int numColumns = minColumnsPerKey + random.nextInt(maxColumnsPerKey - minColumnsPerKey + 1);
byte[][] columns = new byte[numColumns][];
for (int i = 0; i < numColumns; ++i) {
- columns[i] = Integer.toString(i).getBytes();
+ columns[i] = Bytes.toBytes(Integer.toString(i));
}
return columns;
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedReader.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedReader.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedReader.java
index 447cca8..6864366 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedReader.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedReader.java
@@ -201,8 +201,7 @@ public class MultiThreadedReader extends MultiThreadedAction
"to read " + k + " is out of range (startKey=" + startKey +
", endKey=" + endKey + ")");
}
- if (k % numThreads != readerId ||
- writer != null && writer.failedToWriteKey(k)) {
+ if (k % numThreads != readerId || (writer != null && writer.failedToWriteKey(k))) {
// Skip keys that this thread should not read, as well as the keys
// that we know the writer failed to write.
continue;
http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestBoundedPriorityBlockingQueue.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestBoundedPriorityBlockingQueue.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestBoundedPriorityBlockingQueue.java
index 7112d50..7746bea 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestBoundedPriorityBlockingQueue.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestBoundedPriorityBlockingQueue.java
@@ -42,7 +42,7 @@ import org.junit.experimental.categories.Category;
public class TestBoundedPriorityBlockingQueue {
private final static int CAPACITY = 16;
- class TestObject {
+ static class TestObject {
private final int priority;
private final int seqId;
@@ -60,7 +60,7 @@ public class TestBoundedPriorityBlockingQueue {
}
}
- class TestObjectComparator implements Comparator<TestObject> {
+ static class TestObjectComparator implements Comparator<TestObject> {
public TestObjectComparator() {}
@Override
@@ -208,6 +208,7 @@ public class TestBoundedPriorityBlockingQueue {
final CyclicBarrier threadsStarted = new CyclicBarrier(2);
ExecutorService executor = Executors.newFixedThreadPool(2);
executor.execute(new Runnable() {
+ @Override
public void run() {
try {
assertNull(queue.poll(1000, TimeUnit.MILLISECONDS));
@@ -221,6 +222,7 @@ public class TestBoundedPriorityBlockingQueue {
});
executor.execute(new Runnable() {
+ @Override
public void run() {
try {
threadsStarted.await();
http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestByteBuffUtils.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestByteBuffUtils.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestByteBuffUtils.java
index 4c6990e..ecc6611 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestByteBuffUtils.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestByteBuffUtils.java
@@ -36,7 +36,7 @@ public class TestByteBuffUtils {
ByteBuffer bb2 = ByteBuffer.allocate(50);
MultiByteBuff src = new MultiByteBuff(bb1, bb2);
for (int i = 0; i < 7; i++) {
- src.putLong(8l);
+ src.putLong(8L);
}
src.put((byte) 1);
src.put((byte) 1);
@@ -58,7 +58,7 @@ public class TestByteBuffUtils {
bb3 = ByteBuffer.allocate(100);
SingleByteBuff sbb = new SingleByteBuff(bb3);
for (int i = 0; i < 7; i++) {
- sbb.putLong(8l);
+ sbb.putLong(8L);
}
sbb.put((byte) 1);
sbb.put((byte) 1);
http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSHDFSUtils.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSHDFSUtils.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSHDFSUtils.java
index 8f503e0..a554e99 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSHDFSUtils.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSHDFSUtils.java
@@ -161,11 +161,12 @@ public class TestFSHDFSUtils {
/**
* Version of DFS that has HDFS-4525 in it.
*/
- class IsFileClosedDistributedFileSystem extends DistributedFileSystem {
+ static class IsFileClosedDistributedFileSystem extends DistributedFileSystem {
/**
* Close status of a file. Copied over from HDFS-4525
* @return true if file is already closed
**/
+ @Override
public boolean isFileClosed(Path f) throws IOException{
return false;
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSVisitor.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSVisitor.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSVisitor.java
index cb23a0b..be302d5 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSVisitor.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSVisitor.java
@@ -77,6 +77,7 @@ public class TestFSVisitor {
final Set<String> families = new HashSet<>();
final Set<String> hfiles = new HashSet<>();
FSVisitor.visitTableStoreFiles(fs, tableDir, new FSVisitor.StoreFileVisitor() {
+ @Override
public void storeFile(final String region, final String family, final String hfileName)
throws IOException {
regions.add(region);
@@ -84,9 +85,9 @@ public class TestFSVisitor {
hfiles.add(hfileName);
}
});
- assertEquals(tableRegions, regions);
- assertEquals(tableFamilies, families);
- assertEquals(tableHFiles, hfiles);
+ assertEquals(regions, tableRegions);
+ assertEquals(families, tableFamilies);
+ assertEquals(hfiles, tableHFiles);
}
/*
http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckEncryption.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckEncryption.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckEncryption.java
index 641f66e..31f1909 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckEncryption.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckEncryption.java
@@ -131,12 +131,12 @@ public class TestHBaseFsckEncryption {
// Insure HBck doesn't consider them corrupt
HBaseFsck res = HbckTestingUtil.doHFileQuarantine(conf, htd.getTableName());
- assertEquals(res.getRetCode(), 0);
+ assertEquals(0, res.getRetCode());
HFileCorruptionChecker hfcc = res.getHFilecorruptionChecker();
- assertEquals(hfcc.getCorrupted().size(), 0);
- assertEquals(hfcc.getFailures().size(), 0);
- assertEquals(hfcc.getQuarantined().size(), 0);
- assertEquals(hfcc.getMissing().size(), 0);
+ assertEquals(0, hfcc.getCorrupted().size());
+ assertEquals(0, hfcc.getFailures().size());
+ assertEquals(0, hfcc.getQuarantined().size());
+ assertEquals(0, hfcc.getMissing().size());
}
private List<Path> findStorefilePaths(TableName tableName) throws Exception {
http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckMOB.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckMOB.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckMOB.java
index 3661207..2548ecf 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckMOB.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckMOB.java
@@ -118,18 +118,18 @@ public class TestHBaseFsckMOB extends BaseTestHBaseFsck {
// A corrupt mob file doesn't abort the start of regions, so we can enable the table.
admin.enableTable(table);
HBaseFsck res = HbckTestingUtil.doHFileQuarantine(conf, table);
- assertEquals(res.getRetCode(), 0);
+ assertEquals(0, res.getRetCode());
HFileCorruptionChecker hfcc = res.getHFilecorruptionChecker();
- assertEquals(hfcc.getHFilesChecked(), 4);
- assertEquals(hfcc.getCorrupted().size(), 0);
- assertEquals(hfcc.getFailures().size(), 0);
- assertEquals(hfcc.getQuarantined().size(), 0);
- assertEquals(hfcc.getMissing().size(), 0);
- assertEquals(hfcc.getMobFilesChecked(), 5);
- assertEquals(hfcc.getCorruptedMobFiles().size(), 1);
- assertEquals(hfcc.getFailureMobFiles().size(), 0);
- assertEquals(hfcc.getQuarantinedMobFiles().size(), 1);
- assertEquals(hfcc.getMissedMobFiles().size(), 0);
+ assertEquals(4, hfcc.getHFilesChecked());
+ assertEquals(0, hfcc.getCorrupted().size());
+ assertEquals(0, hfcc.getFailures().size());
+ assertEquals(0, hfcc.getQuarantined().size());
+ assertEquals(0, hfcc.getMissing().size());
+ assertEquals(5, hfcc.getMobFilesChecked());
+ assertEquals(1, hfcc.getCorruptedMobFiles().size());
+ assertEquals(0, hfcc.getFailureMobFiles().size());
+ assertEquals(1, hfcc.getQuarantinedMobFiles().size());
+ assertEquals(0, hfcc.getMissedMobFiles().size());
String quarantinedMobFile = hfcc.getQuarantinedMobFiles().iterator().next().getName();
assertEquals(corruptMobFile, quarantinedMobFile);
} finally {
http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestJSONMetricUtil.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestJSONMetricUtil.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestJSONMetricUtil.java
index 13c6df5..bb68898 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestJSONMetricUtil.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestJSONMetricUtil.java
@@ -86,7 +86,7 @@ public class TestJSONMetricUtil {
Hashtable<String, String> properties = JSONMetricUtil.buldKeyValueTable(keys, values);
ObjectName testObject = JSONMetricUtil.buildObjectName(JSONMetricUtil.JAVA_LANG_DOMAIN,
properties);
- assertEquals(testObject.getDomain(), JSONMetricUtil.JAVA_LANG_DOMAIN);
+ assertEquals(JSONMetricUtil.JAVA_LANG_DOMAIN, testObject.getDomain());
assertEquals(testObject.getKeyPropertyList(), properties);
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMiniClusterLoadParallel.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMiniClusterLoadParallel.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMiniClusterLoadParallel.java
index 7b1cd2d..e6b05e9 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMiniClusterLoadParallel.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMiniClusterLoadParallel.java
@@ -40,6 +40,7 @@ public class TestMiniClusterLoadParallel
super(isMultiPut, encoding);
}
+ @Override
@Test(timeout=TIMEOUT_MS)
public void loadTest() throws Exception {
prepareForLoadTest();
http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestRegionSplitCalculator.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestRegionSplitCalculator.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestRegionSplitCalculator.java
index fd86beb..eee3030 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestRegionSplitCalculator.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestRegionSplitCalculator.java
@@ -68,6 +68,7 @@ public class TestRegionSplitCalculator {
return end;
}
+ @Override
public String toString() {
return "[" + Bytes.toString(start) + ", " + Bytes.toString(end) + "]";
}
@@ -135,8 +136,7 @@ public class TestRegionSplitCalculator {
LOG.info("Standard");
String res = dump(sc.getSplits(), regions);
checkDepths(sc.getSplits(), regions, 1, 1, 1, 0);
- assertEquals(res, "A:\t[A, B]\t\n" + "B:\t[B, C]\t\n" + "C:\t[C, D]\t\n"
- + "D:\t\n");
+ assertEquals("A:\t[A, B]\t\n" + "B:\t[B, C]\t\n" + "C:\t[C, D]\t\nD:\t\n", res);
}
@Test
http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/test/LoadTestDataGenerator.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/test/LoadTestDataGenerator.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/test/LoadTestDataGenerator.java
index 3650706..1aab1f8 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/test/LoadTestDataGenerator.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/test/LoadTestDataGenerator.java
@@ -20,10 +20,11 @@ import java.io.IOException;
import java.util.Random;
import java.util.Set;
-import org.apache.yetus.audience.InterfaceAudience;
+import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.hbase.client.Get;
import org.apache.hadoop.hbase.client.Mutation;
import org.apache.hadoop.hbase.util.LoadTestKVGenerator;
+import org.apache.yetus.audience.InterfaceAudience;
/**
* A generator of random data (keys/cfs/columns/values) for load testing.
@@ -35,11 +36,11 @@ public abstract class LoadTestDataGenerator {
// The mutate info column stores information
// about update done to this column family this row.
- public final static byte[] MUTATE_INFO = "mutate_info".getBytes();
+ public final static byte[] MUTATE_INFO = Bytes.toBytes("mutate_info");
// The increment column always has a long value,
// which can be incremented later on during updates.
- public final static byte[] INCREMENT = "increment".getBytes();
+ public final static byte[] INCREMENT = Bytes.toBytes("increment");
protected String[] args;
http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALFactory.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALFactory.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALFactory.java
index f0f35e7..ceb43d5 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALFactory.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALFactory.java
@@ -476,7 +476,7 @@ public class TestWALFactory {
reader.close();
// Reset the lease period
- setLeasePeriod.invoke(cluster, new Object[]{new Long(60000), new Long(3600000)});
+ setLeasePeriod.invoke(cluster, new Object[]{ 60000L, 3600000L });
}
/**
http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplit.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplit.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplit.java
index a1206aa..dded506 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplit.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplit.java
@@ -131,9 +131,9 @@ public class TestWALSplit {
private static final String FILENAME_BEING_SPLIT = "testfile";
private static final TableName TABLE_NAME =
TableName.valueOf("t1");
- private static final byte[] FAMILY = "f1".getBytes();
- private static final byte[] QUALIFIER = "q1".getBytes();
- private static final byte[] VALUE = "v1".getBytes();
+ private static final byte[] FAMILY = Bytes.toBytes("f1");
+ private static final byte[] QUALIFIER = Bytes.toBytes("q1");
+ private static final byte[] VALUE = Bytes.toBytes("v1");
private static final String WAL_FILE_PREFIX = "wal.dat.";
private static List<String> REGIONS = new ArrayList<>();
private static final String HBASE_SKIP_ERRORS = "hbase.hlog.split.skip.errors";
@@ -341,7 +341,7 @@ public class TestWALSplit {
while (!stop.get()) {
try {
long seq = appendEntry(writer, TABLE_NAME, regionBytes,
- ("r" + editsCount.get()).getBytes(), regionBytes, QUALIFIER, VALUE, 0);
+ Bytes.toBytes("r" + editsCount.get()), regionBytes, QUALIFIER, VALUE, 0);
long count = editsCount.incrementAndGet();
LOG.info(getName() + " sync count=" + count + ", seq=" + seq);
try {
@@ -407,7 +407,7 @@ public class TestWALSplit {
TableName.META_TABLE_NAME, 1, now, HConstants.DEFAULT_CLUSTER_ID),
new WALEdit());
Path parent = WALSplitter.getRegionDirRecoveredEditsDir(regiondir);
- assertEquals(parent.getName(), HConstants.RECOVERED_EDITS_DIR);
+ assertEquals(HConstants.RECOVERED_EDITS_DIR, parent.getName());
fs.createNewFile(parent); // create a recovered.edits file
Path p = WALSplitter.getRegionSplitEditsPath(fs, entry, HBASEDIR,
@@ -588,8 +588,8 @@ public class TestWALSplit {
archivedLogs.add(log.getPath().getName());
}
LOG.debug(archived.toString());
- assertEquals(failureType.name() + ": expected to find all of our wals corrupt.",
- walDirContents, archivedLogs);
+ assertEquals(failureType.name() + ": expected to find all of our wals corrupt.", archivedLogs,
+ walDirContents);
}
}
@@ -687,7 +687,7 @@ public class TestWALSplit {
// should not have stored the EOF files as corrupt
FileStatus[] archivedLogs = fs.listStatus(CORRUPTDIR);
- assertEquals(archivedLogs.length, 0);
+ assertEquals(0, archivedLogs.length);
}
@@ -749,7 +749,7 @@ public class TestWALSplit {
InstrumentedLogWriter.activateFailure = false;
appendEntry(writer, TABLE_NAME, Bytes.toBytes(region),
- ("r" + 999).getBytes(), FAMILY, QUALIFIER, VALUE, 0);
+ Bytes.toBytes("r" + 999), FAMILY, QUALIFIER, VALUE, 0);
writer.close();
try {
@@ -1206,8 +1206,8 @@ public class TestWALSplit {
int prefix = 0;
for (String region : REGIONS) {
String row_key = region + prefix++ + i + j;
- appendEntry(ws[i], TABLE_NAME, region.getBytes(), row_key.getBytes(), FAMILY, QUALIFIER,
- VALUE, seq++);
+ appendEntry(ws[i], TABLE_NAME, Bytes.toBytes(region), Bytes.toBytes(row_key), FAMILY,
+ QUALIFIER, VALUE, seq++);
if (numRegionEventsAdded < regionEvents) {
numRegionEventsAdded ++;
@@ -1233,7 +1233,7 @@ public class TestWALSplit {
Path tdir = FSUtils.getTableDir(rootdir, table);
@SuppressWarnings("deprecation")
Path editsdir = WALSplitter.getRegionDirRecoveredEditsDir(HRegion.getRegionDir(tdir,
- Bytes.toString(region.getBytes())));
+ Bytes.toString(Bytes.toBytes(region))));
FileStatus[] files = fs.listStatus(editsdir, new PathFilter() {
@Override
public boolean accept(Path p) {
@@ -1260,46 +1260,46 @@ public class TestWALSplit {
in.close();
switch (corruption) {
- case APPEND_GARBAGE:
- fs.delete(path, false);
- out = fs.create(path);
- out.write(corrupted_bytes);
- out.write("-----".getBytes());
- closeOrFlush(close, out);
- break;
-
- case INSERT_GARBAGE_ON_FIRST_LINE:
- fs.delete(path, false);
- out = fs.create(path);
- out.write(0);
- out.write(corrupted_bytes);
- closeOrFlush(close, out);
- break;
-
- case INSERT_GARBAGE_IN_THE_MIDDLE:
- fs.delete(path, false);
- out = fs.create(path);
- int middle = (int) Math.floor(corrupted_bytes.length / 2);
- out.write(corrupted_bytes, 0, middle);
- out.write(0);
- out.write(corrupted_bytes, middle, corrupted_bytes.length - middle);
- closeOrFlush(close, out);
- break;
-
- case TRUNCATE:
- fs.delete(path, false);
- out = fs.create(path);
- out.write(corrupted_bytes, 0, fileSize
- - (32 + ProtobufLogReader.PB_WAL_COMPLETE_MAGIC.length + Bytes.SIZEOF_INT));
- closeOrFlush(close, out);
- break;
-
- case TRUNCATE_TRAILER:
- fs.delete(path, false);
- out = fs.create(path);
- out.write(corrupted_bytes, 0, fileSize - Bytes.SIZEOF_INT);// trailer is truncated.
- closeOrFlush(close, out);
- break;
+ case APPEND_GARBAGE:
+ fs.delete(path, false);
+ out = fs.create(path);
+ out.write(corrupted_bytes);
+ out.write(Bytes.toBytes("-----"));
+ closeOrFlush(close, out);
+ break;
+
+ case INSERT_GARBAGE_ON_FIRST_LINE:
+ fs.delete(path, false);
+ out = fs.create(path);
+ out.write(0);
+ out.write(corrupted_bytes);
+ closeOrFlush(close, out);
+ break;
+
+ case INSERT_GARBAGE_IN_THE_MIDDLE:
+ fs.delete(path, false);
+ out = fs.create(path);
+ int middle = (int) Math.floor(corrupted_bytes.length / 2);
+ out.write(corrupted_bytes, 0, middle);
+ out.write(0);
+ out.write(corrupted_bytes, middle, corrupted_bytes.length - middle);
+ closeOrFlush(close, out);
+ break;
+
+ case TRUNCATE:
+ fs.delete(path, false);
+ out = fs.create(path);
+ out.write(corrupted_bytes, 0, fileSize
+ - (32 + ProtobufLogReader.PB_WAL_COMPLETE_MAGIC.length + Bytes.SIZEOF_INT));
+ closeOrFlush(close, out);
+ break;
+
+ case TRUNCATE_TRAILER:
+ fs.delete(path, false);
+ out = fs.create(path);
+ out.write(corrupted_bytes, 0, fileSize - Bytes.SIZEOF_INT);// trailer is truncated.
+ closeOrFlush(close, out);
+ break;
}
}
@@ -1360,14 +1360,14 @@ public class TestWALSplit {
WALProtos.RegionEventDescriptor regionOpenDesc = ProtobufUtil.toRegionEventDescriptor(
WALProtos.RegionEventDescriptor.EventType.REGION_OPEN,
TABLE_NAME.toBytes(),
- region.getBytes(),
- String.valueOf(region.hashCode()).getBytes(),
+ Bytes.toBytes(region),
+ Bytes.toBytes(String.valueOf(region.hashCode())),
1,
ServerName.parseServerName("ServerName:9099"), ImmutableMap.<byte[], List<Path>>of());
final long time = EnvironmentEdgeManager.currentTime();
- KeyValue kv = new KeyValue(region.getBytes(), WALEdit.METAFAMILY, WALEdit.REGION_EVENT,
+ KeyValue kv = new KeyValue(Bytes.toBytes(region), WALEdit.METAFAMILY, WALEdit.REGION_EVENT,
time, regionOpenDesc.toByteArray());
- final WALKeyImpl walKey = new WALKeyImpl(region.getBytes(), TABLE_NAME, 1, time,
+ final WALKeyImpl walKey = new WALKeyImpl(Bytes.toBytes(region), TABLE_NAME, 1, time,
HConstants.DEFAULT_CLUSTER_ID);
w.append(
new Entry(walKey, new WALEdit().add(kv)));
http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplitBoundedLogWriterCreation.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplitBoundedLogWriterCreation.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplitBoundedLogWriterCreation.java
index 844cb3a..400d12b 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplitBoundedLogWriterCreation.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplitBoundedLogWriterCreation.java
@@ -35,6 +35,7 @@ public class TestWALSplitBoundedLogWriterCreation extends TestWALSplit{
/**
* The logic of this test has conflict with the limit writers split logic, skip this test
*/
+ @Override
@Test(timeout=300000)
@Ignore
public void testThreadingSlowWriterSmallBuffer() throws Exception {
http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZooKeeperACL.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZooKeeperACL.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZooKeeperACL.java
index 4e67b91..3e9e650 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZooKeeperACL.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZooKeeperACL.java
@@ -129,10 +129,10 @@ public class TestZooKeeperACL {
List<ACL> acls = zkw.getRecoverableZooKeeper().getZooKeeper()
.getACL("/hbase", new Stat());
- assertEquals(acls.size(),1);
- assertEquals(acls.get(0).getId().getScheme(),"sasl");
- assertEquals(acls.get(0).getId().getId(),"hbase");
- assertEquals(acls.get(0).getPerms(), ZooDefs.Perms.ALL);
+ assertEquals(1, acls.size());
+ assertEquals("sasl", acls.get(0).getId().getScheme());
+ assertEquals("hbase", acls.get(0).getId().getId());
+ assertEquals(ZooDefs.Perms.ALL, acls.get(0).getPerms());
}
/**
@@ -148,20 +148,20 @@ public class TestZooKeeperACL {
List<ACL> acls = zkw.getRecoverableZooKeeper().getZooKeeper()
.getACL("/hbase/root-region-server", new Stat());
- assertEquals(acls.size(),2);
+ assertEquals(2, acls.size());
boolean foundWorldReadableAcl = false;
boolean foundHBaseOwnerAcl = false;
for(int i = 0; i < 2; i++) {
if (acls.get(i).getId().getScheme().equals("world") == true) {
- assertEquals(acls.get(0).getId().getId(),"anyone");
- assertEquals(acls.get(0).getPerms(), ZooDefs.Perms.READ);
+ assertEquals("anyone", acls.get(0).getId().getId());
+ assertEquals(ZooDefs.Perms.READ, acls.get(0).getPerms());
foundWorldReadableAcl = true;
}
else {
if (acls.get(i).getId().getScheme().equals("sasl") == true) {
- assertEquals(acls.get(1).getId().getId(),"hbase");
- assertEquals(acls.get(1).getId().getScheme(),"sasl");
+ assertEquals("hbase", acls.get(1).getId().getId());
+ assertEquals("sasl", acls.get(1).getId().getScheme());
foundHBaseOwnerAcl = true;
} else { // error: should not get here: test fails.
assertTrue(false);
@@ -185,19 +185,19 @@ public class TestZooKeeperACL {
List<ACL> acls = zkw.getRecoverableZooKeeper().getZooKeeper()
.getACL("/hbase/master", new Stat());
- assertEquals(acls.size(),2);
+ assertEquals(2, acls.size());
boolean foundWorldReadableAcl = false;
boolean foundHBaseOwnerAcl = false;
for(int i = 0; i < 2; i++) {
if (acls.get(i).getId().getScheme().equals("world") == true) {
- assertEquals(acls.get(0).getId().getId(),"anyone");
- assertEquals(acls.get(0).getPerms(), ZooDefs.Perms.READ);
+ assertEquals("anyone", acls.get(0).getId().getId());
+ assertEquals(ZooDefs.Perms.READ, acls.get(0).getPerms());
foundWorldReadableAcl = true;
} else {
if (acls.get(i).getId().getScheme().equals("sasl") == true) {
- assertEquals(acls.get(1).getId().getId(),"hbase");
- assertEquals(acls.get(1).getId().getScheme(),"sasl");
+ assertEquals("hbase", acls.get(1).getId().getId());
+ assertEquals("sasl", acls.get(1).getId().getScheme());
foundHBaseOwnerAcl = true;
} else { // error: should not get here: test fails.
assertTrue(false);
@@ -221,19 +221,19 @@ public class TestZooKeeperACL {
List<ACL> acls = zkw.getRecoverableZooKeeper().getZooKeeper()
.getACL("/hbase/hbaseid", new Stat());
- assertEquals(acls.size(),2);
+ assertEquals(2, acls.size());
boolean foundWorldReadableAcl = false;
boolean foundHBaseOwnerAcl = false;
for(int i = 0; i < 2; i++) {
if (acls.get(i).getId().getScheme().equals("world") == true) {
- assertEquals(acls.get(0).getId().getId(),"anyone");
- assertEquals(acls.get(0).getPerms(), ZooDefs.Perms.READ);
+ assertEquals("anyone", acls.get(0).getId().getId());
+ assertEquals(ZooDefs.Perms.READ, acls.get(0).getPerms());
foundWorldReadableAcl = true;
} else {
if (acls.get(i).getId().getScheme().equals("sasl") == true) {
- assertEquals(acls.get(1).getId().getId(),"hbase");
- assertEquals(acls.get(1).getId().getScheme(),"sasl");
+ assertEquals("hbase", acls.get(1).getId().getId());
+ assertEquals("sasl", acls.get(1).getId().getScheme());
foundHBaseOwnerAcl = true;
} else { // error: should not get here: test fails.
assertTrue(false);
@@ -257,10 +257,10 @@ public class TestZooKeeperACL {
ZKUtil.createWithParents(zkw, "/testACLNode");
List<ACL> acls = zkw.getRecoverableZooKeeper().getZooKeeper()
.getACL("/testACLNode", new Stat());
- assertEquals(acls.size(),1);
- assertEquals(acls.get(0).getId().getScheme(),"sasl");
- assertEquals(acls.get(0).getId().getId(),"hbase");
- assertEquals(acls.get(0).getPerms(), ZooDefs.Perms.ALL);
+ assertEquals(1, acls.size());
+ assertEquals("sasl", acls.get(0).getId().getScheme());
+ assertEquals("hbase", acls.get(0).getId().getId());
+ assertEquals(ZooDefs.Perms.ALL, acls.get(0).getPerms());
}
/**
@@ -281,7 +281,7 @@ public class TestZooKeeperACL {
saslConfFile.getAbsolutePath());
testJaasConfig = ZKUtil.isSecureZooKeeper(new Configuration(TEST_UTIL.getConfiguration()));
- assertEquals(testJaasConfig, false);
+ assertEquals(false, testJaasConfig);
saslConfFile.delete();
}
@@ -295,13 +295,13 @@ public class TestZooKeeperACL {
Configuration config = new Configuration(HBaseConfiguration.create());
boolean testJaasConfig = ZKUtil.isSecureZooKeeper(config);
- assertEquals(testJaasConfig, false);
+ assertEquals(false, testJaasConfig);
// Now set authentication scheme to Kerberos still it should return false
// because no configuration set
config.set("hbase.security.authentication", "kerberos");
testJaasConfig = ZKUtil.isSecureZooKeeper(config);
- assertEquals(testJaasConfig, false);
+ assertEquals(false, testJaasConfig);
// Now set programmatic options related to security
config.set(HConstants.ZK_CLIENT_KEYTAB_FILE, "/dummy/file");