You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2017/03/07 19:23:17 UTC
[04/22] hbase git commit: HBASE-17532 Replaced explicit type with
diamond operator
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreScanner.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreScanner.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreScanner.java
index 52efe63..ccbf067 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreScanner.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreScanner.java
@@ -173,7 +173,7 @@ public class TestStoreScanner {
* @return
*/
NavigableSet<byte[]> getCols(String ...strCols) {
- NavigableSet<byte[]> cols = new TreeSet<byte[]>(Bytes.BYTES_COMPARATOR);
+ NavigableSet<byte[]> cols = new TreeSet<>(Bytes.BYTES_COMPARATOR);
for (String col : strCols) {
byte[] bytes = Bytes.toBytes(col);
cols.add(bytes);
@@ -189,7 +189,7 @@ public class TestStoreScanner {
Scan scan = new Scan(get);
CellGridStoreScanner scanner = new CellGridStoreScanner(scan, this.scanInfo, this.scanType);
try {
- List<Cell> results = new ArrayList<Cell>();
+ List<Cell> results = new ArrayList<>();
while (scanner.next(results)) {
continue;
}
@@ -213,7 +213,7 @@ public class TestStoreScanner {
Scan scan = new Scan(get);
CellGridStoreScanner scanner = new CellGridStoreScanner(scan, this.scanInfo, this.scanType);
try {
- List<Cell> results = new ArrayList<Cell>();
+ List<Cell> results = new ArrayList<>();
while (scanner.next(results)) {
continue;
}
@@ -242,7 +242,7 @@ public class TestStoreScanner {
scan.addColumn(CF, ONE);
CellGridStoreScanner scanner = new CellGridStoreScanner(scan, this.scanInfo, this.scanType);
try {
- List<Cell> results = new ArrayList<Cell>();
+ List<Cell> results = new ArrayList<>();
while (scanner.next(results)) {
continue;
}
@@ -276,7 +276,7 @@ public class TestStoreScanner {
Scan scan = new Scan(get);
CellGridStoreScanner scanner = new CellGridStoreScanner(scan, this.scanInfo, this.scanType);
try {
- List<Cell> results = new ArrayList<Cell>();
+ List<Cell> results = new ArrayList<>();
// For a Get there should be no more next's after the first call.
Assert.assertEquals(false, scanner.next(results));
// Should be one result only.
@@ -307,7 +307,7 @@ public class TestStoreScanner {
Scan scan = new Scan(get);
CellGridStoreScanner scanner = new CellGridStoreScanner(scan, this.scanInfo, this.scanType);
try {
- List<Cell> results = new ArrayList<Cell>();
+ List<Cell> results = new ArrayList<>();
// For a Get there should be no more next's after the first call.
Assert.assertEquals(false, scanner.next(results));
// Should be one result only.
@@ -341,7 +341,7 @@ public class TestStoreScanner {
List<Cell> results = null;
try (StoreScanner scan =
new StoreScanner(scanSpec, scanInfo, scanType, getCols("a"), scanners)) {
- results = new ArrayList<Cell>();
+ results = new ArrayList<>();
Assert.assertEquals(true, scan.next(results));
Assert.assertEquals(5, results.size());
Assert.assertEquals(kvs[kvs.length - 1], results.get(0));
@@ -352,7 +352,7 @@ public class TestStoreScanner {
scanSpec.setMaxVersions();
try (StoreScanner scan =
new StoreScanner(scanSpec, scanInfo, scanType, getCols("a"), scanners)) {
- results = new ArrayList<Cell>();
+ results = new ArrayList<>();
Assert.assertEquals(true, scan.next(results));
Assert.assertEquals(2, results.size());
}
@@ -362,7 +362,7 @@ public class TestStoreScanner {
scanSpec.setMaxVersions();
try (StoreScanner scan =
new StoreScanner(scanSpec, scanInfo, scanType, getCols("a"), scanners)) {
- results = new ArrayList<Cell>();
+ results = new ArrayList<>();
Assert.assertEquals(true, scan.next(results));
Assert.assertEquals(1, results.size());
}
@@ -373,7 +373,7 @@ public class TestStoreScanner {
scanSpec.setMaxVersions(3);
try (StoreScanner scan = new StoreScanner(scanSpec, scanInfo, scanType, getCols("a"),
scanners)) {
- results = new ArrayList<Cell>();
+ results = new ArrayList<>();
Assert.assertEquals(true, scan.next(results));
Assert.assertEquals(3, results.size());
}
@@ -395,7 +395,7 @@ public class TestStoreScanner {
// this only uses maxVersions (default=1) and TimeRange (default=all)
try (StoreScanner scan =
new StoreScanner(scanSpec, scanInfo, scanType, getCols("a"), scanners)) {
- List<Cell> results = new ArrayList<Cell>();
+ List<Cell> results = new ArrayList<>();
Assert.assertEquals(true, scan.next(results));
Assert.assertEquals(1, results.size());
Assert.assertEquals(kvs[0], results.get(0));
@@ -423,7 +423,7 @@ public class TestStoreScanner {
// this only uses maxVersions (default=1) and TimeRange (default=all)
try (StoreScanner scan =
new StoreScanner(scanSpec, scanInfo, scanType, getCols("a"), scanners)) {
- List<Cell> results = new ArrayList<Cell>();
+ List<Cell> results = new ArrayList<>();
scan.next(results);
Assert.assertEquals(1, results.size());
Assert.assertEquals(kvs[0], results.get(0));
@@ -451,7 +451,7 @@ public class TestStoreScanner {
Scan scanSpec = new Scan(Bytes.toBytes("R1"));
try (StoreScanner scan =
new StoreScanner(scanSpec, scanInfo, scanType, getCols("a"), scanners)) {
- List<Cell> results = new ArrayList<Cell>();
+ List<Cell> results = new ArrayList<>();
Assert.assertFalse(scan.next(results));
Assert.assertEquals(0, results.size());
}
@@ -472,7 +472,7 @@ public class TestStoreScanner {
Scan scanSpec = new Scan(Bytes.toBytes("R1"));
try (StoreScanner scan =
new StoreScanner(scanSpec, scanInfo, scanType, getCols("a"), scanners)) {
- List<Cell> results = new ArrayList<Cell>();
+ List<Cell> results = new ArrayList<>();
Assert.assertEquals(true, scan.next(results));
Assert.assertEquals(0, results.size());
@@ -499,7 +499,7 @@ public class TestStoreScanner {
try (StoreScanner scan = new StoreScanner(new Scan(Bytes.toBytes("R1")), scanInfo, scanType,
getCols("a"), scanners)) {
- List<Cell> results = new ArrayList<Cell>();
+ List<Cell> results = new ArrayList<>();
// the two put at ts=now will be masked by the 1 delete, and
// since the scan default returns 1 version we'll return the newest
// key, which is kvs[2], now-100.
@@ -525,7 +525,7 @@ public class TestStoreScanner {
Scan scanSpec = new Scan(Bytes.toBytes("R1")).setMaxVersions(2);
try (StoreScanner scan = new StoreScanner(scanSpec, scanInfo, scanType,
getCols("a"), scanners)) {
- List<Cell> results = new ArrayList<Cell>();
+ List<Cell> results = new ArrayList<>();
Assert.assertEquals(true, scan.next(results));
Assert.assertEquals(2, results.size());
Assert.assertEquals(kvs2[1], results.get(0));
@@ -543,7 +543,7 @@ public class TestStoreScanner {
List<KeyValueScanner> scanners = scanFixture(kvs);
try (StoreScanner scan =
new StoreScanner(new Scan(Bytes.toBytes("R1")), scanInfo, scanType, null, scanners)) {
- List<Cell> results = new ArrayList<Cell>();
+ List<Cell> results = new ArrayList<>();
Assert.assertEquals(true, scan.next(results));
Assert.assertEquals(2, results.size());
Assert.assertEquals(kvs[0], results.get(0));
@@ -574,7 +574,7 @@ public class TestStoreScanner {
List<KeyValueScanner> scanners = scanFixture(kvs);
try (StoreScanner scan = new StoreScanner(new Scan().setMaxVersions(2),
scanInfo, scanType, null, scanners)) {
- List<Cell> results = new ArrayList<Cell>();
+ List<Cell> results = new ArrayList<>();
Assert.assertEquals(true, scan.next(results));
Assert.assertEquals(5, results.size());
Assert.assertEquals(kvs[0], results.get(0));
@@ -605,7 +605,7 @@ public class TestStoreScanner {
try (StoreScanner scan =
new StoreScanner(new Scan().setMaxVersions(Integer.MAX_VALUE), scanInfo, scanType, null,
scanners)) {
- List<Cell> results = new ArrayList<Cell>();
+ List<Cell> results = new ArrayList<>();
Assert.assertEquals(true, scan.next(results));
Assert.assertEquals(0, results.size());
Assert.assertEquals(true, scan.next(results));
@@ -627,7 +627,7 @@ public class TestStoreScanner {
List<KeyValueScanner> scanners = scanFixture(kvs);
try (StoreScanner scan = new StoreScanner(new Scan(), scanInfo, scanType, null,
scanners)) {
- List<Cell> results = new ArrayList<Cell>();
+ List<Cell> results = new ArrayList<>();
Assert.assertEquals(true, scan.next(results));
Assert.assertEquals(1, results.size());
Assert.assertEquals(kvs[3], results.get(0));
@@ -652,7 +652,7 @@ public class TestStoreScanner {
List<KeyValueScanner> scanners = scanFixture(kvs);
try (StoreScanner scan =
new StoreScanner(new Scan(), scanInfo, scanType, getCols("a", "d"), scanners)) {
- List<Cell> results = new ArrayList<Cell>();
+ List<Cell> results = new ArrayList<>();
Assert.assertEquals(true, scan.next(results));
Assert.assertEquals(2, results.size());
Assert.assertEquals(kvs[0], results.get(0));
@@ -692,7 +692,7 @@ public class TestStoreScanner {
CellComparator.COMPARATOR);
ScanType scanType = ScanType.USER_SCAN;
try (StoreScanner scanner = new StoreScanner(scan, scanInfo, scanType, null, scanners)) {
- List<Cell> results = new ArrayList<Cell>();
+ List<Cell> results = new ArrayList<>();
Assert.assertEquals(true, scanner.next(results));
Assert.assertEquals(2, results.size());
Assert.assertEquals(kvs[1], results.get(0));
@@ -720,9 +720,9 @@ public class TestStoreScanner {
// normally cause an NPE because scan.store is null. So as long as we get through these
// two calls we are good and the bug was quashed.
- scan.updateReaders(new ArrayList<StoreFile>());
+ scan.updateReaders(new ArrayList<>());
- scan.updateReaders(new ArrayList<StoreFile>());
+ scan.updateReaders(new ArrayList<>());
scan.peek();
}
@@ -767,7 +767,7 @@ public class TestStoreScanner {
try (StoreScanner scanner =
new StoreScanner(scan, scanInfo, scanType, null, scanners)) {
- List<Cell> results = new ArrayList<Cell>();
+ List<Cell> results = new ArrayList<>();
Assert.assertEquals(true, scanner.next(results));
Assert.assertEquals(1, results.size());
Assert.assertEquals(kvs[1], results.get(0));
@@ -833,8 +833,8 @@ public class TestStoreScanner {
new StoreScanner(scan, scanInfo,
ScanType.COMPACT_DROP_DELETES, null, scanners,
HConstants.OLDEST_TIMESTAMP)) {
- List<Cell> results = new ArrayList<Cell>();
- results = new ArrayList<Cell>();
+ List<Cell> results = new ArrayList<>();
+ results = new ArrayList<>();
Assert.assertEquals(true, scanner.next(results));
Assert.assertEquals(kvs[0], results.get(0));
Assert.assertEquals(kvs[2], results.get(1));
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreEngine.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreEngine.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreEngine.java
index 3e3eef9..b2739e1 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreEngine.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreEngine.java
@@ -78,7 +78,7 @@ public class TestStripeStoreEngine {
mockCompactor.compact(any(CompactionRequest.class), anyInt(), anyLong(), any(byte[].class),
any(byte[].class), any(byte[].class), any(byte[].class),
any(ThroughputController.class), any(User.class)))
- .thenReturn(new ArrayList<Path>());
+ .thenReturn(new ArrayList<>());
// Produce 3 L0 files.
StoreFile sf = createFile();
@@ -118,6 +118,6 @@ public class TestStripeStoreEngine {
}
private static ArrayList<StoreFile> al(StoreFile... sfs) {
- return new ArrayList<StoreFile>(Arrays.asList(sfs));
+ return new ArrayList<>(Arrays.asList(sfs));
}
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreFileManager.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreFileManager.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreFileManager.java
index c533257..a6ce270 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreFileManager.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreFileManager.java
@@ -115,7 +115,7 @@ public class TestStripeStoreFileManager {
}
private static ArrayList<StoreFile> dumpIterator(Iterator<StoreFile> iter) {
- ArrayList<StoreFile> result = new ArrayList<StoreFile>();
+ ArrayList<StoreFile> result = new ArrayList<>();
for (; iter.hasNext(); result.add(iter.next()));
return result;
}
@@ -233,7 +233,7 @@ public class TestStripeStoreFileManager {
private void verifySplitPointScenario(int splitPointAfter, boolean shouldSplitStripe,
float splitRatioToVerify, int... sizes) throws Exception {
assertTrue(sizes.length > 1);
- ArrayList<StoreFile> sfs = new ArrayList<StoreFile>();
+ ArrayList<StoreFile> sfs = new ArrayList<>();
for (int sizeIx = 0; sizeIx < sizes.length; ++sizeIx) {
byte[] startKey = (sizeIx == 0) ? OPEN_KEY : Bytes.toBytes(sizeIx - 1);
byte[] endKey = (sizeIx == sizes.length - 1) ? OPEN_KEY : Bytes.toBytes(sizeIx);
@@ -525,7 +525,7 @@ public class TestStripeStoreFileManager {
sfm.insertNewFiles(al(createFile()));
}
for (int i = 0; i < filesInStripe; ++i) {
- ArrayList<StoreFile> stripe = new ArrayList<StoreFile>();
+ ArrayList<StoreFile> stripe = new ArrayList<>();
for (int j = 0; j < stripes; ++j) {
stripe.add(createFile(
(j == 0) ? OPEN_KEY : keys[j - 1], (j == stripes - 1) ? OPEN_KEY : keys[j]));
@@ -597,7 +597,7 @@ public class TestStripeStoreFileManager {
}
private static StripeStoreFileManager createManager() throws Exception {
- return createManager(new ArrayList<StoreFile>());
+ return createManager(new ArrayList<>());
}
private static StripeStoreFileManager createManager(ArrayList<StoreFile> sfs) throws Exception {
@@ -615,11 +615,11 @@ public class TestStripeStoreFileManager {
}
private static ArrayList<StoreFile> al(StoreFile... sfs) {
- return new ArrayList<StoreFile>(Arrays.asList(sfs));
+ return new ArrayList<>(Arrays.asList(sfs));
}
private static ArrayList<StoreFile> flattenLists(ArrayList<StoreFile>... sfls) {
- ArrayList<StoreFile> result = new ArrayList<StoreFile>();
+ ArrayList<StoreFile> result = new ArrayList<>();
for (ArrayList<StoreFile> sfl : sfls) {
result.addAll(sfl);
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestTags.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestTags.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestTags.java
index 4d531ac..40eebb6 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestTags.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestTags.java
@@ -421,7 +421,7 @@ public class TestTags {
assertEquals(5L, Bytes.toLong(kv.getValueArray(), kv.getValueOffset(), kv.getValueLength()));
assertEquals(2, tags.size());
// We cannot assume the ordering of tags
- List<String> tagValues = new ArrayList<String>();
+ List<String> tagValues = new ArrayList<>();
for (Tag tag: tags) {
tagValues.add(Bytes.toString(TagUtil.cloneValue(tag)));
}
@@ -557,7 +557,7 @@ public class TestTags {
private void updateMutationAddingTags(final Mutation m) {
byte[] attribute = m.getAttribute("visibility");
byte[] cf = null;
- List<Cell> updatedCells = new ArrayList<Cell>();
+ List<Cell> updatedCells = new ArrayList<>();
if (attribute != null) {
for (List<? extends Cell> edits : m.getFamilyCellMap().values()) {
for (Cell cell : edits) {
@@ -566,7 +566,7 @@ public class TestTags {
cf = CellUtil.cloneFamily(kv);
}
Tag tag = new ArrayBackedTag((byte) 1, attribute);
- List<Tag> tagList = new ArrayList<Tag>();
+ List<Tag> tagList = new ArrayList<>();
tagList.add(tag);
KeyValue newKV = new KeyValue(CellUtil.cloneRow(kv), 0, kv.getRowLength(),
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWALLockup.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWALLockup.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWALLockup.java
index 4821c74..4f247b0 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWALLockup.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWALLockup.java
@@ -218,7 +218,7 @@ public class TestWALLockup {
HTableDescriptor htd = new HTableDescriptor(TableName.META_TABLE_NAME);
final HRegion region = initHRegion(tableName, null, null, dodgyWAL);
byte [] bytes = Bytes.toBytes(getName());
- NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>(
+ NavigableMap<byte[], Integer> scopes = new TreeMap<>(
Bytes.BYTES_COMPARATOR);
scopes.put(COLUMN_FAMILY_BYTES, 0);
MultiVersionConcurrencyControl mvcc = new MultiVersionConcurrencyControl();
@@ -398,7 +398,7 @@ public class TestWALLockup {
HTableDescriptor htd = new HTableDescriptor(TableName.META_TABLE_NAME);
final HRegion region = initHRegion(tableName, null, null, dodgyWAL1);
byte[] bytes = Bytes.toBytes(getName());
- NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>(
+ NavigableMap<byte[], Integer> scopes = new TreeMap<>(
Bytes.BYTES_COMPARATOR);
scopes.put(COLUMN_FAMILY_BYTES, 0);
MultiVersionConcurrencyControl mvcc = new MultiVersionConcurrencyControl();
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWideScanner.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWideScanner.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWideScanner.java
index f598a8d..cdf84d2 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWideScanner.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWideScanner.java
@@ -94,7 +94,7 @@ public class TestWideScanner extends HBaseTestCase {
try {
this.r = createNewHRegion(TESTTABLEDESC, null, null);
int inserted = addWideContent(this.r);
- List<Cell> results = new ArrayList<Cell>();
+ List<Cell> results = new ArrayList<>();
Scan scan = new Scan();
scan.addFamily(A);
scan.addFamily(B);
@@ -130,7 +130,7 @@ public class TestWideScanner extends HBaseTestCase {
((HRegion.RegionScannerImpl)s).storeHeap.getHeap().iterator();
while (scanners.hasNext()) {
StoreScanner ss = (StoreScanner)scanners.next();
- ss.updateReaders(new ArrayList<StoreFile>());
+ ss.updateReaders(new ArrayList<>());
}
} while (more);
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/ConstantSizeFileListGenerator.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/ConstantSizeFileListGenerator.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/ConstantSizeFileListGenerator.java
index 68d57af..5014b41 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/ConstantSizeFileListGenerator.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/ConstantSizeFileListGenerator.java
@@ -49,7 +49,7 @@ class ConstantSizeFileListGenerator extends StoreFileListGenerator {
@Override
public List<StoreFile> next() {
count += 1;
- ArrayList<StoreFile> files = new ArrayList<StoreFile>(NUM_FILES_GEN);
+ ArrayList<StoreFile> files = new ArrayList<>(NUM_FILES_GEN);
for (int i = 0; i < NUM_FILES_GEN; i++) {
files.add(createMockStoreFile(FILESIZE));
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/EverythingPolicy.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/EverythingPolicy.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/EverythingPolicy.java
index 9a4bb8e..46bb639 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/EverythingPolicy.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/EverythingPolicy.java
@@ -46,9 +46,9 @@ public class EverythingPolicy extends RatioBasedCompactionPolicy {
final boolean mayUseOffPeak, final boolean mayBeStuck) throws IOException {
if (candidates.size() < comConf.getMinFilesToCompact()) {
- return new ArrayList<StoreFile>(0);
+ return new ArrayList<>(0);
}
- return new ArrayList<StoreFile>(candidates);
+ return new ArrayList<>(candidates);
}
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/GaussianFileListGenerator.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/GaussianFileListGenerator.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/GaussianFileListGenerator.java
index a19e9ad..fb8c30a 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/GaussianFileListGenerator.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/GaussianFileListGenerator.java
@@ -47,7 +47,7 @@ class GaussianFileListGenerator extends StoreFileListGenerator {
@Override
public List<StoreFile> next() {
count += 1;
- ArrayList<StoreFile> files = new ArrayList<StoreFile>(NUM_FILES_GEN);
+ ArrayList<StoreFile> files = new ArrayList<>(NUM_FILES_GEN);
for (int i = 0; i < NUM_FILES_GEN; i++) {
files.add(createMockStoreFile(
(int) Math.ceil(Math.max(0, gen.nextNormalizedDouble() * 32 + 32)))
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/MockStoreFileGenerator.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/MockStoreFileGenerator.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/MockStoreFileGenerator.java
index 663714a..cb97d27 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/MockStoreFileGenerator.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/MockStoreFileGenerator.java
@@ -46,7 +46,7 @@ class MockStoreFileGenerator {
}
protected List<StoreFile> createStoreFileList(final int[] fs) {
- List<StoreFile> storeFiles = new LinkedList<StoreFile>();
+ List<StoreFile> storeFiles = new LinkedList<>();
for (int fileSize : fs) {
storeFiles.add(createMockStoreFile(fileSize));
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/PerfTestCompactionPolicies.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/PerfTestCompactionPolicies.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/PerfTestCompactionPolicies.java
index 3fcd3fe..0a84fe9 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/PerfTestCompactionPolicies.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/PerfTestCompactionPolicies.java
@@ -82,7 +82,7 @@ public class PerfTestCompactionPolicies extends MockStoreFileGenerator {
int[] minFilesValues = new int[] {3};
float[] ratioValues = new float[] {1.2f};
- List<Object[]> params = new ArrayList<Object[]>(
+ List<Object[]> params = new ArrayList<>(
maxFileValues.length
* minFilesValues.length
* fileListGenClasses.length
@@ -152,7 +152,7 @@ public class PerfTestCompactionPolicies extends MockStoreFileGenerator {
public final void testSelection() throws Exception {
long fileDiff = 0;
for (List<StoreFile> storeFileList : generator) {
- List<StoreFile> currentFiles = new ArrayList<StoreFile>(18);
+ List<StoreFile> currentFiles = new ArrayList<>(18);
for (StoreFile file : storeFileList) {
currentFiles.add(file);
currentFiles = runIteration(currentFiles);
@@ -175,16 +175,16 @@ public class PerfTestCompactionPolicies extends MockStoreFileGenerator {
private List<StoreFile> runIteration(List<StoreFile> startingStoreFiles) throws IOException {
- List<StoreFile> storeFiles = new ArrayList<StoreFile>(startingStoreFiles);
+ List<StoreFile> storeFiles = new ArrayList<>(startingStoreFiles);
CompactionRequest req = cp.selectCompaction(
- storeFiles, new ArrayList<StoreFile>(), false, false, false);
+ storeFiles, new ArrayList<>(), false, false, false);
long newFileSize = 0;
Collection<StoreFile> filesToCompact = req.getFiles();
if (!filesToCompact.isEmpty()) {
- storeFiles = new ArrayList<StoreFile>(storeFiles);
+ storeFiles = new ArrayList<>(storeFiles);
storeFiles.removeAll(filesToCompact);
for (StoreFile storeFile : filesToCompact) {
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SemiConstantSizeFileListGenerator.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SemiConstantSizeFileListGenerator.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SemiConstantSizeFileListGenerator.java
index ed4531a..5fe47f3 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SemiConstantSizeFileListGenerator.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SemiConstantSizeFileListGenerator.java
@@ -42,7 +42,7 @@ class SemiConstantSizeFileListGenerator extends StoreFileListGenerator {
@Override
public List<StoreFile> next() {
count += 1;
- ArrayList<StoreFile> files = new ArrayList<StoreFile>(NUM_FILES_GEN);
+ ArrayList<StoreFile> files = new ArrayList<>(NUM_FILES_GEN);
for (int i = 0; i < NUM_FILES_GEN; i++) {
files.add(createMockStoreFile(random.nextInt(5) + 30));
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SinusoidalFileListGenerator.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SinusoidalFileListGenerator.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SinusoidalFileListGenerator.java
index 6afbb2f..f5f36ac 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SinusoidalFileListGenerator.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SinusoidalFileListGenerator.java
@@ -43,7 +43,7 @@ class SinusoidalFileListGenerator extends StoreFileListGenerator {
@Override
public List<StoreFile> next() {
count += 1;
- ArrayList<StoreFile> files = new ArrayList<StoreFile>(NUM_FILES_GEN);
+ ArrayList<StoreFile> files = new ArrayList<>(NUM_FILES_GEN);
for (int x = 0; x < NUM_FILES_GEN; x++) {
int fileSize = (int) Math.abs(64 * Math.sin((Math.PI * x) / 50.0)) + 1;
files.add(createMockStoreFile(fileSize));
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SpikyFileListGenerator.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SpikyFileListGenerator.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SpikyFileListGenerator.java
index ebaa711..5201eb7 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SpikyFileListGenerator.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SpikyFileListGenerator.java
@@ -43,7 +43,7 @@ class SpikyFileListGenerator extends StoreFileListGenerator {
@Override
public List<StoreFile> next() {
count += 1;
- ArrayList<StoreFile> files = new ArrayList<StoreFile>(NUM_FILES_GEN);
+ ArrayList<StoreFile> files = new ArrayList<>(NUM_FILES_GEN);
for (int x = 0; x < NUM_FILES_GEN; x++) {
int fileSize = random.nextInt(5) + 1;
if ( x % 10 == 0) {
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestCompactedHFilesDischarger.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestCompactedHFilesDischarger.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestCompactedHFilesDischarger.java
index aa5a20e..08fc7bf 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestCompactedHFilesDischarger.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestCompactedHFilesDischarger.java
@@ -77,7 +77,7 @@ public class TestCompactedHFilesDischarger {
Path path = testUtil.getDataTestDir(getClass().getSimpleName());
region = HBaseTestingUtility.createRegionAndWAL(info, path, testUtil.getConfiguration(), htd);
rss = mock(RegionServerServices.class);
- List<Region> regions = new ArrayList<Region>(1);
+ List<Region> regions = new ArrayList<>(1);
regions.add(region);
when(rss.getOnlineRegions()).thenReturn(regions);
}
@@ -379,7 +379,7 @@ public class TestCompactedHFilesDischarger {
RegionScanner resScanner = null;
try {
resScanner = region.getScanner(scan);
- List<Cell> results = new ArrayList<Cell>();
+ List<Cell> results = new ArrayList<>();
boolean next = resScanner.next(results);
try {
counter.incrementAndGet();
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestCompactor.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestCompactor.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestCompactor.java
index 89f61d0..dff6919 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestCompactor.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestCompactor.java
@@ -82,12 +82,12 @@ public class TestCompactor {
public static class StoreFileWritersCapture
implements Answer<StoreFileWriter>, StripeMultiFileWriter.WriterFactory {
public static class Writer {
- public ArrayList<KeyValue> kvs = new ArrayList<KeyValue>();
- public TreeMap<byte[], byte[]> data = new TreeMap<byte[], byte[]>(Bytes.BYTES_COMPARATOR);
+ public ArrayList<KeyValue> kvs = new ArrayList<>();
+ public TreeMap<byte[], byte[]> data = new TreeMap<>(Bytes.BYTES_COMPARATOR);
public boolean hasMetadata;
}
- private List<Writer> writers = new ArrayList<Writer>();
+ private List<Writer> writers = new ArrayList<>();
@Override
public StoreFileWriter createWriter() throws IOException {
@@ -192,7 +192,7 @@ public class TestCompactor {
private final ArrayList<KeyValue> kvs;
public Scanner(KeyValue... kvs) {
- this.kvs = new ArrayList<KeyValue>(Arrays.asList(kvs));
+ this.kvs = new ArrayList<>(Arrays.asList(kvs));
}
@Override
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestDateTieredCompactor.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestDateTieredCompactor.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestDateTieredCompactor.java
index 38d9f99..e590639 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestDateTieredCompactor.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestDateTieredCompactor.java
@@ -156,7 +156,7 @@ public class TestDateTieredCompactor {
StoreFileWritersCapture writers = new StoreFileWritersCapture();
CompactionRequest request = createDummyRequest();
DateTieredCompactor dtc = createCompactor(writers, new KeyValue[0],
- new ArrayList<StoreFile>(request.getFiles()));
+ new ArrayList<>(request.getFiles()));
List<Path> paths = dtc.compact(request, Arrays.asList(Long.MIN_VALUE, Long.MAX_VALUE),
NoLimitThroughputController.INSTANCE, null);
assertEquals(1, paths.size());
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestStripeCompactionPolicy.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestStripeCompactionPolicy.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestStripeCompactionPolicy.java
index 5fadee8..f2d00b3 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestStripeCompactionPolicy.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestStripeCompactionPolicy.java
@@ -415,7 +415,7 @@ public class TestStripeCompactionPolicy {
private static StripeCompactionPolicy.StripeInformationProvider createStripesWithFiles(
List<StoreFile>... stripeFiles) throws Exception {
return createStripesWithFiles(createBoundaries(stripeFiles.length),
- Lists.newArrayList(stripeFiles), new ArrayList<StoreFile>());
+ Lists.newArrayList(stripeFiles), new ArrayList<>());
}
@Test
@@ -433,7 +433,7 @@ public class TestStripeCompactionPolicy {
verifySingleStripeCompaction(policy, si, 0, false);
// Unless there are enough to cause L0 compaction.
si = createStripesWithSizes(6, 2, stripes);
- ConcatenatedLists<StoreFile> sfs = new ConcatenatedLists<StoreFile>();
+ ConcatenatedLists<StoreFile> sfs = new ConcatenatedLists<>();
sfs.addSublist(si.getLevel0Files());
sfs.addSublist(si.getStripes().get(0));
verifyCompaction(
@@ -446,7 +446,7 @@ public class TestStripeCompactionPolicy {
// if all files of stripe aren't selected, delete must not be dropped.
stripes = new Long[][] { new Long[] { 100L, 3L, 2L, 2L, 2L }, new Long[] { 6L } };
si = createStripesWithSizes(0, 0, stripes);
- List<StoreFile> compact_file = new ArrayList<StoreFile>();
+ List<StoreFile> compact_file = new ArrayList<>();
Iterator<StoreFile> iter = si.getStripes().get(0).listIterator(1);
while (iter.hasNext()) {
compact_file.add(iter.next());
@@ -472,7 +472,7 @@ public class TestStripeCompactionPolicy {
}
private static ArrayList<StoreFile> al(StoreFile... sfs) {
- return new ArrayList<StoreFile>(Arrays.asList(sfs));
+ return new ArrayList<>(Arrays.asList(sfs));
}
private void verifyMergeCompatcion(StripeCompactionPolicy policy, StripeInformationProvider si,
@@ -619,7 +619,7 @@ public class TestStripeCompactionPolicy {
private static List<StoreFile> getAllFiles(
StripeInformationProvider si, int fromStripe, int toStripe) {
- ArrayList<StoreFile> expected = new ArrayList<StoreFile>();
+ ArrayList<StoreFile> expected = new ArrayList<>();
for (int i = fromStripe; i <= toStripe; ++i) {
expected.addAll(si.getStripes().get(i));
}
@@ -633,11 +633,11 @@ public class TestStripeCompactionPolicy {
*/
private static StripeInformationProvider createStripes(
int l0Count, byte[]... boundaries) throws Exception {
- List<Long> l0Sizes = new ArrayList<Long>();
+ List<Long> l0Sizes = new ArrayList<>();
for (int i = 0; i < l0Count; ++i) {
l0Sizes.add(5L);
}
- List<List<Long>> sizes = new ArrayList<List<Long>>();
+ List<List<Long>> sizes = new ArrayList<>();
for (int i = 0; i <= boundaries.length; ++i) {
sizes.add(Arrays.asList(Long.valueOf(5)));
}
@@ -651,11 +651,11 @@ public class TestStripeCompactionPolicy {
*/
private static StripeInformationProvider createStripesL0Only(
int l0Count, long l0Size) throws Exception {
- List<Long> l0Sizes = new ArrayList<Long>();
+ List<Long> l0Sizes = new ArrayList<>();
for (int i = 0; i < l0Count; ++i) {
l0Sizes.add(l0Size);
}
- return createStripes(null, new ArrayList<List<Long>>(), l0Sizes);
+ return createStripes(null, new ArrayList<>(), l0Sizes);
}
/**
@@ -666,7 +666,7 @@ public class TestStripeCompactionPolicy {
*/
private static StripeInformationProvider createStripesWithSizes(
int l0Count, long l0Size, Long[]... sizes) throws Exception {
- ArrayList<List<Long>> sizeList = new ArrayList<List<Long>>(sizes.length);
+ ArrayList<List<Long>> sizeList = new ArrayList<>(sizes.length);
for (Long[] size : sizes) {
sizeList.add(Arrays.asList(size));
}
@@ -676,7 +676,7 @@ public class TestStripeCompactionPolicy {
private static StripeInformationProvider createStripesWithSizes(
int l0Count, long l0Size, List<List<Long>> sizes) throws Exception {
List<byte[]> boundaries = createBoundaries(sizes.size());
- List<Long> l0Sizes = new ArrayList<Long>();
+ List<Long> l0Sizes = new ArrayList<>();
for (int i = 0; i < l0Count; ++i) {
l0Sizes.add(l0Size);
}
@@ -686,22 +686,22 @@ public class TestStripeCompactionPolicy {
private static List<byte[]> createBoundaries(int stripeCount) {
byte[][] keys = new byte[][] { KEY_A, KEY_B, KEY_C, KEY_D, KEY_E };
assert stripeCount <= keys.length + 1;
- List<byte[]> boundaries = new ArrayList<byte[]>();
+ List<byte[]> boundaries = new ArrayList<>();
boundaries.addAll(Arrays.asList(keys).subList(0, stripeCount - 1));
return boundaries;
}
private static StripeInformationProvider createStripes(List<byte[]> boundaries,
List<List<Long>> stripeSizes, List<Long> l0Sizes) throws Exception {
- List<List<StoreFile>> stripeFiles = new ArrayList<List<StoreFile>>(stripeSizes.size());
+ List<List<StoreFile>> stripeFiles = new ArrayList<>(stripeSizes.size());
for (List<Long> sizes : stripeSizes) {
- List<StoreFile> sfs = new ArrayList<StoreFile>(sizes.size());
+ List<StoreFile> sfs = new ArrayList<>(sizes.size());
for (Long size : sizes) {
sfs.add(createFile(size));
}
stripeFiles.add(sfs);
}
- List<StoreFile> l0Files = new ArrayList<StoreFile>();
+ List<StoreFile> l0Files = new ArrayList<>();
for (Long size : l0Sizes) {
l0Files.add(createFile(size));
}
@@ -713,8 +713,8 @@ public class TestStripeCompactionPolicy {
*/
private static StripeInformationProvider createStripesWithFiles(List<byte[]> boundaries,
List<List<StoreFile>> stripeFiles, List<StoreFile> l0Files) throws Exception {
- ArrayList<ImmutableList<StoreFile>> stripes = new ArrayList<ImmutableList<StoreFile>>();
- ArrayList<byte[]> boundariesList = new ArrayList<byte[]>();
+ ArrayList<ImmutableList<StoreFile>> stripes = new ArrayList<>();
+ ArrayList<byte[]> boundariesList = new ArrayList<>();
StripeInformationProvider si = mock(StripeInformationProvider.class);
if (!stripeFiles.isEmpty()) {
assert stripeFiles.size() == (boundaries.size() + 1);
@@ -731,7 +731,7 @@ public class TestStripeCompactionPolicy {
when(si.getEndRow(eq(i))).thenReturn(endKey);
}
}
- ConcatenatedLists<StoreFile> sfs = new ConcatenatedLists<StoreFile>();
+ ConcatenatedLists<StoreFile> sfs = new ConcatenatedLists<>();
sfs.addAllSublists(stripes);
sfs.addSublist(l0Files);
when(si.getStorefiles()).thenReturn(sfs);
@@ -803,7 +803,7 @@ public class TestStripeCompactionPolicy {
private final ArrayList<KeyValue> kvs;
public Scanner(KeyValue... kvs) {
- this.kvs = new ArrayList<KeyValue>(Arrays.asList(kvs));
+ this.kvs = new ArrayList<>(Arrays.asList(kvs));
}
@Override
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestStripeCompactor.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestStripeCompactor.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestStripeCompactor.java
index 4b82940..088c958 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestStripeCompactor.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestStripeCompactor.java
@@ -177,7 +177,7 @@ public class TestStripeCompactor {
null, NoLimitThroughputController.INSTANCE, null);
assertEquals(output.length, paths.size());
writers.verifyKvs(output, true, true);
- List<byte[]> boundaries = new ArrayList<byte[]>(output.length + 2);
+ List<byte[]> boundaries = new ArrayList<>(output.length + 2);
boundaries.add(left);
for (int i = 1; i < output.length; ++i) {
boundaries.add(CellUtil.cloneRow(output[i][0]));
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestCompactionScanQueryMatcher.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestCompactionScanQueryMatcher.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestCompactionScanQueryMatcher.java
index 055fe1c..af8c27d 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestCompactionScanQueryMatcher.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestCompactionScanQueryMatcher.java
@@ -79,8 +79,7 @@ public class TestCompactionScanQueryMatcher extends AbstractTestScanQueryMatcher
CompactionScanQueryMatcher qm = CompactionScanQueryMatcher.create(scanInfo,
ScanType.COMPACT_RETAIN_DELETES, Long.MAX_VALUE, HConstants.OLDEST_TIMESTAMP,
HConstants.OLDEST_TIMESTAMP, now, from, to, null);
- List<ScanQueryMatcher.MatchCode> actual = new ArrayList<ScanQueryMatcher.MatchCode>(
- rows.length);
+ List<ScanQueryMatcher.MatchCode> actual = new ArrayList<>(rows.length);
byte[] prevRow = null;
for (byte[] row : rows) {
if (prevRow == null || !Bytes.equals(prevRow, row)) {
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestExplicitColumnTracker.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestExplicitColumnTracker.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestExplicitColumnTracker.java
index 3480597..4e07f80 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestExplicitColumnTracker.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestExplicitColumnTracker.java
@@ -49,7 +49,7 @@ public class TestExplicitColumnTracker {
ColumnTracker exp = new ExplicitColumnTracker(trackColumns, 0, maxVersions, Long.MIN_VALUE);
// Initialize result
- List<ScanQueryMatcher.MatchCode> result = new ArrayList<ScanQueryMatcher.MatchCode>(scannerColumns.size());
+ List<ScanQueryMatcher.MatchCode> result = new ArrayList<>(scannerColumns.size());
long timestamp = 0;
// "Match"
@@ -67,11 +67,11 @@ public class TestExplicitColumnTracker {
@Test
public void testGetSingleVersion() throws IOException {
// Create tracker
- TreeSet<byte[]> columns = new TreeSet<byte[]>(Bytes.BYTES_COMPARATOR);
+ TreeSet<byte[]> columns = new TreeSet<>(Bytes.BYTES_COMPARATOR);
// Looking for every other
columns.add(col2);
columns.add(col4);
- List<MatchCode> expected = new ArrayList<ScanQueryMatcher.MatchCode>(5);
+ List<MatchCode> expected = new ArrayList<>(5);
expected.add(ScanQueryMatcher.MatchCode.SEEK_NEXT_COL); // col1
expected.add(ScanQueryMatcher.MatchCode.INCLUDE_AND_SEEK_NEXT_COL); // col2
expected.add(ScanQueryMatcher.MatchCode.SEEK_NEXT_COL); // col3
@@ -80,7 +80,7 @@ public class TestExplicitColumnTracker {
int maxVersions = 1;
// Create "Scanner"
- List<byte[]> scanner = new ArrayList<byte[]>(5);
+ List<byte[]> scanner = new ArrayList<>(5);
scanner.add(col1);
scanner.add(col2);
scanner.add(col3);
@@ -93,12 +93,12 @@ public class TestExplicitColumnTracker {
@Test
public void testGetMultiVersion() throws IOException {
// Create tracker
- TreeSet<byte[]> columns = new TreeSet<byte[]>(Bytes.BYTES_COMPARATOR);
+ TreeSet<byte[]> columns = new TreeSet<>(Bytes.BYTES_COMPARATOR);
// Looking for every other
columns.add(col2);
columns.add(col4);
- List<ScanQueryMatcher.MatchCode> expected = new ArrayList<ScanQueryMatcher.MatchCode>(15);
+ List<ScanQueryMatcher.MatchCode> expected = new ArrayList<>(15);
expected.add(ScanQueryMatcher.MatchCode.SEEK_NEXT_COL);
expected.add(ScanQueryMatcher.MatchCode.SEEK_NEXT_COL);
expected.add(ScanQueryMatcher.MatchCode.SEEK_NEXT_COL);
@@ -121,7 +121,7 @@ public class TestExplicitColumnTracker {
int maxVersions = 2;
// Create "Scanner"
- List<byte[]> scanner = new ArrayList<byte[]>(15);
+ List<byte[]> scanner = new ArrayList<>(15);
scanner.add(col1);
scanner.add(col1);
scanner.add(col1);
@@ -148,7 +148,7 @@ public class TestExplicitColumnTracker {
@Test
public void testStackOverflow() throws IOException {
int maxVersions = 1;
- TreeSet<byte[]> columns = new TreeSet<byte[]>(Bytes.BYTES_COMPARATOR);
+ TreeSet<byte[]> columns = new TreeSet<>(Bytes.BYTES_COMPARATOR);
for (int i = 0; i < 100000; i++) {
columns.add(Bytes.toBytes("col" + i));
}
@@ -173,7 +173,7 @@ public class TestExplicitColumnTracker {
*/
@Test
public void testInfiniteLoop() throws IOException {
- TreeSet<byte[]> columns = new TreeSet<byte[]>(Bytes.BYTES_COMPARATOR);
+ TreeSet<byte[]> columns = new TreeSet<>(Bytes.BYTES_COMPARATOR);
columns.addAll(Arrays.asList(new byte[][] { col2, col3, col5 }));
List<byte[]> scanner = Arrays.<byte[]> asList(new byte[][] { col1, col4 });
List<ScanQueryMatcher.MatchCode> expected = Arrays.<ScanQueryMatcher.MatchCode> asList(
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestScanWildcardColumnTracker.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestScanWildcardColumnTracker.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestScanWildcardColumnTracker.java
index 2852947..6d6e58e 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestScanWildcardColumnTracker.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestScanWildcardColumnTracker.java
@@ -44,20 +44,20 @@ public class TestScanWildcardColumnTracker {
ScanWildcardColumnTracker tracker = new ScanWildcardColumnTracker(0, VERSIONS, Long.MIN_VALUE);
// Create list of qualifiers
- List<byte[]> qualifiers = new ArrayList<byte[]>(4);
+ List<byte[]> qualifiers = new ArrayList<>(4);
qualifiers.add(Bytes.toBytes("qualifier1"));
qualifiers.add(Bytes.toBytes("qualifier2"));
qualifiers.add(Bytes.toBytes("qualifier3"));
qualifiers.add(Bytes.toBytes("qualifier4"));
// Setting up expected result
- List<MatchCode> expected = new ArrayList<MatchCode>(4);
+ List<MatchCode> expected = new ArrayList<>(4);
expected.add(ScanQueryMatcher.MatchCode.INCLUDE);
expected.add(ScanQueryMatcher.MatchCode.INCLUDE);
expected.add(ScanQueryMatcher.MatchCode.INCLUDE);
expected.add(ScanQueryMatcher.MatchCode.INCLUDE);
- List<ScanQueryMatcher.MatchCode> actual = new ArrayList<MatchCode>(qualifiers.size());
+ List<ScanQueryMatcher.MatchCode> actual = new ArrayList<>(qualifiers.size());
for (byte[] qualifier : qualifiers) {
ScanQueryMatcher.MatchCode mc = ScanQueryMatcher.checkColumn(tracker, qualifier, 0,
@@ -76,20 +76,20 @@ public class TestScanWildcardColumnTracker {
ScanWildcardColumnTracker tracker = new ScanWildcardColumnTracker(0, VERSIONS, Long.MIN_VALUE);
// Create list of qualifiers
- List<byte[]> qualifiers = new ArrayList<byte[]>(4);
+ List<byte[]> qualifiers = new ArrayList<>(4);
qualifiers.add(Bytes.toBytes("qualifier1"));
qualifiers.add(Bytes.toBytes("qualifier1"));
qualifiers.add(Bytes.toBytes("qualifier1"));
qualifiers.add(Bytes.toBytes("qualifier2"));
// Setting up expected result
- List<ScanQueryMatcher.MatchCode> expected = new ArrayList<MatchCode>(4);
+ List<ScanQueryMatcher.MatchCode> expected = new ArrayList<>(4);
expected.add(ScanQueryMatcher.MatchCode.INCLUDE);
expected.add(ScanQueryMatcher.MatchCode.INCLUDE);
expected.add(ScanQueryMatcher.MatchCode.SEEK_NEXT_COL);
expected.add(ScanQueryMatcher.MatchCode.INCLUDE);
- List<MatchCode> actual = new ArrayList<ScanQueryMatcher.MatchCode>(qualifiers.size());
+ List<MatchCode> actual = new ArrayList<>(qualifiers.size());
long timestamp = 0;
for (byte[] qualifier : qualifiers) {
@@ -109,7 +109,7 @@ public class TestScanWildcardColumnTracker {
ScanWildcardColumnTracker tracker = new ScanWildcardColumnTracker(0, VERSIONS, Long.MIN_VALUE);
// Create list of qualifiers
- List<byte[]> qualifiers = new ArrayList<byte[]>(2);
+ List<byte[]> qualifiers = new ArrayList<>(2);
qualifiers.add(Bytes.toBytes("qualifier2"));
qualifiers.add(Bytes.toBytes("qualifier1"));
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestUserScanQueryMatcher.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestUserScanQueryMatcher.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestUserScanQueryMatcher.java
index 0831404..b4e4311 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestUserScanQueryMatcher.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestUserScanQueryMatcher.java
@@ -69,7 +69,7 @@ public class TestUserScanQueryMatcher extends AbstractTestScanQueryMatcher {
// of just byte []
// Expected result
- List<MatchCode> expected = new ArrayList<ScanQueryMatcher.MatchCode>(6);
+ List<MatchCode> expected = new ArrayList<>(6);
expected.add(ScanQueryMatcher.MatchCode.SEEK_NEXT_COL);
expected.add(ScanQueryMatcher.MatchCode.INCLUDE_AND_SEEK_NEXT_COL);
expected.add(ScanQueryMatcher.MatchCode.SEEK_NEXT_COL);
@@ -83,7 +83,7 @@ public class TestUserScanQueryMatcher extends AbstractTestScanQueryMatcher {
new ScanInfo(this.conf, fam2, 0, 1, ttl, KeepDeletedCells.FALSE, 0, rowComparator),
get.getFamilyMap().get(fam2), now - ttl, now, null);
- List<KeyValue> memstore = new ArrayList<KeyValue>(6);
+ List<KeyValue> memstore = new ArrayList<>(6);
memstore.add(new KeyValue(row1, fam2, col1, 1, data));
memstore.add(new KeyValue(row1, fam2, col2, 1, data));
memstore.add(new KeyValue(row1, fam2, col3, 1, data));
@@ -92,7 +92,7 @@ public class TestUserScanQueryMatcher extends AbstractTestScanQueryMatcher {
memstore.add(new KeyValue(row2, fam1, col1, data));
- List<ScanQueryMatcher.MatchCode> actual = new ArrayList<ScanQueryMatcher.MatchCode>(memstore.size());
+ List<ScanQueryMatcher.MatchCode> actual = new ArrayList<>(memstore.size());
KeyValue k = memstore.get(0);
qm.setToNewRow(k);
@@ -113,7 +113,7 @@ public class TestUserScanQueryMatcher extends AbstractTestScanQueryMatcher {
// of just byte []
// Expected result
- List<MatchCode> expected = new ArrayList<ScanQueryMatcher.MatchCode>(6);
+ List<MatchCode> expected = new ArrayList<>(6);
expected.add(ScanQueryMatcher.MatchCode.INCLUDE);
expected.add(ScanQueryMatcher.MatchCode.INCLUDE);
expected.add(ScanQueryMatcher.MatchCode.INCLUDE);
@@ -126,7 +126,7 @@ public class TestUserScanQueryMatcher extends AbstractTestScanQueryMatcher {
new ScanInfo(this.conf, fam2, 0, 1, ttl, KeepDeletedCells.FALSE, 0, rowComparator), null,
now - ttl, now, null);
- List<KeyValue> memstore = new ArrayList<KeyValue>(6);
+ List<KeyValue> memstore = new ArrayList<>(6);
memstore.add(new KeyValue(row1, fam2, col1, 1, data));
memstore.add(new KeyValue(row1, fam2, col2, 1, data));
memstore.add(new KeyValue(row1, fam2, col3, 1, data));
@@ -134,7 +134,7 @@ public class TestUserScanQueryMatcher extends AbstractTestScanQueryMatcher {
memstore.add(new KeyValue(row1, fam2, col5, 1, data));
memstore.add(new KeyValue(row2, fam1, col1, 1, data));
- List<ScanQueryMatcher.MatchCode> actual = new ArrayList<ScanQueryMatcher.MatchCode>(memstore.size());
+ List<ScanQueryMatcher.MatchCode> actual = new ArrayList<>(memstore.size());
KeyValue k = memstore.get(0);
qm.setToNewRow(k);
@@ -181,7 +181,7 @@ public class TestUserScanQueryMatcher extends AbstractTestScanQueryMatcher {
KeyValue k = kvs[0];
qm.setToNewRow(k);
- List<MatchCode> actual = new ArrayList<MatchCode>(kvs.length);
+ List<MatchCode> actual = new ArrayList<>(kvs.length);
for (KeyValue kv : kvs) {
actual.add(qm.match(kv));
}
@@ -222,7 +222,7 @@ public class TestUserScanQueryMatcher extends AbstractTestScanQueryMatcher {
KeyValue k = kvs[0];
qm.setToNewRow(k);
- List<ScanQueryMatcher.MatchCode> actual = new ArrayList<ScanQueryMatcher.MatchCode>(kvs.length);
+ List<ScanQueryMatcher.MatchCode> actual = new ArrayList<>(kvs.length);
for (KeyValue kv : kvs) {
actual.add(qm.match(kv));
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestFSWAL.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestFSWAL.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestFSWAL.java
index fb0b514..0be7b31 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestFSWAL.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestFSWAL.java
@@ -255,11 +255,11 @@ public abstract class AbstractTestFSWAL {
new HRegionInfo(t2.getTableName(), HConstants.EMPTY_START_ROW, HConstants.EMPTY_END_ROW);
// add edits and roll the wal
MultiVersionConcurrencyControl mvcc = new MultiVersionConcurrencyControl();
- NavigableMap<byte[], Integer> scopes1 = new TreeMap<byte[], Integer>(Bytes.BYTES_COMPARATOR);
+ NavigableMap<byte[], Integer> scopes1 = new TreeMap<>(Bytes.BYTES_COMPARATOR);
for (byte[] fam : t1.getFamiliesKeys()) {
scopes1.put(fam, 0);
}
- NavigableMap<byte[], Integer> scopes2 = new TreeMap<byte[], Integer>(Bytes.BYTES_COMPARATOR);
+ NavigableMap<byte[], Integer> scopes2 = new TreeMap<>(Bytes.BYTES_COMPARATOR);
for (byte[] fam : t2.getFamiliesKeys()) {
scopes2.put(fam, 0);
}
@@ -362,7 +362,7 @@ public abstract class AbstractTestFSWAL {
HBaseTestingUtility.closeRegionAndWAL(r);
final int countPerFamily = 10;
final AtomicBoolean goslow = new AtomicBoolean(false);
- NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>(Bytes.BYTES_COMPARATOR);
+ NavigableMap<byte[], Integer> scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR);
for (byte[] fam : htd.getFamiliesKeys()) {
scopes.put(fam, 0);
}
@@ -402,7 +402,7 @@ public abstract class AbstractTestFSWAL {
}
}
// Add any old cluster id.
- List<UUID> clusterIds = new ArrayList<UUID>(1);
+ List<UUID> clusterIds = new ArrayList<>(1);
clusterIds.add(UUID.randomUUID());
// Now make appends run slow.
goslow.set(true);
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestLogRollPeriod.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestLogRollPeriod.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestLogRollPeriod.java
index f70bcc8..04a4bbc 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestLogRollPeriod.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestLogRollPeriod.java
@@ -130,7 +130,7 @@ public abstract class AbstractTestLogRollPeriod {
private void checkMinLogRolls(final WAL log, final int minRolls)
throws Exception {
- final List<Path> paths = new ArrayList<Path>();
+ final List<Path> paths = new ArrayList<>();
log.registerWALActionsListener(new WALActionsListener.Base() {
@Override
public void postLogRoll(Path oldFile, Path newFile) {
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestWALReplay.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestWALReplay.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestWALReplay.java
index 90eacf0..237d24a 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestWALReplay.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestWALReplay.java
@@ -299,8 +299,7 @@ public abstract class AbstractTestWALReplay {
// Add 1k to each family.
final int countPerFamily = 1000;
- NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>(
- Bytes.BYTES_COMPARATOR);
+ NavigableMap<byte[], Integer> scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR);
for(byte[] fam : htd.getFamiliesKeys()) {
scopes.put(fam, 0);
}
@@ -366,7 +365,7 @@ public abstract class AbstractTestWALReplay {
Path f = new Path(basedir, "hfile");
HFileTestUtil.createHFile(this.conf, fs, f, family, family, Bytes.toBytes(""),
Bytes.toBytes("z"), 10);
- List<Pair<byte[], String>> hfs = new ArrayList<Pair<byte[], String>>(1);
+ List<Pair<byte[], String>> hfs = new ArrayList<>(1);
hfs.add(Pair.newPair(family, f.toString()));
region.bulkLoadHFiles(hfs, true, null);
@@ -434,7 +433,7 @@ public abstract class AbstractTestWALReplay {
region.put((new Put(row)).addColumn(family, family, family));
wal.sync();
- List <Pair<byte[],String>> hfs= new ArrayList<Pair<byte[],String>>(1);
+ List <Pair<byte[],String>> hfs= new ArrayList<>(1);
for (int i = 0; i < 3; i++) {
Path f = new Path(basedir, "hfile"+i);
HFileTestUtil.createHFile(this.conf, fs, f, family, family, Bytes.toBytes(i + "00"),
@@ -700,8 +699,7 @@ public abstract class AbstractTestWALReplay {
HRegion region =
HRegion.openHRegion(this.hbaseRootDir, hri, htd, wal, customConf, rsServices, null);
int writtenRowCount = 10;
- List<HColumnDescriptor> families = new ArrayList<HColumnDescriptor>(
- htd.getFamilies());
+ List<HColumnDescriptor> families = new ArrayList<>(htd.getFamilies());
for (int i = 0; i < writtenRowCount; i++) {
Put put = new Put(Bytes.toBytes(tableName + Integer.toString(i)));
put.addColumn(families.get(i % families.size()).getName(), Bytes.toBytes("q"),
@@ -759,7 +757,7 @@ public abstract class AbstractTestWALReplay {
private int getScannedCount(RegionScanner scanner) throws IOException {
int scannedCount = 0;
- List<Cell> results = new ArrayList<Cell>();
+ List<Cell> results = new ArrayList<>();
while (true) {
boolean existMore = scanner.next(results);
if (!results.isEmpty())
@@ -794,9 +792,8 @@ public abstract class AbstractTestWALReplay {
// Add 1k to each family.
final int countPerFamily = 1000;
- Set<byte[]> familyNames = new HashSet<byte[]>();
- NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>(
- Bytes.BYTES_COMPARATOR);
+ Set<byte[]> familyNames = new HashSet<>();
+ NavigableMap<byte[], Integer> scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR);
for(byte[] fam : htd.getFamiliesKeys()) {
scopes.put(fam, 0);
}
@@ -1048,7 +1045,7 @@ public abstract class AbstractTestWALReplay {
deleteDir(basedir);
final HTableDescriptor htd = createBasic1FamilyHTD(tableName);
- NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>(Bytes.BYTES_COMPARATOR);
+ NavigableMap<byte[], Integer> scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR);
for (byte[] fam : htd.getFamiliesKeys()) {
scopes.put(fam, 0);
}
@@ -1196,7 +1193,7 @@ public abstract class AbstractTestWALReplay {
static List<Put> addRegionEdits(final byte[] rowName, final byte[] family, final int count,
EnvironmentEdge ee, final Region r, final String qualifierPrefix) throws IOException {
- List<Put> puts = new ArrayList<Put>();
+ List<Put> puts = new ArrayList<>();
for (int j = 0; j < count; j++) {
byte[] qualifier = Bytes.toBytes(qualifierPrefix + Integer.toString(j));
Put p = new Put(rowName);
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/FaultyProtobufLogReader.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/FaultyProtobufLogReader.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/FaultyProtobufLogReader.java
index c654c16..1af21d2 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/FaultyProtobufLogReader.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/FaultyProtobufLogReader.java
@@ -31,7 +31,7 @@ public class FaultyProtobufLogReader extends ProtobufLogReader {
BEGINNING, MIDDLE, END, NONE
}
- Queue<Entry> nextQueue = new LinkedList<Entry>();
+ Queue<Entry> nextQueue = new LinkedList<>();
int numberOfFileEntries = 0;
FailureType getFailureType() {
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestFSHLog.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestFSHLog.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestFSHLog.java
index 9e546e6..8847c4c 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestFSHLog.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestFSHLog.java
@@ -102,7 +102,7 @@ public class TestFSHLog extends AbstractTestFSWAL {
syncRunnerIndexField.set(ringBufferEventHandler, Integer.MAX_VALUE - 1);
HTableDescriptor htd =
new HTableDescriptor(TableName.valueOf(this.name.getMethodName())).addFamily(new HColumnDescriptor("row"));
- NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>(Bytes.BYTES_COMPARATOR);
+ NavigableMap<byte[], Integer> scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR);
for (byte[] fam : htd.getFamiliesKeys()) {
scopes.put(fam, 0);
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestKeyValueCompression.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestKeyValueCompression.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestKeyValueCompression.java
index 104f897..4a256a6 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestKeyValueCompression.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestKeyValueCompression.java
@@ -107,7 +107,7 @@ public class TestKeyValueCompression {
byte[] cf = Bytes.toBytes("myCF");
byte[] q = Bytes.toBytes("myQualifier");
byte[] value = Bytes.toBytes("myValue");
- List<Tag> tags = new ArrayList<Tag>(noOfTags);
+ List<Tag> tags = new ArrayList<>(noOfTags);
for (int i = 1; i <= noOfTags; i++) {
tags.add(new ArrayBackedTag((byte) i, Bytes.toBytes("tagValue" + i)));
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollAbort.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollAbort.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollAbort.java
index 22395c8..ccb00c7 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollAbort.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollAbort.java
@@ -199,8 +199,7 @@ public class TestLogRollAbort {
kvs.add(new KeyValue(Bytes.toBytes(i), tableName.getName(), tableName.getName()));
HTableDescriptor htd = new HTableDescriptor(tableName);
htd.addFamily(new HColumnDescriptor("column"));
- NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>(
- Bytes.BYTES_COMPARATOR);
+ NavigableMap<byte[], Integer> scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR);
for(byte[] fam : htd.getFamiliesKeys()) {
scopes.put(fam, 0);
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRolling.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRolling.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRolling.java
index 19c534e..5bc4c9b 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRolling.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRolling.java
@@ -245,8 +245,8 @@ public class TestLogRolling extends AbstractTestLogRolling {
server = TEST_UTIL.getRSForFirstRegionInTable(desc.getTableName());
HRegionInfo region = server.getOnlineRegions(desc.getTableName()).get(0).getRegionInfo();
final WAL log = server.getWAL(region);
- final List<Path> paths = new ArrayList<Path>(1);
- final List<Integer> preLogRolledCalled = new ArrayList<Integer>();
+ final List<Path> paths = new ArrayList<>(1);
+ final List<Integer> preLogRolledCalled = new ArrayList<>();
paths.add(AbstractFSWALProvider.getCurrentFileName(log));
log.registerWALActionsListener(new WALActionsListener.Base() {
@@ -307,7 +307,7 @@ public class TestLogRolling extends AbstractTestLogRolling {
preLogRolledCalled.size() >= 1);
// read back the data written
- Set<String> loggedRows = new HashSet<String>();
+ Set<String> loggedRows = new HashSet<>();
FSUtils fsUtils = FSUtils.getInstance(fs, TEST_UTIL.getConfiguration());
for (Path p : paths) {
LOG.debug("recovering lease for " + p);
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollingNoCluster.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollingNoCluster.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollingNoCluster.java
index 7412128..d3d582c 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollingNoCluster.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollingNoCluster.java
@@ -157,8 +157,7 @@ public class TestLogRollingNoCluster {
edit.add(new KeyValue(bytes, bytes, bytes, now, EMPTY_1K_ARRAY));
final HRegionInfo hri = HRegionInfo.FIRST_META_REGIONINFO;
final HTableDescriptor htd = TEST_UTIL.getMetaTableDescriptor();
- NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>(
- Bytes.BYTES_COMPARATOR);
+ NavigableMap<byte[], Integer> scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR);
for(byte[] fam : htd.getFamiliesKeys()) {
scopes.put(fam, 0);
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestSequenceIdAccounting.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestSequenceIdAccounting.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestSequenceIdAccounting.java
index 9fd0cb1..9f5acbd 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestSequenceIdAccounting.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestSequenceIdAccounting.java
@@ -38,7 +38,7 @@ public class TestSequenceIdAccounting {
private static final byte [] FAMILY_NAME = Bytes.toBytes("cf");
private static final Set<byte[]> FAMILIES;
static {
- FAMILIES = new HashSet<byte[]>();
+ FAMILIES = new HashSet<>();
FAMILIES.add(FAMILY_NAME);
}
@@ -46,7 +46,7 @@ public class TestSequenceIdAccounting {
public void testStartCacheFlush() {
SequenceIdAccounting sida = new SequenceIdAccounting();
sida.getOrCreateLowestSequenceIds(ENCODED_REGION_NAME);
- Map<byte[], Long> m = new HashMap<byte[], Long>();
+ Map<byte[], Long> m = new HashMap<>();
m.put(ENCODED_REGION_NAME, HConstants.NO_SEQNUM);
assertEquals(HConstants.NO_SEQNUM, (long)sida.startCacheFlush(ENCODED_REGION_NAME, FAMILIES));
sida.completeCacheFlush(ENCODED_REGION_NAME);
@@ -57,7 +57,7 @@ public class TestSequenceIdAccounting {
sida.completeCacheFlush(ENCODED_REGION_NAME);
long currentSequenceId = sequenceid;
sida.update(ENCODED_REGION_NAME, FAMILIES, sequenceid, true);
- final Set<byte[]> otherFamily = new HashSet<byte[]>(1);
+ final Set<byte[]> otherFamily = new HashSet<>(1);
otherFamily.add(Bytes.toBytes("otherCf"));
sida.update(ENCODED_REGION_NAME, FAMILIES, ++sequenceid, true);
// Should return oldest sequence id in the region.
@@ -69,7 +69,7 @@ public class TestSequenceIdAccounting {
public void testAreAllLower() {
SequenceIdAccounting sida = new SequenceIdAccounting();
sida.getOrCreateLowestSequenceIds(ENCODED_REGION_NAME);
- Map<byte[], Long> m = new HashMap<byte[], Long>();
+ Map<byte[], Long> m = new HashMap<>();
m.put(ENCODED_REGION_NAME, HConstants.NO_SEQNUM);
assertTrue(sida.areAllLower(m));
long sequenceid = 1;
@@ -117,7 +117,7 @@ public class TestSequenceIdAccounting {
public void testFindLower() {
SequenceIdAccounting sida = new SequenceIdAccounting();
sida.getOrCreateLowestSequenceIds(ENCODED_REGION_NAME);
- Map<byte[], Long> m = new HashMap<byte[], Long>();
+ Map<byte[], Long> m = new HashMap<>();
m.put(ENCODED_REGION_NAME, HConstants.NO_SEQNUM);
long sequenceid = 1;
sida.update(ENCODED_REGION_NAME, FAMILIES, sequenceid, true);
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALActionsListener.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALActionsListener.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALActionsListener.java
index 9ac9f0e..9f9e2df 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALActionsListener.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALActionsListener.java
@@ -95,7 +95,7 @@ public class TestWALActionsListener {
@Test
public void testActionListener() throws Exception {
DummyWALActionsListener observer = new DummyWALActionsListener();
- List<WALActionsListener> list = new ArrayList<WALActionsListener>(1);
+ List<WALActionsListener> list = new ArrayList<>(1);
list.add(observer);
final WALFactory wals = new WALFactory(conf, list, "testActionListener");
DummyWALActionsListener laterobserver = new DummyWALActionsListener();
@@ -110,8 +110,7 @@ public class TestWALActionsListener {
edit.add(kv);
HTableDescriptor htd = new HTableDescriptor(TableName.valueOf(SOME_BYTES));
htd.addFamily(new HColumnDescriptor(b));
- NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>(
- Bytes.BYTES_COMPARATOR);
+ NavigableMap<byte[], Integer> scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR);
for(byte[] fam : htd.getFamiliesKeys()) {
scopes.put(fam, 0);
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALCellCodecWithCompression.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALCellCodecWithCompression.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALCellCodecWithCompression.java
index c4329b8..8a246be 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALCellCodecWithCompression.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALCellCodecWithCompression.java
@@ -102,7 +102,7 @@ public class TestWALCellCodecWithCompression {
byte[] cf = Bytes.toBytes("myCF");
byte[] q = Bytes.toBytes("myQualifier");
byte[] value = Bytes.toBytes("myValue");
- List<Tag> tags = new ArrayList<Tag>(noOfTags);
+ List<Tag> tags = new ArrayList<>(noOfTags);
for (int i = 1; i <= noOfTags; i++) {
tags.add(new ArrayBackedTag((byte) i, Bytes.toBytes("tagValue" + i)));
}
@@ -114,7 +114,7 @@ public class TestWALCellCodecWithCompression {
byte[] cf = Bytes.toBytes("myCF");
byte[] q = Bytes.toBytes("myQualifier");
byte[] value = Bytes.toBytes("myValue");
- List<Tag> tags = new ArrayList<Tag>(noOfTags);
+ List<Tag> tags = new ArrayList<>(noOfTags);
for (int i = 1; i <= noOfTags; i++) {
tags.add(new ArrayBackedTag((byte) i, Bytes.toBytes("tagValue" + i)));
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestNamespaceReplication.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestNamespaceReplication.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestNamespaceReplication.java
index 3814562..e296f87 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestNamespaceReplication.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestNamespaceReplication.java
@@ -165,7 +165,7 @@ public class TestNamespaceReplication extends TestReplicationBase {
namespaces.add(ns2);
rpc.setNamespaces(namespaces);
Map<TableName, List<String>> tableCfs = new HashMap<>();
- tableCfs.put(tabAName, new ArrayList<String>());
+ tableCfs.put(tabAName, new ArrayList<>());
tableCfs.get(tabAName).add("f1");
rpc.setTableCFsMap(tableCfs);
admin.updatePeerConfig("2", rpc);
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestPerTableCFReplication.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestPerTableCFReplication.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestPerTableCFReplication.java
index a7a4cd8..abf2db3 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestPerTableCFReplication.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestPerTableCFReplication.java
@@ -283,7 +283,7 @@ public class TestPerTableCFReplication {
// 1. null or empty string, result should be null
assertNull(ReplicationSerDeHelper.convert(tabCFsMap));
- tabCFsMap = new HashMap<TableName, List<String>>();
+ tabCFsMap = new HashMap<>();
tableCFs = ReplicationSerDeHelper.convert(tabCFsMap);
assertEquals(0, tableCFs.length);
@@ -301,7 +301,7 @@ public class TestPerTableCFReplication {
assertEquals(0, tableCFs[0].getFamiliesCount());
tabCFsMap.clear();
- tabCFsMap.put(tableName2, new ArrayList<String>());
+ tabCFsMap.put(tableName2, new ArrayList<>());
tabCFsMap.get(tableName2).add("cf1");
tableCFs = ReplicationSerDeHelper.convert(tabCFsMap);
assertEquals(1, tableCFs.length); // only one table
@@ -311,7 +311,7 @@ public class TestPerTableCFReplication {
assertEquals("cf1", tableCFs[0].getFamilies(0).toStringUtf8());
tabCFsMap.clear();
- tabCFsMap.put(tableName3, new ArrayList<String>());
+ tabCFsMap.put(tableName3, new ArrayList<>());
tabCFsMap.get(tableName3).add("cf1");
tabCFsMap.get(tableName3).add("cf3");
tableCFs = ReplicationSerDeHelper.convert(tabCFsMap);
@@ -324,9 +324,9 @@ public class TestPerTableCFReplication {
tabCFsMap.clear();
tabCFsMap.put(tableName1, null);
- tabCFsMap.put(tableName2, new ArrayList<String>());
+ tabCFsMap.put(tableName2, new ArrayList<>());
tabCFsMap.get(tableName2).add("cf1");
- tabCFsMap.put(tableName3, new ArrayList<String>());
+ tabCFsMap.put(tableName3, new ArrayList<>());
tabCFsMap.get(tableName3).add("cf1");
tabCFsMap.get(tableName3).add("cf3");
@@ -406,7 +406,7 @@ public class TestPerTableCFReplication {
rpc2.setClusterKey(utility2.getClusterKey());
Map<TableName, List<String>> tableCFs = new HashMap<>();
tableCFs.put(tabCName, null);
- tableCFs.put(tabBName, new ArrayList<String>());
+ tableCFs.put(tabBName, new ArrayList<>());
tableCFs.get(tabBName).add("f1");
tableCFs.get(tabBName).add("f3");
replicationAdmin.addPeer("2", rpc2, tableCFs);
@@ -415,7 +415,7 @@ public class TestPerTableCFReplication {
rpc3.setClusterKey(utility3.getClusterKey());
tableCFs.clear();
tableCFs.put(tabAName, null);
- tableCFs.put(tabBName, new ArrayList<String>());
+ tableCFs.put(tabBName, new ArrayList<>());
tableCFs.get(tabBName).add("f1");
tableCFs.get(tabBName).add("f2");
replicationAdmin.addPeer("3", rpc3, tableCFs);
@@ -462,17 +462,17 @@ public class TestPerTableCFReplication {
// B. change peers' replicable table-cf config
tableCFs.clear();
- tableCFs.put(tabAName, new ArrayList<String>());
+ tableCFs.put(tabAName, new ArrayList<>());
tableCFs.get(tabAName).add("f1");
tableCFs.get(tabAName).add("f2");
- tableCFs.put(tabCName, new ArrayList<String>());
+ tableCFs.put(tabCName, new ArrayList<>());
tableCFs.get(tabCName).add("f2");
tableCFs.get(tabCName).add("f3");
replicationAdmin.setPeerTableCFs("2", tableCFs);
tableCFs.clear();
tableCFs.put(tabBName, null);
- tableCFs.put(tabCName, new ArrayList<String>());
+ tableCFs.put(tabCName, new ArrayList<>());
tableCFs.get(tabCName).add("f3");
replicationAdmin.setPeerTableCFs("3", tableCFs);
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationBase.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationBase.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationBase.java
index 474039b..caad544 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationBase.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationBase.java
@@ -142,8 +142,7 @@ public class TestReplicationBase {
table.addFamily(fam);
fam = new HColumnDescriptor(noRepfamName);
table.addFamily(fam);
- scopes = new TreeMap<byte[], Integer>(
- Bytes.BYTES_COMPARATOR);
+ scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR);
for(HColumnDescriptor f : table.getColumnFamilies()) {
scopes.put(f.getName(), f.getScope());
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationEndpoint.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationEndpoint.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationEndpoint.java
index 5e8d569..4925aab 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationEndpoint.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationEndpoint.java
@@ -106,7 +106,7 @@ public class TestReplicationEndpoint extends TestReplicationBase {
@Override
public String explainFailure() throws Exception {
- List<String> logRollInProgressRsList = new ArrayList<String>();
+ List<String> logRollInProgressRsList = new ArrayList<>();
for (RegionServerThread rs : rsThreads) {
if (!rs.getRegionServer().walRollRequestFinished()) {
logRollInProgressRsList.add(rs.getRegionServer().toString());
@@ -462,7 +462,7 @@ public class TestReplicationEndpoint extends TestReplicationBase {
public static class ReplicationEndpointReturningFalse extends ReplicationEndpointForTest {
static int COUNT = 10;
- static AtomicReference<Exception> ex = new AtomicReference<Exception>(null);
+ static AtomicReference<Exception> ex = new AtomicReference<>(null);
static AtomicBoolean replicated = new AtomicBoolean(false);
@Override
public boolean replicate(ReplicateContext replicateContext) {
@@ -483,7 +483,7 @@ public class TestReplicationEndpoint extends TestReplicationBase {
// return a WALEntry filter which only accepts "row", but not other rows
public static class ReplicationEndpointWithWALEntryFilter extends ReplicationEndpointForTest {
- static AtomicReference<Exception> ex = new AtomicReference<Exception>(null);
+ static AtomicReference<Exception> ex = new AtomicReference<>(null);
@Override
public boolean replicate(ReplicateContext replicateContext) {
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.java
index 9536f9f..1c5a994 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.java
@@ -413,7 +413,7 @@ public class TestReplicationSmallTests extends TestReplicationBase {
@Test(timeout=300000)
public void testLoading() throws Exception {
LOG.info("Writing out rows to table1 in testLoading");
- List<Put> puts = new ArrayList<Put>(NB_ROWS_IN_BIG_BATCH);
+ List<Put> puts = new ArrayList<>(NB_ROWS_IN_BIG_BATCH);
for (int i = 0; i < NB_ROWS_IN_BIG_BATCH; i++) {
Put put = new Put(Bytes.toBytes(i));
put.addColumn(famName, row, row);
@@ -519,8 +519,7 @@ public class TestReplicationSmallTests extends TestReplicationBase {
fam.setMaxVersions(100);
fam.setScope(HConstants.REPLICATION_SCOPE_GLOBAL);
table.addFamily(fam);
- scopes = new TreeMap<byte[], Integer>(
- Bytes.BYTES_COMPARATOR);
+ scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR);
for (HColumnDescriptor f : table.getColumnFamilies()) {
scopes.put(f.getName(), f.getScope());
}
@@ -818,7 +817,7 @@ public class TestReplicationSmallTests extends TestReplicationBase {
HRegion region = utility1.getMiniHBaseCluster().getRegions(tableName).get(0);
HRegionInfo hri = region.getRegionInfo();
- NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>(Bytes.BYTES_COMPARATOR);
+ NavigableMap<byte[], Integer> scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR);
for (byte[] fam : htable1.getTableDescriptor().getFamiliesKeys()) {
scopes.put(fam, 1);
}