You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2013/01/19 23:45:12 UTC
svn commit: r1435747 [4/4] - in /hbase/trunk:
hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/
hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/
hbase-protocol/src/main/protobuf/ hbase-server/src/main/jav...
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompaction.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompaction.java?rev=1435747&r1=1435746&r2=1435747&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompaction.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompaction.java Sat Jan 19 22:45:11 2013
@@ -127,7 +127,7 @@ public class TestCompaction extends HBas
do {
List<KeyValue> results = new ArrayList<KeyValue>();
boolean result = s.next(results);
- r.delete(new Delete(results.get(0).getRow()), null, false);
+ r.delete(new Delete(results.get(0).getRow()), false);
if (!result) break;
} while(true);
s.close();
@@ -199,7 +199,7 @@ public class TestCompaction extends HBas
// Default is that there only 3 (MAXVERSIONS) versions allowed per column.
//
// Assert == 3 when we ask for versions.
- Result result = r.get(new Get(STARTROW).addFamily(COLUMN_FAMILY_TEXT).setMaxVersions(100), null);
+ Result result = r.get(new Get(STARTROW).addFamily(COLUMN_FAMILY_TEXT).setMaxVersions(100));
assertEquals(compactionThreshold, result.size());
// see if CompactionProgress is in place but null
@@ -229,7 +229,7 @@ public class TestCompaction extends HBas
// Always 3 versions if that is what max versions is.
result = r.get(new Get(secondRowBytes).addFamily(COLUMN_FAMILY_TEXT).
- setMaxVersions(100), null);
+ setMaxVersions(100));
LOG.debug("Row " + Bytes.toStringBinary(secondRowBytes) + " after " +
"initial compaction: " + result);
assertEquals("Invalid number of versions of row "
@@ -242,32 +242,32 @@ public class TestCompaction extends HBas
// should result in a compacted store file that has no references to the
// deleted row.
LOG.debug("Adding deletes to memstore and flushing");
- Delete delete = new Delete(secondRowBytes, System.currentTimeMillis(), null);
+ Delete delete = new Delete(secondRowBytes, System.currentTimeMillis());
byte [][] famAndQf = {COLUMN_FAMILY, null};
delete.deleteFamily(famAndQf[0]);
- r.delete(delete, null, true);
+ r.delete(delete, true);
// Assert deleted.
- result = r.get(new Get(secondRowBytes).addFamily(COLUMN_FAMILY_TEXT).setMaxVersions(100), null );
+ result = r.get(new Get(secondRowBytes).addFamily(COLUMN_FAMILY_TEXT).setMaxVersions(100));
assertTrue("Second row should have been deleted", result.isEmpty());
r.flushcache();
- result = r.get(new Get(secondRowBytes).addFamily(COLUMN_FAMILY_TEXT).setMaxVersions(100), null );
+ result = r.get(new Get(secondRowBytes).addFamily(COLUMN_FAMILY_TEXT).setMaxVersions(100));
assertTrue("Second row should have been deleted", result.isEmpty());
// Add a bit of data and flush. Start adding at 'bbb'.
createSmallerStoreFile(this.r);
r.flushcache();
// Assert that the second row is still deleted.
- result = r.get(new Get(secondRowBytes).addFamily(COLUMN_FAMILY_TEXT).setMaxVersions(100), null );
+ result = r.get(new Get(secondRowBytes).addFamily(COLUMN_FAMILY_TEXT).setMaxVersions(100));
assertTrue("Second row should still be deleted", result.isEmpty());
// Force major compaction.
r.compactStores(true);
assertEquals(r.getStore(COLUMN_FAMILY_TEXT).getStorefiles().size(), 1);
- result = r.get(new Get(secondRowBytes).addFamily(COLUMN_FAMILY_TEXT).setMaxVersions(100), null );
+ result = r.get(new Get(secondRowBytes).addFamily(COLUMN_FAMILY_TEXT).setMaxVersions(100));
assertTrue("Second row should still be deleted", result.isEmpty());
// Make sure the store files do have some 'aaa' keys in them -- exactly 3.
@@ -406,22 +406,22 @@ public class TestCompaction extends HBas
r.flushcache();
}
- Result result = r.get(new Get(firstRowBytes).addColumn(fam1, col1).setMaxVersions(100), null);
+ Result result = r.get(new Get(firstRowBytes).addColumn(fam1, col1).setMaxVersions(100));
assertEquals(compactionThreshold, result.size());
- result = r.get(new Get(secondRowBytes).addColumn(fam2, col2).setMaxVersions(100), null);
+ result = r.get(new Get(secondRowBytes).addColumn(fam2, col2).setMaxVersions(100));
assertEquals(compactionThreshold, result.size());
// Now add deletes to memstore and then flush it. That will put us over
// the compaction threshold of 3 store files. Compacting these store files
// should result in a compacted store file that has no references to the
// deleted row.
- r.delete(delete, null, true);
+ r.delete(delete, true);
// Make sure that we have only deleted family2 from secondRowBytes
- result = r.get(new Get(secondRowBytes).addColumn(fam2, col2).setMaxVersions(100), null);
+ result = r.get(new Get(secondRowBytes).addColumn(fam2, col2).setMaxVersions(100));
assertEquals(expectedResultsAfterDelete, result.size());
// but we still have firstrow
- result = r.get(new Get(firstRowBytes).addColumn(fam1, col1).setMaxVersions(100), null);
+ result = r.get(new Get(firstRowBytes).addColumn(fam1, col1).setMaxVersions(100));
assertEquals(compactionThreshold, result.size());
r.flushcache();
@@ -429,10 +429,10 @@ public class TestCompaction extends HBas
// Let us check again
// Make sure that we have only deleted family2 from secondRowBytes
- result = r.get(new Get(secondRowBytes).addColumn(fam2, col2).setMaxVersions(100), null);
+ result = r.get(new Get(secondRowBytes).addColumn(fam2, col2).setMaxVersions(100));
assertEquals(expectedResultsAfterDelete, result.size());
// but we still have firstrow
- result = r.get(new Get(firstRowBytes).addColumn(fam1, col1).setMaxVersions(100), null);
+ result = r.get(new Get(firstRowBytes).addColumn(fam1, col1).setMaxVersions(100));
assertEquals(compactionThreshold, result.size());
// do a compaction
@@ -447,10 +447,10 @@ public class TestCompaction extends HBas
assertTrue("Was not supposed to be a major compaction", numFiles2 > 1);
// Make sure that we have only deleted family2 from secondRowBytes
- result = r.get(new Get(secondRowBytes).addColumn(fam2, col2).setMaxVersions(100), null);
+ result = r.get(new Get(secondRowBytes).addColumn(fam2, col2).setMaxVersions(100));
assertEquals(expectedResultsAfterDelete, result.size());
// but we still have firstrow
- result = r.get(new Get(firstRowBytes).addColumn(fam1, col1).setMaxVersions(100), null);
+ result = r.get(new Get(firstRowBytes).addColumn(fam1, col1).setMaxVersions(100));
assertEquals(compactionThreshold, result.size());
}
@@ -530,7 +530,7 @@ public class TestCompaction extends HBas
Delete delete = new Delete(Bytes.add(STARTROW, Bytes.toBytes(i)));
byte [][] famAndQf = {COLUMN_FAMILY, null};
delete.deleteFamily(famAndQf[0]);
- r.delete(delete, null, true);
+ r.delete(delete, true);
}
r.flushcache();
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestGetClosestAtOrBefore.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestGetClosestAtOrBefore.java?rev=1435747&r1=1435746&r2=1435747&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestGetClosestAtOrBefore.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestGetClosestAtOrBefore.java Sat Jan 19 22:45:11 2013
@@ -112,7 +112,7 @@ public class TestGetClosestAtOrBefore ex
try {
List<KeyValue> keys = new ArrayList<KeyValue>();
while (s.next(keys)) {
- mr.delete(new Delete(keys.get(0).getRow()), null, false);
+ mr.delete(new Delete(keys.get(0).getRow()), false);
keys.clear();
}
} finally {
@@ -207,7 +207,7 @@ public class TestGetClosestAtOrBefore ex
Delete d = new Delete(T20);
d.deleteColumn(c0, c0);
- region.delete(d, null, false);
+ region.delete(d, false);
r = region.getClosestRowBefore(T20, c0);
assertTrue(Bytes.equals(T10, r.getRow()));
@@ -221,7 +221,7 @@ public class TestGetClosestAtOrBefore ex
d = new Delete(T30);
d.deleteColumn(c0, c0);
- region.delete(d, null, false);
+ region.delete(d, false);
r = region.getClosestRowBefore(T30, c0);
assertTrue(Bytes.equals(T10, r.getRow()));
@@ -257,7 +257,7 @@ public class TestGetClosestAtOrBefore ex
// in memory; make sure we get back t10 again.
d = new Delete(T20);
d.deleteColumn(c1, c1);
- region.delete(d, null, false);
+ region.delete(d, false);
r = region.getClosestRowBefore(T30, c0);
assertTrue(Bytes.equals(T10, r.getRow()));
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHBase7051.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHBase7051.java?rev=1435747&r1=1435746&r2=1435747&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHBase7051.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHBase7051.java Sat Jan 19 22:45:11 2013
@@ -130,7 +130,7 @@ public class TestHBase7051 {
}
testStep = TestStep.CHECKANDPUT_STARTED;
region.checkAndMutate(Bytes.toBytes("r1"), Bytes.toBytes(family), Bytes.toBytes("q1"),
- CompareOp.EQUAL, new BinaryComparator(Bytes.toBytes("10")), put, null, true);
+ CompareOp.EQUAL, new BinaryComparator(Bytes.toBytes("10")), put, true);
testStep = TestStep.CHECKANDPUT_COMPLETED;
}
}
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java?rev=1435747&r1=1435746&r2=1435747&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java Sat Jan 19 22:45:11 2013
@@ -169,7 +169,7 @@ public class TestHRegion extends HBaseTe
RegionScanner scanner1 = region.getScanner(scan);
Delete delete = new Delete(Bytes.toBytes("r1"));
- region.delete(delete, null, false);
+ region.delete(delete, false);
region.flushcache();
// open the second scanner
@@ -274,7 +274,7 @@ public class TestHRegion extends HBaseTe
long seqId = region.replayRecoveredEditsIfAny(regiondir, maxSeqIdInStores, null, status);
assertEquals(maxSeqId, seqId);
Get get = new Get(row);
- Result result = region.get(get, null);
+ Result result = region.get(get);
for (long i = minSeqId; i <= maxSeqId; i += 10) {
List<KeyValue> kvs = result.getColumn(family, Bytes.toBytes(i));
assertEquals(1, kvs.size());
@@ -327,7 +327,7 @@ public class TestHRegion extends HBaseTe
long seqId = region.replayRecoveredEditsIfAny(regiondir, maxSeqIdInStores, null, status);
assertEquals(maxSeqId, seqId);
Get get = new Get(row);
- Result result = region.get(get, null);
+ Result result = region.get(get);
for (long i = minSeqId; i <= maxSeqId; i += 10) {
List<KeyValue> kvs = result.getColumn(family, Bytes.toBytes(i));
if (i < recoverSeqId) {
@@ -461,7 +461,7 @@ public class TestHRegion extends HBaseTe
public void run() {
while (!this.done.get()) {
try {
- assertTrue(region.get(g, null).size() > 0);
+ assertTrue(region.get(g).size() > 0);
this.count.incrementAndGet();
} catch (Exception e) {
this.e = e;
@@ -532,7 +532,7 @@ public class TestHRegion extends HBaseTe
break;
Delete delete = new Delete(results.get(0).getRow());
delete.deleteColumn(Bytes.toBytes("trans-tags"), Bytes.toBytes("qual2"));
- r.delete(delete, null, false);
+ r.delete(delete, false);
results.clear();
} while (more);
assertEquals("Did not perform correct number of deletes", 3, count);
@@ -771,7 +771,6 @@ public class TestHRegion extends HBaseTe
byte [] emptyVal = new byte[] {};
byte [] val1 = Bytes.toBytes("value1");
byte [] val2 = Bytes.toBytes("value2");
- Integer lockId = null;
//Setting up region
String method = this.getName();
@@ -783,7 +782,7 @@ public class TestHRegion extends HBaseTe
//checkAndPut with empty value
boolean res = region.checkAndMutate(row1, fam1, qf1, CompareOp.EQUAL,
- new BinaryComparator(emptyVal), put, lockId, true);
+ new BinaryComparator(emptyVal), put, true);
assertTrue(res);
//Putting data in key
@@ -792,25 +791,25 @@ public class TestHRegion extends HBaseTe
//checkAndPut with correct value
res = region.checkAndMutate(row1, fam1, qf1, CompareOp.EQUAL,
- new BinaryComparator(emptyVal), put, lockId, true);
+ new BinaryComparator(emptyVal), put, true);
assertTrue(res);
// not empty anymore
res = region.checkAndMutate(row1, fam1, qf1, CompareOp.EQUAL,
- new BinaryComparator(emptyVal), put, lockId, true);
+ new BinaryComparator(emptyVal), put, true);
assertFalse(res);
Delete delete = new Delete(row1);
delete.deleteColumn(fam1, qf1);
res = region.checkAndMutate(row1, fam1, qf1, CompareOp.EQUAL,
- new BinaryComparator(emptyVal), delete, lockId, true);
+ new BinaryComparator(emptyVal), delete, true);
assertFalse(res);
put = new Put(row1);
put.add(fam1, qf1, val2);
//checkAndPut with correct value
res = region.checkAndMutate(row1, fam1, qf1, CompareOp.EQUAL,
- new BinaryComparator(val1), put, lockId, true);
+ new BinaryComparator(val1), put, true);
assertTrue(res);
//checkAndDelete with correct value
@@ -818,12 +817,12 @@ public class TestHRegion extends HBaseTe
delete.deleteColumn(fam1, qf1);
delete.deleteColumn(fam1, qf1);
res = region.checkAndMutate(row1, fam1, qf1, CompareOp.EQUAL,
- new BinaryComparator(val2), delete, lockId, true);
+ new BinaryComparator(val2), delete, true);
assertTrue(res);
delete = new Delete(row1);
res = region.checkAndMutate(row1, fam1, qf1, CompareOp.EQUAL,
- new BinaryComparator(emptyVal), delete, lockId, true);
+ new BinaryComparator(emptyVal), delete, true);
assertTrue(res);
//checkAndPut looking for a null value
@@ -831,7 +830,7 @@ public class TestHRegion extends HBaseTe
put.add(fam1, qf1, val1);
res = region.checkAndMutate(row1, fam1, qf1, CompareOp.EQUAL,
- new NullComparator(), put, lockId, true);
+ new NullComparator(), put, true);
assertTrue(res);
} finally {
HRegion.closeHRegion(this.region);
@@ -846,7 +845,6 @@ public class TestHRegion extends HBaseTe
byte [] qf1 = Bytes.toBytes("qualifier");
byte [] val1 = Bytes.toBytes("value1");
byte [] val2 = Bytes.toBytes("value2");
- Integer lockId = null;
//Setting up region
String method = this.getName();
@@ -859,14 +857,14 @@ public class TestHRegion extends HBaseTe
//checkAndPut with wrong value
boolean res = region.checkAndMutate(row1, fam1, qf1, CompareOp.EQUAL,
- new BinaryComparator(val2), put, lockId, true);
+ new BinaryComparator(val2), put, true);
assertEquals(false, res);
//checkAndDelete with wrong value
Delete delete = new Delete(row1);
delete.deleteFamily(fam1);
res = region.checkAndMutate(row1, fam1, qf1, CompareOp.EQUAL,
- new BinaryComparator(val2), delete, lockId, true);
+ new BinaryComparator(val2), delete, true);
assertEquals(false, res);
} finally {
HRegion.closeHRegion(this.region);
@@ -880,7 +878,6 @@ public class TestHRegion extends HBaseTe
byte [] fam1 = Bytes.toBytes("fam1");
byte [] qf1 = Bytes.toBytes("qualifier");
byte [] val1 = Bytes.toBytes("value1");
- Integer lockId = null;
//Setting up region
String method = this.getName();
@@ -893,14 +890,14 @@ public class TestHRegion extends HBaseTe
//checkAndPut with correct value
boolean res = region.checkAndMutate(row1, fam1, qf1, CompareOp.EQUAL,
- new BinaryComparator(val1), put, lockId, true);
+ new BinaryComparator(val1), put, true);
assertEquals(true, res);
//checkAndDelete with correct value
Delete delete = new Delete(row1);
delete.deleteColumn(fam1, qf1);
res = region.checkAndMutate(row1, fam1, qf1, CompareOp.EQUAL,
- new BinaryComparator(val1), put, lockId, true);
+ new BinaryComparator(val1), put, true);
assertEquals(true, res);
} finally {
HRegion.closeHRegion(this.region);
@@ -916,7 +913,6 @@ public class TestHRegion extends HBaseTe
byte [] qf1 = Bytes.toBytes("qualifier");
byte [] val1 = Bytes.toBytes("value1");
byte [] val2 = Bytes.toBytes("value2");
- Integer lockId = null;
byte [][] families = {fam1, fam2};
@@ -940,13 +936,13 @@ public class TestHRegion extends HBaseTe
store.memstore.kvset.size();
boolean res = region.checkAndMutate(row1, fam1, qf1, CompareOp.EQUAL,
- new BinaryComparator(val1), put, lockId, true);
+ new BinaryComparator(val1), put, true);
assertEquals(true, res);
store.memstore.kvset.size();
Get get = new Get(row1);
get.addColumn(fam2, qf1);
- KeyValue [] actual = region.get(get, null).raw();
+ KeyValue [] actual = region.get(get).raw();
KeyValue [] expected = {kv};
@@ -967,7 +963,7 @@ public class TestHRegion extends HBaseTe
put.add(fam1, qual1, value1);
try {
boolean res = region.checkAndMutate(row, fam1, qual1, CompareOp.EQUAL,
- new BinaryComparator(value2), put, null, false);
+ new BinaryComparator(value2), put, false);
fail();
} catch (DoNotRetryIOException expected) {
// expected exception.
@@ -990,7 +986,6 @@ public class TestHRegion extends HBaseTe
byte [] val2 = Bytes.toBytes("value2");
byte [] val3 = Bytes.toBytes("value3");
byte[] emptyVal = new byte[] { };
- Integer lockId = null;
byte [][] families = {fam1, fam2};
@@ -1018,14 +1013,14 @@ public class TestHRegion extends HBaseTe
delete.deleteColumn(fam2, qf1);
delete.deleteColumn(fam1, qf3);
boolean res = region.checkAndMutate(row1, fam1, qf1, CompareOp.EQUAL,
- new BinaryComparator(val2), delete, lockId, true);
+ new BinaryComparator(val2), delete, true);
assertEquals(true, res);
Get get = new Get(row1);
get.addColumn(fam1, qf1);
get.addColumn(fam1, qf3);
get.addColumn(fam2, qf2);
- Result r = region.get(get, null);
+ Result r = region.get(get);
assertEquals(2, r.size());
assertEquals(val1, r.getValue(fam1, qf1));
assertEquals(val2, r.getValue(fam2, qf2));
@@ -1034,21 +1029,21 @@ public class TestHRegion extends HBaseTe
delete = new Delete(row1);
delete.deleteFamily(fam2);
res = region.checkAndMutate(row1, fam2, qf1, CompareOp.EQUAL,
- new BinaryComparator(emptyVal), delete, lockId, true);
+ new BinaryComparator(emptyVal), delete, true);
assertEquals(true, res);
get = new Get(row1);
- r = region.get(get, null);
+ r = region.get(get);
assertEquals(1, r.size());
assertEquals(val1, r.getValue(fam1, qf1));
//Row delete
delete = new Delete(row1);
res = region.checkAndMutate(row1, fam1, qf1, CompareOp.EQUAL,
- new BinaryComparator(val1), delete, lockId, true);
+ new BinaryComparator(val1), delete, true);
assertEquals(true, res);
get = new Get(row1);
- r = region.get(get, null);
+ r = region.get(get);
assertEquals(0, r.size());
} finally {
HRegion.closeHRegion(this.region);
@@ -1079,11 +1074,11 @@ public class TestHRegion extends HBaseTe
Delete delete = new Delete(row1);
delete.deleteColumn(fam1, qual);
delete.deleteColumn(fam1, qual);
- region.delete(delete, null, false);
+ region.delete(delete, false);
Get get = new Get(row1);
get.addFamily(fam1);
- Result r = region.get(get, null);
+ Result r = region.get(get);
assertEquals(0, r.size());
} finally {
HRegion.closeHRegion(this.region);
@@ -1165,19 +1160,19 @@ public class TestHRegion extends HBaseTe
// ok now delete a split:
Delete delete = new Delete(row);
delete.deleteColumns(fam, splitA);
- region.delete(delete, null, true);
+ region.delete(delete, true);
// assert some things:
Get get = new Get(row).addColumn(fam, serverinfo);
- Result result = region.get(get, null);
+ Result result = region.get(get);
assertEquals(1, result.size());
get = new Get(row).addColumn(fam, splitA);
- result = region.get(get, null);
+ result = region.get(get);
assertEquals(0, result.size());
get = new Get(row).addColumn(fam, splitB);
- result = region.get(get, null);
+ result = region.get(get);
assertEquals(1, result.size());
// Assert that after a delete, I can put.
@@ -1185,16 +1180,16 @@ public class TestHRegion extends HBaseTe
put.add(fam, splitA, Bytes.toBytes("reference_A"));
region.put(put);
get = new Get(row);
- result = region.get(get, null);
+ result = region.get(get);
assertEquals(3, result.size());
// Now delete all... then test I can add stuff back
delete = new Delete(row);
- region.delete(delete, null, false);
- assertEquals(0, region.get(get, null).size());
+ region.delete(delete, false);
+ assertEquals(0, region.get(get).size());
region.put(new Put(row).add(fam, splitA, Bytes.toBytes("reference_A")));
- result = region.get(get, null);
+ result = region.get(get);
assertEquals(1, result.size());
} finally {
HRegion.closeHRegion(this.region);
@@ -1220,20 +1215,20 @@ public class TestHRegion extends HBaseTe
// now delete something in the present
Delete delete = new Delete(row);
- region.delete(delete, null, true);
+ region.delete(delete, true);
// make sure we still see our data
Get get = new Get(row).addColumn(fam, serverinfo);
- Result result = region.get(get, null);
+ Result result = region.get(get);
assertEquals(1, result.size());
// delete the future row
- delete = new Delete(row,HConstants.LATEST_TIMESTAMP-3,null);
- region.delete(delete, null, true);
+ delete = new Delete(row,HConstants.LATEST_TIMESTAMP-3);
+ region.delete(delete, true);
// make sure it is gone
get = new Get(row).addColumn(fam, serverinfo);
- result = region.get(get, null);
+ result = region.get(get);
assertEquals(0, result.size());
} finally {
HRegion.closeHRegion(this.region);
@@ -1263,7 +1258,7 @@ public class TestHRegion extends HBaseTe
// Make sure it shows up with an actual timestamp
Get get = new Get(row).addColumn(fam, qual);
- Result result = region.get(get, null);
+ Result result = region.get(get);
assertEquals(1, result.size());
KeyValue kv = result.raw()[0];
LOG.info("Got: " + kv);
@@ -1279,7 +1274,7 @@ public class TestHRegion extends HBaseTe
// Make sure it shows up with an actual timestamp
get = new Get(row).addColumn(fam, qual);
- result = region.get(get, null);
+ result = region.get(get);
assertEquals(1, result.size());
kv = result.raw()[0];
LOG.info("Got: " + kv);
@@ -1344,7 +1339,7 @@ public class TestHRegion extends HBaseTe
Delete delete = new Delete(rowA);
delete.deleteFamily(fam1);
- region.delete(delete, null, true);
+ region.delete(delete, true);
// now create data.
Put put = new Put(rowA);
@@ -1395,7 +1390,7 @@ public class TestHRegion extends HBaseTe
region.put(put);
// now delete the value:
- region.delete(delete, null, true);
+ region.delete(delete, true);
// ok put data:
@@ -1407,7 +1402,7 @@ public class TestHRegion extends HBaseTe
Get get = new Get(row);
get.addColumn(fam1, qual1);
- Result r = region.get(get, null);
+ Result r = region.get(get);
assertEquals(1, r.size());
assertByteEquals(value2, r.getValue(fam1, qual1));
@@ -1487,7 +1482,7 @@ public class TestHRegion extends HBaseTe
//Test
try {
- region.get(get, null);
+ region.get(get);
} catch (DoNotRetryIOException e) {
assertFalse(false);
return;
@@ -1531,7 +1526,7 @@ public class TestHRegion extends HBaseTe
KeyValue [] expected = {kv1, kv2};
//Test
- Result res = region.get(get, null);
+ Result res = region.get(get);
assertEquals(expected.length, res.size());
for(int i=0; i<res.size(); i++){
assertEquals(0,
@@ -1547,7 +1542,7 @@ public class TestHRegion extends HBaseTe
Get g = new Get(row1);
final int count = 2;
g.setFilter(new ColumnCountGetFilter(count));
- res = region.get(g, null);
+ res = region.get(g);
assertEquals(count, res.size());
} finally {
HRegion.closeHRegion(this.region);
@@ -1565,7 +1560,7 @@ public class TestHRegion extends HBaseTe
try {
Get get = new Get(row);
get.addFamily(fam);
- Result r = region.get(get, null);
+ Result r = region.get(get);
assertTrue(r.isEmpty());
} finally {
@@ -1598,7 +1593,7 @@ public class TestHRegion extends HBaseTe
KeyValue [] expected = {kv1};
//Test from memstore
- Result res = region.get(get, null);
+ Result res = region.get(get);
assertEquals(expected.length, res.size());
for(int i=0; i<res.size(); i++){
@@ -1615,7 +1610,7 @@ public class TestHRegion extends HBaseTe
region.flushcache();
//test2
- res = region.get(get, null);
+ res = region.get(get);
assertEquals(expected.length, res.size());
for(int i=0; i<res.size(); i++){
@@ -1652,74 +1647,6 @@ public class TestHRegion extends HBaseTe
}
//////////////////////////////////////////////////////////////////////////////
- // Lock test
- //////////////////////////////////////////////////////////////////////////////
- public void testLocks() throws IOException{
- byte [] tableName = Bytes.toBytes("testtable");
- byte [][] families = {fam1, fam2, fam3};
-
- Configuration hc = initSplit();
- //Setting up region
- String method = this.getName();
- this.region = initHRegion(tableName, method, hc, families);
- try {
- final int threadCount = 10;
- final int lockCount = 10;
-
- List<Thread>threads = new ArrayList<Thread>(threadCount);
- for (int i = 0; i < threadCount; i++) {
- threads.add(new Thread(Integer.toString(i)) {
- @Override
- public void run() {
- Integer [] lockids = new Integer[lockCount];
- // Get locks.
- for (int i = 0; i < lockCount; i++) {
- try {
- byte [] rowid = Bytes.toBytes(Integer.toString(i));
- lockids[i] = region.obtainRowLock(rowid);
- assertEquals(rowid, region.getRowFromLock(lockids[i]));
- LOG.debug(getName() + " locked " + Bytes.toString(rowid));
- } catch (IOException e) {
- e.printStackTrace();
- }
- }
- LOG.debug(getName() + " set " +
- Integer.toString(lockCount) + " locks");
-
- // Abort outstanding locks.
- for (int i = lockCount - 1; i >= 0; i--) {
- region.releaseRowLock(lockids[i]);
- LOG.debug(getName() + " unlocked " + i);
- }
- LOG.debug(getName() + " released " +
- Integer.toString(lockCount) + " locks");
- }
- });
- }
-
- // Startup all our threads.
- for (Thread t : threads) {
- t.start();
- }
-
- // Now wait around till all are done.
- for (Thread t: threads) {
- while (t.isAlive()) {
- try {
- Thread.sleep(1);
- } catch (InterruptedException e) {
- // Go around again.
- }
- }
- }
- LOG.info("locks completed.");
- } finally {
- HRegion.closeHRegion(this.region);
- this.region = null;
- }
- }
-
- //////////////////////////////////////////////////////////////////////////////
// Merge test
//////////////////////////////////////////////////////////////////////////////
public void testMerge() throws IOException {
@@ -2390,7 +2317,7 @@ public class TestHRegion extends HBaseTe
// run a get and see?
Get get = new Get(row);
get.addColumn(familiy, qualifier);
- Result result = region.get(get, null);
+ Result result = region.get(get);
assertEquals(1, result.size());
KeyValue kv = result.raw()[0];
@@ -2405,7 +2332,7 @@ public class TestHRegion extends HBaseTe
// run a get and see?
Get get = new Get(row);
get.addColumn(familiy, qualifier);
- Result result = region.get(get, null);
+ Result result = region.get(get);
assertEquals(1, result.size());
KeyValue kv = result.raw()[0];
@@ -3051,8 +2978,8 @@ public class TestHRegion extends HBaseTe
numPutsFinished++;
if (numPutsFinished > 0 && numPutsFinished % 47 == 0) {
System.out.println("put iteration = " + numPutsFinished);
- Delete delete = new Delete(row, (long)numPutsFinished-30, null);
- region.delete(delete, null, true);
+ Delete delete = new Delete(row, (long)numPutsFinished-30);
+ region.delete(delete, true);
}
numPutsFinished++;
}
@@ -3137,7 +3064,7 @@ public class TestHRegion extends HBaseTe
for (int i = 0; i < testCount; i++) {
boolean previousEmpty = result == null || result.isEmpty();
- result = region.get(get, null);
+ result = region.get(get);
if (!result.isEmpty() || !previousEmpty || i > compactInterval) {
assertEquals("i=" + i, expectedCount, result.size());
// TODO this was removed, now what dangit?!
@@ -3196,14 +3123,14 @@ public class TestHRegion extends HBaseTe
byte[] rowNotServed = Bytes.toBytes("a");
Get g = new Get(rowNotServed);
try {
- region.get(g, null);
+ region.get(g);
fail();
} catch (WrongRegionException x) {
// OK
}
byte[] row = Bytes.toBytes("y");
g = new Get(row);
- region.get(g, null);
+ region.get(g);
} finally {
HRegion.closeHRegion(this.region);
this.region = null;
@@ -3224,9 +3151,9 @@ public class TestHRegion extends HBaseTe
region.flushcache();
- Delete delete = new Delete(Bytes.toBytes(1L), 1L, null);
+ Delete delete = new Delete(Bytes.toBytes(1L), 1L);
//delete.deleteColumn(family, qual1);
- region.delete(delete, null, true);
+ region.delete(delete, true);
put = new Put(Bytes.toBytes(2L));
put.add(family, qual1, 2L, Bytes.toBytes(2L));
@@ -3350,7 +3277,7 @@ public class TestHRegion extends HBaseTe
//Get rows
Get get = new Get(row);
get.setMaxVersions();
- KeyValue[] kvs = region.get(get, null).raw();
+ KeyValue[] kvs = region.get(get).raw();
//Check if rows are correct
assertEquals(4, kvs.length);
@@ -3394,14 +3321,14 @@ public class TestHRegion extends HBaseTe
region.flushcache();
Delete del = new Delete(row);
- region.delete(del, null, true);
+ region.delete(del, true);
region.flushcache();
// Get remaining rows (should have none)
Get get = new Get(row);
get.addColumn(familyName, col);
- KeyValue[] keyValues = region.get(get, null).raw();
+ KeyValue[] keyValues = region.get(get).raw();
assertTrue(keyValues.length == 0);
} finally {
HRegion.closeHRegion(this.region);
@@ -3580,7 +3507,7 @@ public class TestHRegion extends HBaseTe
inc.addColumn(family, qualifier, ONE);
count++;
try {
- region.increment(inc, null, true);
+ region.increment(inc, true);
} catch (IOException e) {
e.printStackTrace();
break;
@@ -3635,7 +3562,7 @@ public class TestHRegion extends HBaseTe
Get get = new Get(Incrementer.incRow);
get.addColumn(Incrementer.family, Incrementer.qualifier);
get.setMaxVersions(1);
- Result res = this.region.get(get, null);
+ Result res = this.region.get(get);
List<KeyValue> kvs = res.getColumn(Incrementer.family,
Incrementer.qualifier);
@@ -3671,7 +3598,7 @@ public class TestHRegion extends HBaseTe
app.add(family, qualifier, CHAR);
count++;
try {
- region.append(app, null, true);
+ region.append(app, true);
} catch (IOException e) {
e.printStackTrace();
break;
@@ -3729,7 +3656,7 @@ public class TestHRegion extends HBaseTe
Get get = new Get(Appender.appendRow);
get.addColumn(Appender.family, Appender.qualifier);
get.setMaxVersions(1);
- Result res = this.region.get(get, null);
+ Result res = this.region.get(get);
List<KeyValue> kvs = res.getColumn(Appender.family,
Appender.qualifier);
@@ -3766,7 +3693,7 @@ public class TestHRegion extends HBaseTe
get = new Get(row);
get.addColumn(family, qualifier);
get.setMaxVersions();
- res = this.region.get(get, null);
+ res = this.region.get(get);
kvs = res.getColumn(family, qualifier);
assertEquals(1, kvs.size());
assertEquals(Bytes.toBytes("value0"), kvs.get(0).getValue());
@@ -3775,7 +3702,7 @@ public class TestHRegion extends HBaseTe
get = new Get(row);
get.addColumn(family, qualifier);
get.setMaxVersions();
- res = this.region.get(get, null);
+ res = this.region.get(get);
kvs = res.getColumn(family, qualifier);
assertEquals(1, kvs.size());
assertEquals(Bytes.toBytes("value0"), kvs.get(0).getValue());
@@ -3787,7 +3714,7 @@ public class TestHRegion extends HBaseTe
get = new Get(row);
get.addColumn(family, qualifier);
get.setMaxVersions();
- res = this.region.get(get, null);
+ res = this.region.get(get);
kvs = res.getColumn(family, qualifier);
assertEquals(1, kvs.size());
assertEquals(Bytes.toBytes("value1"), kvs.get(0).getValue());
@@ -3796,7 +3723,7 @@ public class TestHRegion extends HBaseTe
get = new Get(row);
get.addColumn(family, qualifier);
get.setMaxVersions();
- res = this.region.get(get, null);
+ res = this.region.get(get);
kvs = res.getColumn(family, qualifier);
assertEquals(1, kvs.size());
assertEquals(Bytes.toBytes("value1"), kvs.get(0).getValue());
@@ -3824,7 +3751,7 @@ public class TestHRegion extends HBaseTe
for(byte [] family : families) {
get.addColumn(family, qf);
}
- Result result = newReg.get(get, null);
+ Result result = newReg.get(get);
KeyValue [] raw = result.raw();
assertEquals(families.length, result.size());
for(int j=0; j<families.length; j++) {
@@ -3839,7 +3766,7 @@ public class TestHRegion extends HBaseTe
throws IOException {
// Now I have k, get values out and assert they are as expected.
Get get = new Get(k).addFamily(family).setMaxVersions();
- KeyValue [] results = r.get(get, null).raw();
+ KeyValue [] results = r.get(get).raw();
for (int j = 0; j < results.length; j++) {
byte [] tmp = results[j].getValue();
// Row should be equal to value every time.
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionBusyWait.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionBusyWait.java?rev=1435747&r1=1435746&r2=1435747&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionBusyWait.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionBusyWait.java Sat Jan 19 22:45:11 2013
@@ -70,7 +70,7 @@ public class TestHRegionBusyWait extends
while (stopped.get()) {
Thread.sleep(100);
}
- region.get(get, null);
+ region.get(get);
fail("Should throw RegionTooBusyException");
} catch (InterruptedException ie) {
fail("test interrupted");
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestKeepDeletes.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestKeepDeletes.java?rev=1435747&r1=1435746&r2=1435747&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestKeepDeletes.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestKeepDeletes.java Sat Jan 19 22:45:11 2013
@@ -70,8 +70,8 @@ public class TestKeepDeletes extends HBa
region.put(p);
// now place a delete marker at ts+2
- Delete d = new Delete(T1, ts+2, null);
- region.delete(d, null, true);
+ Delete d = new Delete(T1, ts+2);
+ region.delete(d, true);
// a raw scan can see the delete markers
// (one for each column family)
@@ -81,14 +81,14 @@ public class TestKeepDeletes extends HBa
Get g = new Get(T1);
g.setMaxVersions();
g.setTimeRange(0L, ts+2);
- Result r = region.get(g, null);
+ Result r = region.get(g);
checkResult(r, c0, c0, T2,T1);
// flush
region.flushcache();
// yep, T2 still there, T1 gone
- r = region.get(g, null);
+ r = region.get(g);
checkResult(r, c0, c0, T2);
// major compact
@@ -100,12 +100,12 @@ public class TestKeepDeletes extends HBa
assertEquals(1, countDeleteMarkers(region));
// still there (even after multiple compactions)
- r = region.get(g, null);
+ r = region.get(g);
checkResult(r, c0, c0, T2);
// a timerange that includes the delete marker won't see past rows
g.setTimeRange(0L, ts+4);
- r = region.get(g, null);
+ r = region.get(g);
assertTrue(r.isEmpty());
// two more puts, this will expire the older puts.
@@ -121,7 +121,7 @@ public class TestKeepDeletes extends HBa
p = new Put(T1, ts);
p.add(c0, c0, T1);
region.put(p);
- r = region.get(g, null);
+ r = region.get(g);
assertTrue(r.isEmpty());
region.flushcache();
@@ -130,7 +130,7 @@ public class TestKeepDeletes extends HBa
// verify that the delete marker itself was collected
region.put(p);
- r = region.get(g, null);
+ r = region.get(g);
checkResult(r, c0, c0, T1);
assertEquals(0, countDeleteMarkers(region));
@@ -156,9 +156,9 @@ public class TestKeepDeletes extends HBa
p.add(c0, c0, T1);
region.put(p);
- Delete d = new Delete(T1, ts, null);
+ Delete d = new Delete(T1, ts);
d.deleteColumn(c0, c0, ts);
- region.delete(d, null, true);
+ region.delete(d, true);
// scan still returns delete markers and deletes rows
Scan s = new Scan();
@@ -199,15 +199,15 @@ public class TestKeepDeletes extends HBa
Put p = new Put(T1, ts);
p.add(c0, c0, T1);
region.put(p);
- Delete d = new Delete(T1, ts+2, null);
+ Delete d = new Delete(T1, ts+2);
d.deleteColumn(c0, c0, ts);
- region.delete(d, null, true);
+ region.delete(d, true);
// "past" get does not see rows behind delete marker
Get g = new Get(T1);
g.setMaxVersions();
g.setTimeRange(0L, ts+1);
- Result r = region.get(g, null);
+ Result r = region.get(g);
assertTrue(r.isEmpty());
// "past" scan does not see rows behind delete marker
@@ -272,16 +272,16 @@ public class TestKeepDeletes extends HBa
p.add(c0, c0, T3);
region.put(p);
- Delete d = new Delete(T1, ts+1, null);
- region.delete(d, null, true);
+ Delete d = new Delete(T1, ts+1);
+ region.delete(d, true);
- d = new Delete(T1, ts+2, null);
+ d = new Delete(T1, ts+2);
d.deleteColumn(c0, c0, ts+2);
- region.delete(d, null, true);
+ region.delete(d, true);
- d = new Delete(T1, ts+3, null);
+ d = new Delete(T1, ts+3);
d.deleteColumns(c0, c0, ts+3);
- region.delete(d, null, true);
+ region.delete(d, true);
Scan s = new Scan();
s.setRaw(true);
@@ -309,21 +309,21 @@ public class TestKeepDeletes extends HBa
long ts = System.currentTimeMillis();
- Delete d = new Delete(T1, ts, null);
+ Delete d = new Delete(T1, ts);
d.deleteColumns(c0, c0, ts);
- region.delete(d, null, true);
+ region.delete(d, true);
- d = new Delete(T1, ts, null);
+ d = new Delete(T1, ts);
d.deleteFamily(c0);
- region.delete(d, null, true);
+ region.delete(d, true);
- d = new Delete(T1, ts, null);
+ d = new Delete(T1, ts);
d.deleteColumn(c0, c0, ts+1);
- region.delete(d, null, true);
+ region.delete(d, true);
- d = new Delete(T1, ts, null);
+ d = new Delete(T1, ts);
d.deleteColumn(c0, c0, ts+2);
- region.delete(d, null, true);
+ region.delete(d, true);
// 1 family marker, 1 column marker, 2 version markers
assertEquals(4, countDeleteMarkers(region));
@@ -361,21 +361,21 @@ public class TestKeepDeletes extends HBa
region.put(p);
// all the following deletes affect the put
- Delete d = new Delete(T1, ts, null);
+ Delete d = new Delete(T1, ts);
d.deleteColumns(c0, c0, ts);
- region.delete(d, null, true);
+ region.delete(d, true);
- d = new Delete(T1, ts, null);
+ d = new Delete(T1, ts);
d.deleteFamily(c0, ts);
- region.delete(d, null, true);
+ region.delete(d, true);
- d = new Delete(T1, ts, null);
+ d = new Delete(T1, ts);
d.deleteColumn(c0, c0, ts+1);
- region.delete(d, null, true);
+ region.delete(d, true);
- d = new Delete(T1, ts, null);
+ d = new Delete(T1, ts);
d.deleteColumn(c0, c0, ts+2);
- region.delete(d, null, true);
+ region.delete(d, true);
// 1 family marker, 1 column marker, 2 version markers
assertEquals(4, countDeleteMarkers(region));
@@ -440,22 +440,22 @@ public class TestKeepDeletes extends HBa
p.add(c1, c1, T2);
region.put(p);
- Delete d = new Delete(T1, ts+2, null);
+ Delete d = new Delete(T1, ts+2);
d.deleteColumns(c0, c0, ts+2);
- region.delete(d, null, true);
+ region.delete(d, true);
- d = new Delete(T1, ts+2, null);
+ d = new Delete(T1, ts+2);
d.deleteFamily(c1, ts+2);
- region.delete(d, null, true);
+ region.delete(d, true);
- d = new Delete(T2, ts+2, null);
+ d = new Delete(T2, ts+2);
d.deleteFamily(c0, ts+2);
- region.delete(d, null, true);
+ region.delete(d, true);
// add an older delete, to make sure it is filtered
- d = new Delete(T1, ts-10, null);
+ d = new Delete(T1, ts-10);
d.deleteFamily(c1, ts-10);
- region.delete(d, null, true);
+ region.delete(d, true);
// ts + 2 does NOT include the delete at ts+2
checkGet(region, T1, c0, c0, ts+2, T2, T1);
@@ -503,18 +503,18 @@ public class TestKeepDeletes extends HBa
p.add(c0, c1, T1);
region.put(p);
- Delete d = new Delete(T1, ts, null);
+ Delete d = new Delete(T1, ts);
// test corner case (Put and Delete have same TS)
d.deleteColumns(c0, c0, ts);
- region.delete(d, null, true);
+ region.delete(d, true);
- d = new Delete(T1, ts+1, null);
+ d = new Delete(T1, ts+1);
d.deleteColumn(c0, c0, ts+1);
- region.delete(d, null, true);
+ region.delete(d, true);
- d = new Delete(T1, ts+3, null);
+ d = new Delete(T1, ts+3);
d.deleteColumn(c0, c0, ts+3);
- region.delete(d, null, true);
+ region.delete(d, true);
region.flushcache();
region.compactStores(true);
@@ -601,11 +601,11 @@ public class TestKeepDeletes extends HBa
region.put(p);
// family markers are each family
- Delete d = new Delete(T1, ts+1, null);
- region.delete(d, null, true);
+ Delete d = new Delete(T1, ts+1);
+ region.delete(d, true);
- d = new Delete(T2, ts+2, null);
- region.delete(d, null, true);
+ d = new Delete(T2, ts+2);
+ region.delete(d, true);
Scan s = new Scan(T1);
s.setTimeRange(0, ts+1);
@@ -652,17 +652,17 @@ public class TestKeepDeletes extends HBa
// all puts now are just retained because of min versions = 3
// place a family delete marker
- Delete d = new Delete(T1, ts-1, null);
- region.delete(d, null, true);
+ Delete d = new Delete(T1, ts-1);
+ region.delete(d, true);
// and a column delete marker
- d = new Delete(T1, ts-2, null);
+ d = new Delete(T1, ts-2);
d.deleteColumns(c0, c0, ts-1);
- region.delete(d, null, true);
+ region.delete(d, true);
Get g = new Get(T1);
g.setMaxVersions();
g.setTimeRange(0L, ts-2);
- Result r = region.get(g, null);
+ Result r = region.get(g);
checkResult(r, c0, c0, T1,T0);
// 3 families, one column delete marker
@@ -672,7 +672,7 @@ public class TestKeepDeletes extends HBa
// no delete marker removes by the flush
assertEquals(4, countDeleteMarkers(region));
- r = region.get(g, null);
+ r = region.get(g);
checkResult(r, c0, c0, T1);
p = new Put(T1, ts+1);
p.add(c0, c0, T4);
@@ -681,7 +681,7 @@ public class TestKeepDeletes extends HBa
assertEquals(4, countDeleteMarkers(region));
- r = region.get(g, null);
+ r = region.get(g);
checkResult(r, c0, c0, T1);
// this will push out the last put before
@@ -709,7 +709,7 @@ public class TestKeepDeletes extends HBa
g.addColumn(fam, col);
g.setMaxVersions();
g.setTimeRange(0L, time);
- Result r = region.get(g, null);
+ Result r = region.get(g);
checkResult(r, fam, col, vals);
}
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMinVersions.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMinVersions.java?rev=1435747&r1=1435746&r2=1435747&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMinVersions.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMinVersions.java Sat Jan 19 22:45:11 2013
@@ -133,13 +133,13 @@ public class TestMinVersions extends HBa
Get g = new Get(T1);
g.setMaxVersions();
- Result r = region.get(g, null); // this'll use ScanWildcardColumnTracker
+ Result r = region.get(g); // this'll use ScanWildcardColumnTracker
checkResult(r, c0, T3,T2,T1);
g = new Get(T1);
g.setMaxVersions();
g.addColumn(c0, c0);
- r = region.get(g, null); // this'll use ExplicitColumnTracker
+ r = region.get(g); // this'll use ExplicitColumnTracker
checkResult(r, c0, T3,T2,T1);
} finally {
HRegion.closeHRegion(region);
@@ -169,18 +169,18 @@ public class TestMinVersions extends HBa
p.add(c0, c0, T3);
region.put(p);
- Delete d = new Delete(T1, ts-1, null);
- region.delete(d, null, true);
+ Delete d = new Delete(T1, ts-1);
+ region.delete(d, true);
Get g = new Get(T1);
g.setMaxVersions();
- Result r = region.get(g, null); // this'll use ScanWildcardColumnTracker
+ Result r = region.get(g); // this'll use ScanWildcardColumnTracker
checkResult(r, c0, T3);
g = new Get(T1);
g.setMaxVersions();
g.addColumn(c0, c0);
- r = region.get(g, null); // this'll use ExplicitColumnTracker
+ r = region.get(g); // this'll use ExplicitColumnTracker
checkResult(r, c0, T3);
// now flush/compact
@@ -190,13 +190,13 @@ public class TestMinVersions extends HBa
// try again
g = new Get(T1);
g.setMaxVersions();
- r = region.get(g, null); // this'll use ScanWildcardColumnTracker
+ r = region.get(g); // this'll use ScanWildcardColumnTracker
checkResult(r, c0, T3);
g = new Get(T1);
g.setMaxVersions();
g.addColumn(c0, c0);
- r = region.get(g, null); // this'll use ExplicitColumnTracker
+ r = region.get(g); // this'll use ExplicitColumnTracker
checkResult(r, c0, T3);
} finally {
HRegion.closeHRegion(region);
@@ -241,18 +241,18 @@ public class TestMinVersions extends HBa
// now the latest change is in the memstore,
// but it is not the latest version
- Result r = region.get(new Get(T1), null);
+ Result r = region.get(new Get(T1));
checkResult(r, c0, T4);
Get g = new Get(T1);
g.setMaxVersions();
- r = region.get(g, null); // this'll use ScanWildcardColumnTracker
+ r = region.get(g); // this'll use ScanWildcardColumnTracker
checkResult(r, c0, T4,T3);
g = new Get(T1);
g.setMaxVersions();
g.addColumn(c0, c0);
- r = region.get(g, null); // this'll use ExplicitColumnTracker
+ r = region.get(g); // this'll use ExplicitColumnTracker
checkResult(r, c0, T4,T3);
p = new Put(T1, ts+1);
@@ -263,13 +263,13 @@ public class TestMinVersions extends HBa
g = new Get(T1);
g.setMaxVersions();
- r = region.get(g, null); // this'll use ScanWildcardColumnTracker
+ r = region.get(g); // this'll use ScanWildcardColumnTracker
checkResult(r, c0, T5,T4);
g = new Get(T1);
g.setMaxVersions();
g.addColumn(c0, c0);
- r = region.get(g, null); // this'll use ExplicitColumnTracker
+ r = region.get(g); // this'll use ExplicitColumnTracker
checkResult(r, c0, T5,T4);
} finally {
HRegion.closeHRegion(region);
@@ -308,30 +308,30 @@ public class TestMinVersions extends HBa
p.add(c0, c0, T4);
region.put(p);
- Result r = region.get(new Get(T1), null);
+ Result r = region.get(new Get(T1));
checkResult(r, c0, T4);
Get g = new Get(T1);
g.setTimeRange(0L, ts+1);
- r = region.get(g, null);
+ r = region.get(g);
checkResult(r, c0, T4);
// oldest version still exists
g.setTimeRange(0L, ts-2);
- r = region.get(g, null);
+ r = region.get(g);
checkResult(r, c0, T1);
// gets see only available versions
// even before compactions
g = new Get(T1);
g.setMaxVersions();
- r = region.get(g, null); // this'll use ScanWildcardColumnTracker
+ r = region.get(g); // this'll use ScanWildcardColumnTracker
checkResult(r, c0, T4,T3);
g = new Get(T1);
g.setMaxVersions();
g.addColumn(c0, c0);
- r = region.get(g, null); // this'll use ExplicitColumnTracker
+ r = region.get(g); // this'll use ExplicitColumnTracker
checkResult(r, c0, T4,T3);
// now flush
@@ -340,7 +340,7 @@ public class TestMinVersions extends HBa
// with HBASE-4241 a flush will eliminate the expired rows
g = new Get(T1);
g.setTimeRange(0L, ts-2);
- r = region.get(g, null);
+ r = region.get(g);
assertTrue(r.isEmpty());
// major compaction
@@ -349,17 +349,17 @@ public class TestMinVersions extends HBa
// after compaction the 4th version is still available
g = new Get(T1);
g.setTimeRange(0L, ts+1);
- r = region.get(g, null);
+ r = region.get(g);
checkResult(r, c0, T4);
// so is the 3rd
g.setTimeRange(0L, ts);
- r = region.get(g, null);
+ r = region.get(g);
checkResult(r, c0, T3);
// but the 2nd and earlier versions are gone
g.setTimeRange(0L, ts-1);
- r = region.get(g, null);
+ r = region.get(g);
assertTrue(r.isEmpty());
} finally {
HRegion.closeHRegion(region);
@@ -407,14 +407,14 @@ public class TestMinVersions extends HBa
g.addColumn(c1,c1);
g.setFilter(new TimestampsFilter(tss));
g.setMaxVersions();
- Result r = region.get(g, null);
+ Result r = region.get(g);
checkResult(r, c1, T2,T1);
g = new Get(T1);
g.addColumn(c0,c0);
g.setFilter(new TimestampsFilter(tss));
g.setMaxVersions();
- r = region.get(g, null);
+ r = region.get(g);
checkResult(r, c0, T2,T1);
// now flush/compact
@@ -425,14 +425,14 @@ public class TestMinVersions extends HBa
g.addColumn(c1,c1);
g.setFilter(new TimestampsFilter(tss));
g.setMaxVersions();
- r = region.get(g, null);
+ r = region.get(g);
checkResult(r, c1, T2);
g = new Get(T1);
g.addColumn(c0,c0);
g.setFilter(new TimestampsFilter(tss));
g.setMaxVersions();
- r = region.get(g, null);
+ r = region.get(g);
checkResult(r, c0, T2);
} finally {
HRegion.closeHRegion(region);
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMultiColumnScanner.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMultiColumnScanner.java?rev=1435747&r1=1435746&r2=1435747&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMultiColumnScanner.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMultiColumnScanner.java Sat Jan 19 22:45:11 2013
@@ -214,7 +214,7 @@ public class TestMultiColumnScanner {
deletedSomething = true;
}
if (deletedSomething)
- region.delete(d, null, true);
+ region.delete(d, true);
}
}
region.flushcache();
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestParallelPut.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestParallelPut.java?rev=1435747&r1=1435746&r2=1435747&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestParallelPut.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestParallelPut.java Sat Jan 19 22:45:11 2013
@@ -167,7 +167,7 @@ public class TestParallelPut extends HBa
// run a get and see if the value matches
Get get = new Get(row);
get.addColumn(familiy, qualifier);
- Result result = region.get(get, null);
+ Result result = region.get(get);
assertEquals(1, result.size());
KeyValue kv = result.raw()[0];
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestResettingCounters.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestResettingCounters.java?rev=1435747&r1=1435746&r2=1435747&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestResettingCounters.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestResettingCounters.java Sat Jan 19 22:45:11 2013
@@ -78,14 +78,14 @@ public class TestResettingCounters {
}
// increment odd qualifiers 5 times and flush
- for (int i=0;i<5;i++) region.increment(odd, null, false);
+ for (int i=0;i<5;i++) region.increment(odd, false);
region.flushcache();
// increment even qualifiers 5 times
- for (int i=0;i<5;i++) region.increment(even, null, false);
+ for (int i=0;i<5;i++) region.increment(even, false);
// increment all qualifiers, should have value=6 for all
- Result result = region.increment(all, null, false);
+ Result result = region.increment(all, false);
assertEquals(numQualifiers, result.size());
KeyValue [] kvs = result.raw();
for (int i=0;i<kvs.length;i++) {
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestScanner.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestScanner.java?rev=1435747&r1=1435746&r2=1435747&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestScanner.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestScanner.java Sat Jan 19 22:45:11 2013
@@ -234,7 +234,7 @@ public class TestScanner extends HBaseTe
// Write information to the meta table
- Put put = new Put(ROW_KEY, System.currentTimeMillis(), null);
+ Put put = new Put(ROW_KEY, System.currentTimeMillis());
put.add(HConstants.CATALOG_FAMILY, HConstants.REGIONINFO_QUALIFIER,
REGION_INFO.toByteArray());
@@ -261,7 +261,7 @@ public class TestScanner extends HBaseTe
String address = HConstants.LOCALHOST_IP + ":" + HBaseTestingUtility.randomFreePort();
- put = new Put(ROW_KEY, System.currentTimeMillis(), null);
+ put = new Put(ROW_KEY, System.currentTimeMillis());
put.add(HConstants.CATALOG_FAMILY, HConstants.SERVER_QUALIFIER,
Bytes.toBytes(address));
@@ -299,7 +299,7 @@ public class TestScanner extends HBaseTe
address = "bar.foo.com:4321";
- put = new Put(ROW_KEY, System.currentTimeMillis(), null);
+ put = new Put(ROW_KEY, System.currentTimeMillis());
put.add(HConstants.CATALOG_FAMILY, HConstants.SERVER_QUALIFIER,
Bytes.toBytes(address));
@@ -428,7 +428,7 @@ public class TestScanner extends HBaseTe
private void getRegionInfo() throws IOException {
Get get = new Get(ROW_KEY);
get.addColumn(HConstants.CATALOG_FAMILY, HConstants.REGIONINFO_QUALIFIER);
- Result result = region.get(get, null);
+ Result result = region.get(get);
byte [] bytes = result.value();
validateRegionInfo(bytes);
}
@@ -498,7 +498,7 @@ public class TestScanner extends HBaseTe
Delete dc = new Delete(firstRowBytes);
/* delete column1 of firstRow */
dc.deleteColumns(fam1, col1);
- r.delete(dc, null, true);
+ r.delete(dc, true);
r.flushcache();
addContent(hri, Bytes.toString(fam1), Bytes.toString(col1),
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSeekOptimizations.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSeekOptimizations.java?rev=1435747&r1=1435746&r2=1435747&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSeekOptimizations.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSeekOptimizations.java Sat Jan 19 22:45:11 2013
@@ -419,7 +419,7 @@ public class TestSeekOptimizations {
region.put(put);
if (!del.isEmpty()) {
- region.delete(del, null, true);
+ region.delete(del, true);
}
// Add remaining timestamps (those we have not deleted) to expected
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALReplay.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALReplay.java?rev=1435747&r1=1435746&r2=1435747&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALReplay.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALReplay.java Sat Jan 19 22:45:11 2013
@@ -380,7 +380,7 @@ public class TestWALReplay {
}
// Now assert edits made it in.
final Get g = new Get(rowName);
- Result result = region.get(g, null);
+ Result result = region.get(g);
assertEquals(countPerFamily * htd.getFamilies().size(),
result.size());
// Now close the region (without flush), split the log, reopen the region and assert that
@@ -395,7 +395,7 @@ public class TestWALReplay {
// HRegionServer usually does this. It knows the largest seqid across all regions.
wal2.setSequenceNumber(seqid2);
assertTrue(seqid + result.size() < seqid2);
- final Result result1b = region2.get(g, null);
+ final Result result1b = region2.get(g);
assertEquals(result.size(), result1b.size());
// Next test. Add more edits, then 'crash' this region by stealing its wal
@@ -405,7 +405,7 @@ public class TestWALReplay {
addRegionEdits(rowName, hcd.getName(), countPerFamily, this.ee, region2, "y");
}
// Get count of edits.
- final Result result2 = region2.get(g, null);
+ final Result result2 = region2.get(g);
assertEquals(2 * result.size(), result2.size());
wal2.sync();
// Set down maximum recovery so we dfsclient doesn't linger retrying something
@@ -432,7 +432,7 @@ public class TestWALReplay {
long seqid3 = region3.initialize();
// HRegionServer usually does this. It knows the largest seqid across all regions.
wal3.setSequenceNumber(seqid3);
- Result result3 = region3.get(g, null);
+ Result result3 = region3.get(g);
// Assert that count of cells is same as before crash.
assertEquals(result2.size(), result3.size());
assertEquals(htd.getFamilies().size() * countPerFamily,
@@ -492,7 +492,7 @@ public class TestWALReplay {
// Now assert edits made it in.
final Get g = new Get(rowName);
- Result result = region.get(g, null);
+ Result result = region.get(g);
assertEquals(countPerFamily * htd.getFamilies().size(),
result.size());
@@ -524,7 +524,7 @@ public class TestWALReplay {
wal2.setSequenceNumber(seqid2);
assertTrue(seqid + result.size() < seqid2);
- final Result result1b = region2.get(g, null);
+ final Result result1b = region2.get(g);
assertEquals(result.size(), result1b.size());
}
@@ -612,7 +612,7 @@ public class TestWALReplay {
assertTrue(seqid > wal.getSequenceNumber());
Get get = new Get(rowName);
- Result result = region.get(get, -1);
+ Result result = region.get(get);
// Make sure we only see the good edits
assertEquals(countPerFamily * (htd.getFamilies().size() - 1),
result.size());
Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMergeTool.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMergeTool.java?rev=1435747&r1=1435746&r2=1435747&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMergeTool.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMergeTool.java Sat Jan 19 22:45:11 2013
@@ -233,7 +233,7 @@ public class TestMergeTool extends HBase
for (int j = 0; j < rows[i].length; j++) {
Get get = new Get(rows[i][j]);
get.addFamily(FAMILY);
- Result result = merged.get(get, null);
+ Result result = merged.get(get);
assertEquals(1, result.size());
byte [] bytes = result.raw()[0].getValue();
assertNotNull(Bytes.toStringBinary(rows[i][j]), bytes);
@@ -253,7 +253,7 @@ public class TestMergeTool extends HBase
for (int j = 0; j < rows[i].length; j++) {
Get get = new Get(rows[i][j]);
get.addFamily(FAMILY);
- Result result = regions[i].get(get, null);
+ Result result = regions[i].get(get);
byte [] bytes = result.raw()[0].getValue();
assertNotNull(bytes);
assertTrue(Bytes.equals(bytes, rows[i][j]));