You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by ps...@apache.org on 2021/09/29 08:04:23 UTC
[hbase] branch master updated: HBASE-26289 Hbase scan
setMaxResultsPerColumnFamily not giving right results (#3688)
This is an automated email from the ASF dual-hosted git repository.
psomogyi pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hbase.git
The following commit(s) were added to refs/heads/master by this push:
new f000b77 HBASE-26289 Hbase scan setMaxResultsPerColumnFamily not giving right results (#3688)
f000b77 is described below
commit f000b775320330eb2f426f6b2a3b5e27a794a707
Author: richardantal <60...@users.noreply.github.com>
AuthorDate: Wed Sep 29 10:03:53 2021 +0200
HBASE-26289 Hbase scan setMaxResultsPerColumnFamily not giving right results (#3688)
Co-authored-by: Duo Zhang <zh...@apache.org>
Signed-off-by: Duo Zhang <zh...@apache.org>
Signed-off-by: Peter Somogyi <ps...@apache.org>
---
.../hadoop/hbase/regionserver/StoreScanner.java | 22 ++--
.../TestMaxResultsPerColumnFamily.java | 117 +++++++++++++++++++++
2 files changed, 128 insertions(+), 11 deletions(-)
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreScanner.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreScanner.java
index bb1354d..a90d9d8 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreScanner.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/StoreScanner.java
@@ -626,22 +626,11 @@ public class StoreScanner extends NonReversedNonLazyKeyValueScanner
case INCLUDE:
case INCLUDE_AND_SEEK_NEXT_ROW:
case INCLUDE_AND_SEEK_NEXT_COL:
-
Filter f = matcher.getFilter();
if (f != null) {
cell = f.transformCell(cell);
}
this.countPerRow++;
- if (storeLimit > -1 && this.countPerRow > (storeLimit + storeOffset)) {
- // do what SEEK_NEXT_ROW does.
- if (!matcher.moreRowsMayExistAfter(cell)) {
- close(false);// Do all cleanup except heap.close()
- return scannerContext.setScannerState(NextState.NO_MORE_VALUES).hasMoreValues();
- }
- matcher.clearCurrentRow();
- seekToNextRow(cell);
- break LOOP;
- }
// add to results only if we have skipped #storeOffset kvs
// also update metric accordingly
@@ -670,6 +659,17 @@ public class StoreScanner extends NonReversedNonLazyKeyValueScanner
LOG.warn(message);
throw new RowTooBigException(message);
}
+
+ if (storeLimit > -1 && this.countPerRow >= (storeLimit + storeOffset)) {
+ // do what SEEK_NEXT_ROW does.
+ if (!matcher.moreRowsMayExistAfter(cell)) {
+ close(false);// Do all cleanup except heap.close()
+ return scannerContext.setScannerState(NextState.NO_MORE_VALUES).hasMoreValues();
+ }
+ matcher.clearCurrentRow();
+ seekToNextRow(cell);
+ break LOOP;
+ }
}
if (qcode == ScanQueryMatcher.MatchCode.INCLUDE_AND_SEEK_NEXT_ROW) {
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMaxResultsPerColumnFamily.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMaxResultsPerColumnFamily.java
new file mode 100644
index 0000000..d8a10bf
--- /dev/null
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMaxResultsPerColumnFamily.java
@@ -0,0 +1,117 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements. See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership. The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.hadoop.hbase.regionserver;
+
+import static org.junit.Assert.assertEquals;
+import org.apache.hadoop.hbase.HBaseClassTestRule;
+import org.apache.hadoop.hbase.HBaseTestingUtil;
+import org.apache.hadoop.hbase.TableName;
+import org.apache.hadoop.hbase.client.Admin;
+import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;
+import org.apache.hadoop.hbase.client.Put;
+import org.apache.hadoop.hbase.client.Result;
+import org.apache.hadoop.hbase.client.ResultScanner;
+import org.apache.hadoop.hbase.client.Scan;
+import org.apache.hadoop.hbase.client.Table;
+import org.apache.hadoop.hbase.client.TableDescriptor;
+import org.apache.hadoop.hbase.client.TableDescriptorBuilder;
+import org.apache.hadoop.hbase.testclassification.MediumTests;
+import org.apache.hadoop.hbase.util.Bytes;
+import org.junit.AfterClass;
+import org.junit.BeforeClass;
+import org.junit.ClassRule;
+import org.junit.Rule;
+import org.junit.Test;
+import org.junit.experimental.categories.Category;
+import org.junit.rules.TestName;
+
+@Category(MediumTests.class)
+public class TestMaxResultsPerColumnFamily {
+
+ @ClassRule
+ public static final HBaseClassTestRule CLASS_RULE =
+ HBaseClassTestRule.forClass(TestMaxResultsPerColumnFamily.class);
+
+ private static final byte [][] FAMILIES = {
+ Bytes.toBytes("1"), Bytes.toBytes("2")
+ };
+
+ private static final byte [][] VALUES = {
+ Bytes.toBytes("testValueOne"), Bytes.toBytes("testValueTwo")
+ };
+
+ private static final HBaseTestingUtil UTIL = new HBaseTestingUtil();
+
+ @BeforeClass
+ public static void setUpBeforeClass() throws Exception {
+ UTIL.startMiniCluster(1);
+ }
+
+ @AfterClass
+ public static void tearDownAfterClass() throws Exception {
+ UTIL.shutdownMiniCluster();
+ }
+
+ @Rule
+ public TestName name = new TestName();
+
+ @Test
+ public void testSetMaxResultsPerColumnFamilySimple() throws Exception {
+ TableName tableName = TableName.valueOf(name.getMethodName());
+ Admin admin = UTIL.getAdmin();
+ ColumnFamilyDescriptorBuilder cfBuilder0 =
+ ColumnFamilyDescriptorBuilder.newBuilder(FAMILIES[0]);
+
+ TableDescriptor tableDescriptor =
+ TableDescriptorBuilder.newBuilder(tableName).setColumnFamily(cfBuilder0.build()).build();
+ admin.createTable(tableDescriptor);
+ try (Table table = UTIL.getConnection().getTable(tableName)) {
+
+ for (int i = 0; i < 30000; i++) {
+ byte[] ROW = Bytes.toBytes("" + i);
+ Put p = new Put(ROW);
+
+ p.addColumn(FAMILIES[0], Bytes.toBytes("" + 1), VALUES[1]);
+ p.addColumn(FAMILIES[0], Bytes.toBytes("" + 2), VALUES[0]);
+
+ table.put(p);
+ }
+ }
+
+ try (Table t = UTIL.getConnection().getTable(tableName)) {
+ int expected = 30000;
+
+ Scan limits = new Scan().setReadType(Scan.ReadType.PREAD);
+ limits.setMaxResultsPerColumnFamily(1);
+
+ int count1 = countScanRows(t, limits);
+ assertEquals(expected, count1);
+ }
+ }
+
+ static int countScanRows(Table t, Scan scan) throws Exception {
+
+ int count = 0;
+ try(ResultScanner scanner = t.getScanner(scan)) {
+ for(Result r:scanner) {
+ count ++;
+ }
+ }
+ return count;
+ }
+}