You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by te...@apache.org on 2014/04/03 12:13:32 UTC
svn commit: r1584335 - in /hbase/trunk:
hbase-client/src/main/java/org/apache/hadoop/hbase/filter/
hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/
hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/
Author: tedyu
Date: Thu Apr 3 10:13:32 2014
New Revision: 1584335
URL: http://svn.apache.org/r1584335
Log:
HBASE-10850 essential column family optimization is broken
Added:
hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSCVFWithMiniCluster.java
Modified:
hbase/trunk/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterWrapper.java
hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
Modified: hbase/trunk/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterWrapper.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterWrapper.java?rev=1584335&r1=1584334&r2=1584335&view=diff
==============================================================================
--- hbase/trunk/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterWrapper.java (original)
+++ hbase/trunk/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterWrapper.java Thu Apr 3 10:13:32 2014
@@ -31,7 +31,6 @@ import org.apache.hadoop.hbase.KeyValueU
import org.apache.hadoop.hbase.exceptions.DeserializationException;
import org.apache.hadoop.hbase.protobuf.ProtobufUtil;
import org.apache.hadoop.hbase.protobuf.generated.FilterProtos;
-import org.apache.zookeeper.KeeperException.UnimplementedException;
import com.google.protobuf.InvalidProtocolBufferException;
@@ -154,15 +153,29 @@ final public class FilterWrapper extends
@Override
public void filterRowCells(List<Cell> kvs) throws IOException {
+ filterRowCellsWithRet(kvs);
+ }
+
+ public enum FilterRowRetCode {
+ NOT_CALLED,
+ INCLUDE, // corresponds to filter.filterRow() returning false
+ EXCLUDE // corresponds to filter.filterRow() returning true
+ }
+ public FilterRowRetCode filterRowCellsWithRet(List<Cell> kvs) throws IOException {
//To fix HBASE-6429,
//Filter with filterRow() returning true is incompatible with scan with limit
//1. hasFilterRow() returns true, if either filterRow() or filterRow(kvs) is implemented.
//2. filterRow() is merged with filterRow(kvs),
//so that to make all those row related filtering stuff in the same function.
this.filter.filterRowCells(kvs);
- if (!kvs.isEmpty() && this.filter.filterRow()) {
- kvs.clear();
+ if (!kvs.isEmpty()) {
+ if (this.filter.filterRow()) {
+ kvs.clear();
+ return FilterRowRetCode.EXCLUDE;
+ }
+ return FilterRowRetCode.INCLUDE;
}
+ return FilterRowRetCode.NOT_CALLED;
}
/**
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java?rev=1584335&r1=1584334&r2=1584335&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java Thu Apr 3 10:13:32 2014
@@ -3636,7 +3636,7 @@ public class HRegion implements HeapSize
// KeyValue indicating that limit is reached when scanning
private final KeyValue KV_LIMIT = new KeyValue();
protected final byte[] stopRow;
- private final Filter filter;
+ private final FilterWrapper filter;
private int batch;
protected int isScan;
private boolean filterClosed = false;
@@ -3916,14 +3916,15 @@ public class HRegion implements HeapSize
isStopRow(nextKv.getRowArray(), nextKv.getRowOffset(), nextKv.getRowLength());
// save that the row was empty before filters applied to it.
final boolean isEmptyRow = results.isEmpty();
-
+
// We have the part of the row necessary for filtering (all of it, usually).
// First filter with the filterRow(List).
+ FilterWrapper.FilterRowRetCode ret = FilterWrapper.FilterRowRetCode.NOT_CALLED;
if (filter != null && filter.hasFilterRow()) {
- filter.filterRowCells(results);
+ ret = filter.filterRowCellsWithRet(results);
}
- if (isEmptyRow || filterRow()) {
+ if ((isEmptyRow || ret == FilterWrapper.FilterRowRetCode.EXCLUDE) || filterRow()) {
results.clear();
boolean moreRows = nextRow(currentRow, offset, length);
if (!moreRows) return false;
Added: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSCVFWithMiniCluster.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSCVFWithMiniCluster.java?rev=1584335&view=auto
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSCVFWithMiniCluster.java (added)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSCVFWithMiniCluster.java Thu Apr 3 10:13:32 2014
@@ -0,0 +1,242 @@
+/*
+ * Copyright The Apache Software Foundation
+ *
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements. See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership. The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.hadoop.hbase.regionserver;
+
+import static junit.framework.Assert.assertEquals;
+
+import java.io.IOException;
+import java.util.ArrayList;
+import java.util.Iterator;
+import java.util.List;
+
+import org.apache.hadoop.hbase.HBaseTestingUtility;
+import org.apache.hadoop.hbase.HColumnDescriptor;
+import org.apache.hadoop.hbase.HConstants;
+import org.apache.hadoop.hbase.HTableDescriptor;
+import org.apache.hadoop.hbase.MediumTests;
+import org.apache.hadoop.hbase.TableExistsException;
+import org.apache.hadoop.hbase.TableName;
+import org.apache.hadoop.hbase.TableNotFoundException;
+import org.apache.hadoop.hbase.client.Durability;
+import org.apache.hadoop.hbase.client.HBaseAdmin;
+import org.apache.hadoop.hbase.client.HTable;
+import org.apache.hadoop.hbase.client.Put;
+import org.apache.hadoop.hbase.client.Result;
+import org.apache.hadoop.hbase.client.ResultScanner;
+import org.apache.hadoop.hbase.client.Scan;
+import org.apache.hadoop.hbase.filter.BinaryComparator;
+import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp;
+import org.apache.hadoop.hbase.filter.Filter;
+import org.apache.hadoop.hbase.filter.SingleColumnValueFilter;
+import org.apache.hadoop.hbase.io.compress.Compression.Algorithm;
+import org.apache.hadoop.hbase.util.Bytes;
+import org.junit.After;
+import org.junit.Before;
+import org.junit.Test;
+import org.junit.experimental.categories.Category;
+
+@Category(MediumTests.class)
+/*
+ * This test verifies that the scenarios illustrated by HBASE-10850 work
+ * w.r.t. essential column family optimization
+ */
+public class TestSCVFWithMiniCluster {
+ private static final String HBASE_TABLE_NAME = "TestSCVFWithMiniCluster";
+
+ private static final byte[] FAMILY_A = Bytes.toBytes("a");
+ private static final byte[] FAMILY_B = Bytes.toBytes("b");
+
+ private static final byte[] QUALIFIER_FOO = Bytes.toBytes("foo");
+ private static final byte[] QUALIFIER_BAR = Bytes.toBytes("bar");
+
+ private HTable htable;
+
+ private Filter scanFilter;
+
+ private int expected = 1;
+
+ @Before
+ public void setUp() throws Exception {
+ HBaseTestingUtility util = new HBaseTestingUtility();
+
+ util.startMiniCluster(1);
+
+ HBaseAdmin admin = util.getHBaseAdmin();
+ destroy(admin, HBASE_TABLE_NAME);
+ create(admin, HBASE_TABLE_NAME, FAMILY_A, FAMILY_B);
+ admin.close();
+ htable = new HTable(util.getConfiguration(), HBASE_TABLE_NAME);
+
+ /* Add some values */
+ List<Put> puts = new ArrayList<Put>();
+
+ /* Add a row with 'a:foo' = false */
+ Put put = new Put(Bytes.toBytes("1"));
+ put.setDurability(Durability.SKIP_WAL);
+ put.add(FAMILY_A, QUALIFIER_FOO, Bytes.toBytes("false"));
+ put.add(FAMILY_A, QUALIFIER_BAR, Bytes.toBytes("_flag_"));
+ put.add(FAMILY_B, QUALIFIER_FOO, Bytes.toBytes("_flag_"));
+ put.add(FAMILY_B, QUALIFIER_BAR, Bytes.toBytes("_flag_"));
+ puts.add(put);
+
+ /* Add a row with 'a:foo' = true */
+ put = new Put(Bytes.toBytes("2"));
+ put.setDurability(Durability.SKIP_WAL);
+ put.add(FAMILY_A, QUALIFIER_FOO, Bytes.toBytes("true"));
+ put.add(FAMILY_A, QUALIFIER_BAR, Bytes.toBytes("_flag_"));
+ put.add(FAMILY_B, QUALIFIER_FOO, Bytes.toBytes("_flag_"));
+ put.add(FAMILY_B, QUALIFIER_BAR, Bytes.toBytes("_flag_"));
+ puts.add(put);
+
+ /* Add a row with 'a:foo' qualifier not set */
+ put = new Put(Bytes.toBytes("3"));
+ put.setDurability(Durability.SKIP_WAL);
+ put.add(FAMILY_A, QUALIFIER_BAR, Bytes.toBytes("_flag_"));
+ put.add(FAMILY_B, QUALIFIER_FOO, Bytes.toBytes("_flag_"));
+ put.add(FAMILY_B, QUALIFIER_BAR, Bytes.toBytes("_flag_"));
+ puts.add(put);
+
+ htable.put(puts);
+ /*
+ * We want to filter out from the scan all rows that do not have the column 'a:foo' with value
+ * 'false'. Only row with key '1' should be returned in the scan.
+ */
+ scanFilter = new SingleColumnValueFilter(FAMILY_A, QUALIFIER_FOO, CompareOp.EQUAL,
+ new BinaryComparator(Bytes.toBytes("false")));
+ ((SingleColumnValueFilter) scanFilter).setFilterIfMissing(true);
+ }
+ @After
+ public void tearDown() throws Exception {
+ htable.close();
+ }
+
+ private void verify(Scan scan) throws IOException {
+ ResultScanner scanner = htable.getScanner(scan);
+ Iterator<Result> it = scanner.iterator();
+
+ /* Then */
+ int count = 0;
+ try {
+ while (it.hasNext()) {
+ it.next();
+ count++;
+ }
+ } finally {
+ scanner.close();
+ }
+ assertEquals(expected, count);
+ }
+ /**
+ * Test the filter by adding all columns of family A in the scan. (OK)
+ */
+ @Test
+ public void scanWithAllQualifiersOfFamiliyA() throws IOException {
+ /* Given */
+ Scan scan = new Scan();
+ scan.addFamily(FAMILY_A);
+ scan.setFilter(scanFilter);
+
+ verify(scan);
+ }
+
+ /**
+ * Test the filter by adding all columns of family A and B in the scan. (KO: row '3' without
+ * 'a:foo' qualifier is returned)
+ */
+ @Test
+ public void scanWithAllQualifiersOfBothFamilies() throws IOException {
+ /* When */
+ Scan scan = new Scan();
+ scan.setFilter(scanFilter);
+
+ verify(scan);
+ }
+
+ /**
+ * Test the filter by adding 2 columns of family A and 1 column of family B in the scan. (KO: row
+ * '3' without 'a:foo' qualifier is returned)
+ */
+ @Test
+ public void scanWithSpecificQualifiers1() throws IOException {
+ /* When */
+ Scan scan = new Scan();
+ scan.addColumn(FAMILY_A, QUALIFIER_FOO);
+ scan.addColumn(FAMILY_A, QUALIFIER_BAR);
+ scan.addColumn(FAMILY_B, QUALIFIER_BAR);
+ scan.addColumn(FAMILY_B, QUALIFIER_FOO);
+ scan.setFilter(scanFilter);
+
+ verify(scan);
+ }
+
+ /**
+ * Test the filter by adding 1 column of family A (the one used in the filter) and 1 column of
+ * family B in the scan. (OK)
+ */
+ @Test
+ public void scanWithSpecificQualifiers2() throws IOException {
+ /* When */
+ Scan scan = new Scan();
+ scan.addColumn(FAMILY_A, QUALIFIER_FOO);
+ scan.addColumn(FAMILY_B, QUALIFIER_BAR);
+ scan.setFilter(scanFilter);
+
+ verify(scan);
+ }
+
+ /**
+ * Test the filter by adding 2 columns of family A in the scan. (OK)
+ */
+ @Test
+ public void scanWithSpecificQualifiers3() throws IOException {
+ /* When */
+ Scan scan = new Scan();
+ scan.addColumn(FAMILY_A, QUALIFIER_FOO);
+ scan.addColumn(FAMILY_A, QUALIFIER_BAR);
+ scan.setFilter(scanFilter);
+
+ verify(scan);
+ }
+
+ private static void create(HBaseAdmin admin, String tableName, byte[]... families)
+ throws IOException {
+ HTableDescriptor desc = new HTableDescriptor(TableName.valueOf(tableName));
+ for (byte[] family : families) {
+ HColumnDescriptor colDesc = new HColumnDescriptor(family);
+ colDesc.setMaxVersions(1);
+ colDesc.setCompressionType(Algorithm.GZ);
+ desc.addFamily(colDesc);
+ }
+ try {
+ admin.createTable(desc);
+ } catch (TableExistsException tee) {
+ /* Ignore */
+ }
+ }
+
+ private static void destroy(HBaseAdmin admin, String tableName) throws IOException {
+ try {
+ admin.disableTable(tableName);
+ admin.deleteTable(tableName);
+ } catch (TableNotFoundException tnfe) {
+ /* Ignore */
+ }
+ }
+}