You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by sy...@apache.org on 2017/01/17 23:40:09 UTC
[1/7] hbase git commit: HBASE-17470 Remove merge region code from
region server (Stephen Yuan Jiang)
Repository: hbase
Updated Branches:
refs/heads/master 287f95a57 -> 805d39fca
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionMergeTransaction.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionMergeTransaction.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionMergeTransaction.java
deleted file mode 100644
index 5bd2ff1..0000000
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionMergeTransaction.java
+++ /dev/null
@@ -1,485 +0,0 @@
-/**
- * Copyright The Apache Software Foundation
- *
- * Licensed to the Apache Software Foundation (ASF) under one or more
- * contributor license agreements. See the NOTICE file distributed with this
- * work for additional information regarding copyright ownership. The ASF
- * licenses this file to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance with the License.
- * You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
- * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
- * License for the specific language governing permissions and limitations
- * under the License.
- */
-package org.apache.hadoop.hbase.regionserver;
-
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
-import static org.junit.Assert.assertTrue;
-import static org.mockito.Mockito.*;
-
-import java.io.IOException;
-
-import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.fs.FileSystem;
-import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.hbase.CoordinatedStateManager;
-import org.apache.hadoop.hbase.CoordinatedStateManagerFactory;
-import org.apache.hadoop.hbase.HBaseTestingUtility;
-import org.apache.hadoop.hbase.HColumnDescriptor;
-import org.apache.hadoop.hbase.HConstants;
-import org.apache.hadoop.hbase.HRegionInfo;
-import org.apache.hadoop.hbase.HTableDescriptor;
-import org.apache.hadoop.hbase.Server;
-import org.apache.hadoop.hbase.TableName;
-import org.apache.hadoop.hbase.client.Durability;
-import org.apache.hadoop.hbase.client.Put;
-import org.apache.hadoop.hbase.testclassification.RegionServerTests;
-import org.apache.hadoop.hbase.testclassification.SmallTests;
-import org.apache.hadoop.hbase.util.Bytes;
-import org.apache.hadoop.hbase.util.FSUtils;
-import org.apache.hadoop.hbase.wal.WALFactory;
-import org.apache.zookeeper.KeeperException;
-import org.junit.After;
-import org.junit.Before;
-import org.junit.Test;
-import org.junit.experimental.categories.Category;
-import org.mockito.Mockito;
-
-import com.google.common.collect.ImmutableList;
-
-/**
- * Test the {@link RegionMergeTransactionImpl} class against two HRegions (as
- * opposed to running cluster).
- */
-@Category({RegionServerTests.class, SmallTests.class})
-public class TestRegionMergeTransaction {
- private final HBaseTestingUtility TEST_UTIL = new HBaseTestingUtility();
- private final Path testdir = TEST_UTIL.getDataTestDir(this.getClass()
- .getName());
- private HRegion region_a;
- private HRegion region_b;
- private HRegion region_c;
- private WALFactory wals;
- private FileSystem fs;
- // Start rows of region_a,region_b,region_c
- private static final byte[] STARTROW_A = new byte[] { 'a', 'a', 'a' };
- private static final byte[] STARTROW_B = new byte[] { 'g', 'g', 'g' };
- private static final byte[] STARTROW_C = new byte[] { 'w', 'w', 'w' };
- private static final byte[] ENDROW = new byte[] { '{', '{', '{' };
- private static final byte[] CF = HConstants.CATALOG_FAMILY;
-
- @Before
- public void setup() throws IOException {
- this.fs = FileSystem.get(TEST_UTIL.getConfiguration());
- this.fs.delete(this.testdir, true);
- final Configuration walConf = new Configuration(TEST_UTIL.getConfiguration());
- FSUtils.setRootDir(walConf, this.testdir);
- this.wals = new WALFactory(walConf, null, TestRegionMergeTransaction.class.getName());
- this.region_a = createRegion(this.testdir, this.wals, STARTROW_A, STARTROW_B);
- this.region_b = createRegion(this.testdir, this.wals, STARTROW_B, STARTROW_C);
- this.region_c = createRegion(this.testdir, this.wals, STARTROW_C, ENDROW);
- assert region_a != null && region_b != null && region_c != null;
- TEST_UTIL.getConfiguration().setBoolean("hbase.testing.nocluster", true);
- }
-
- @After
- public void teardown() throws IOException {
- for (HRegion region : new HRegion[] { region_a, region_b, region_c }) {
- if (region != null && !region.isClosed()) region.close();
- if (this.fs.exists(region.getRegionFileSystem().getRegionDir())
- && !this.fs.delete(region.getRegionFileSystem().getRegionDir(), true)) {
- throw new IOException("Failed deleting of "
- + region.getRegionFileSystem().getRegionDir());
- }
- }
- if (this.wals != null) {
- this.wals.close();
- }
- this.fs.delete(this.testdir, true);
- }
-
- /**
- * Test straight prepare works. Tries to merge on {@link #region_a} and
- * {@link #region_b}
- * @throws IOException
- */
- @Test
- public void testPrepare() throws IOException {
- prepareOnGoodRegions();
- }
-
- private RegionMergeTransactionImpl prepareOnGoodRegions() throws IOException {
- RegionMergeTransactionImpl mt = new RegionMergeTransactionImpl(region_a, region_b,
- false);
- RegionMergeTransactionImpl spyMT = Mockito.spy(mt);
- doReturn(false).when(spyMT).hasMergeQualifierInMeta(null,
- region_a.getRegionInfo().getRegionName());
- doReturn(false).when(spyMT).hasMergeQualifierInMeta(null,
- region_b.getRegionInfo().getRegionName());
- assertTrue(spyMT.prepare(null));
- return spyMT;
- }
-
- /**
- * Test merging the same region
- */
- @Test
- public void testPrepareWithSameRegion() throws IOException {
- RegionMergeTransactionImpl mt = new RegionMergeTransactionImpl(this.region_a,
- this.region_a, true);
- assertFalse("should not merge the same region even if it is forcible ",
- mt.prepare(null));
- }
-
- /**
- * Test merging two not adjacent regions under a common merge
- */
- @Test
- public void testPrepareWithRegionsNotAdjacent() throws IOException {
- RegionMergeTransactionImpl mt = new RegionMergeTransactionImpl(this.region_a,
- this.region_c, false);
- assertFalse("should not merge two regions if they are adjacent except it is forcible",
- mt.prepare(null));
- }
-
- /**
- * Test merging two not adjacent regions under a compulsory merge
- */
- @Test
- public void testPrepareWithRegionsNotAdjacentUnderCompulsory()
- throws IOException {
- RegionMergeTransactionImpl mt = new RegionMergeTransactionImpl(region_a, region_c,
- true);
- RegionMergeTransactionImpl spyMT = Mockito.spy(mt);
- doReturn(false).when(spyMT).hasMergeQualifierInMeta(null,
- region_a.getRegionInfo().getRegionName());
- doReturn(false).when(spyMT).hasMergeQualifierInMeta(null,
- region_c.getRegionInfo().getRegionName());
- assertTrue("Since focible is true, should merge two regions even if they are not adjacent",
- spyMT.prepare(null));
- }
-
- /**
- * Pass a reference store
- */
- @Test
- public void testPrepareWithRegionsWithReference() throws IOException {
- HStore storeMock = Mockito.mock(HStore.class);
- when(storeMock.hasReferences()).thenReturn(true);
- when(storeMock.getFamily()).thenReturn(new HColumnDescriptor("cf"));
- when(storeMock.getSizeToFlush()).thenReturn(new MemstoreSize());
- when(storeMock.close()).thenReturn(ImmutableList.<StoreFile>of());
- this.region_a.stores.put(Bytes.toBytes(""), storeMock);
- RegionMergeTransactionImpl mt = new RegionMergeTransactionImpl(this.region_a,
- this.region_b, false);
- assertFalse(
- "a region should not be mergeable if it has instances of store file references",
- mt.prepare(null));
- }
-
- @Test
- public void testPrepareWithClosedRegion() throws IOException {
- this.region_a.close();
- RegionMergeTransactionImpl mt = new RegionMergeTransactionImpl(this.region_a,
- this.region_b, false);
- assertFalse(mt.prepare(null));
- }
-
- /**
- * Test merging regions which are merged regions and has reference in hbase:meta all
- * the same
- */
- @Test
- public void testPrepareWithRegionsWithMergeReference() throws IOException {
- RegionMergeTransactionImpl mt = new RegionMergeTransactionImpl(region_a, region_b,
- false);
- RegionMergeTransactionImpl spyMT = Mockito.spy(mt);
- doReturn(true).when(spyMT).hasMergeQualifierInMeta(null,
- region_a.getRegionInfo().getRegionName());
- doReturn(true).when(spyMT).hasMergeQualifierInMeta(null,
- region_b.getRegionInfo().getRegionName());
- assertFalse(spyMT.prepare(null));
- }
-
- /**
- * Test RegionMergeTransactionListener
- */
- @Test public void testRegionMergeTransactionListener() throws IOException {
- RegionMergeTransactionImpl mt = new RegionMergeTransactionImpl(region_a, region_b,
- false);
- RegionMergeTransactionImpl spyMT = Mockito.spy(mt);
- doReturn(false).when(spyMT).hasMergeQualifierInMeta(null,
- region_a.getRegionInfo().getRegionName());
- doReturn(false).when(spyMT).hasMergeQualifierInMeta(null,
- region_b.getRegionInfo().getRegionName());
- RegionMergeTransaction.TransactionListener listener =
- Mockito.mock(RegionMergeTransaction.TransactionListener.class);
- mt.registerTransactionListener(listener);
- mt.prepare(null);
- TEST_UTIL.getConfiguration().setInt(HConstants.REGIONSERVER_PORT, 0);
- CoordinatedStateManager cp = CoordinatedStateManagerFactory.getCoordinatedStateManager(
- TEST_UTIL.getConfiguration());
- Server mockServer = new HRegionServer(TEST_UTIL.getConfiguration(), cp);
- mt.execute(mockServer, null);
- verify(listener).transition(mt,
- RegionMergeTransaction.RegionMergeTransactionPhase.STARTED,
- RegionMergeTransaction.RegionMergeTransactionPhase.PREPARED);
- verify(listener, times(10)).transition(any(RegionMergeTransaction.class),
- any(RegionMergeTransaction.RegionMergeTransactionPhase.class),
- any(RegionMergeTransaction.RegionMergeTransactionPhase.class));
- verifyNoMoreInteractions(listener);
- }
-
- @Test
- public void testWholesomeMerge() throws IOException, InterruptedException {
- final int rowCountOfRegionA = loadRegion(this.region_a, CF, true);
- final int rowCountOfRegionB = loadRegion(this.region_b, CF, true);
- assertTrue(rowCountOfRegionA > 0 && rowCountOfRegionB > 0);
- assertEquals(rowCountOfRegionA, TEST_UTIL.countRows(this.region_a));
- assertEquals(rowCountOfRegionB, TEST_UTIL.countRows(this.region_b));
-
- // Start transaction.
- RegionMergeTransactionImpl mt = prepareOnGoodRegions();
-
- // Run the execute. Look at what it returns.
- TEST_UTIL.getConfiguration().setInt(HConstants.REGIONSERVER_PORT, 0);
- CoordinatedStateManager cp = CoordinatedStateManagerFactory.getCoordinatedStateManager(
- TEST_UTIL.getConfiguration());
- Server mockServer = new HRegionServer(TEST_UTIL.getConfiguration(), cp);
- HRegion mergedRegion = (HRegion)mt.execute(mockServer, null);
- // Do some assertions about execution.
- assertTrue(this.fs.exists(mt.getMergesDir()));
- // Assert region_a and region_b is closed.
- assertTrue(region_a.isClosed());
- assertTrue(region_b.isClosed());
-
- // Assert mergedir is empty -- because its content will have been moved out
- // to be under the merged region dirs.
- assertEquals(0, this.fs.listStatus(mt.getMergesDir()).length);
- // Check merged region have correct key span.
- assertTrue(Bytes.equals(this.region_a.getRegionInfo().getStartKey(),
- mergedRegion.getRegionInfo().getStartKey()));
- assertTrue(Bytes.equals(this.region_b.getRegionInfo().getEndKey(),
- mergedRegion.getRegionInfo().getEndKey()));
- // Count rows. merged region are already open
- try {
- int mergedRegionRowCount = TEST_UTIL.countRows(mergedRegion);
- assertEquals((rowCountOfRegionA + rowCountOfRegionB),
- mergedRegionRowCount);
- } finally {
- HBaseTestingUtility.closeRegionAndWAL(mergedRegion);
- }
- // Assert the write lock is no longer held on region_a and region_b
- assertTrue(!this.region_a.lock.writeLock().isHeldByCurrentThread());
- assertTrue(!this.region_b.lock.writeLock().isHeldByCurrentThread());
- }
-
- @Test
- public void testRollback() throws IOException, InterruptedException {
- final int rowCountOfRegionA = loadRegion(this.region_a, CF, true);
- final int rowCountOfRegionB = loadRegion(this.region_b, CF, true);
- assertTrue(rowCountOfRegionA > 0 && rowCountOfRegionB > 0);
- assertEquals(rowCountOfRegionA, TEST_UTIL.countRows(this.region_a));
- assertEquals(rowCountOfRegionB, TEST_UTIL.countRows(this.region_b));
-
- // Start transaction.
- RegionMergeTransactionImpl mt = prepareOnGoodRegions();
-
- when(mt.createMergedRegionFromMerges(region_a, region_b,
- mt.getMergedRegionInfo())).thenThrow(
- new MockedFailedMergedRegionCreation());
-
- // Run the execute. Look at what it returns.
- boolean expectedException = false;
- TEST_UTIL.getConfiguration().setInt(HConstants.REGIONSERVER_PORT, 0);
- CoordinatedStateManager cp = CoordinatedStateManagerFactory.getCoordinatedStateManager(
- TEST_UTIL.getConfiguration());
- Server mockServer = new HRegionServer(TEST_UTIL.getConfiguration(), cp);
- try {
- mt.execute(mockServer, null);
- } catch (MockedFailedMergedRegionCreation e) {
- expectedException = true;
- }
- assertTrue(expectedException);
- // Run rollback
- assertTrue(mt.rollback(null, null));
-
- // Assert I can scan region_a and region_b.
- int rowCountOfRegionA2 = TEST_UTIL.countRows(this.region_a);
- assertEquals(rowCountOfRegionA, rowCountOfRegionA2);
- int rowCountOfRegionB2 = TEST_UTIL.countRows(this.region_b);
- assertEquals(rowCountOfRegionB, rowCountOfRegionB2);
-
- // Assert rollback cleaned up stuff in fs
- assertTrue(!this.fs.exists(HRegion.getRegionDir(this.testdir,
- mt.getMergedRegionInfo())));
-
- assertTrue(!this.region_a.lock.writeLock().isHeldByCurrentThread());
- assertTrue(!this.region_b.lock.writeLock().isHeldByCurrentThread());
-
- // Now retry the merge but do not throw an exception this time.
- assertTrue(mt.prepare(null));
- HRegion mergedRegion = (HRegion)mt.execute(mockServer, null);
- // Count rows. daughters are already open
- // Count rows. merged region are already open
- try {
- int mergedRegionRowCount = TEST_UTIL.countRows(mergedRegion);
- assertEquals((rowCountOfRegionA + rowCountOfRegionB),
- mergedRegionRowCount);
- } finally {
- HBaseTestingUtility.closeRegionAndWAL(mergedRegion);
- }
- // Assert the write lock is no longer held on region_a and region_b
- assertTrue(!this.region_a.lock.writeLock().isHeldByCurrentThread());
- assertTrue(!this.region_b.lock.writeLock().isHeldByCurrentThread());
- }
-
- @Test
- public void testFailAfterPONR() throws IOException, KeeperException, InterruptedException {
- final int rowCountOfRegionA = loadRegion(this.region_a, CF, true);
- final int rowCountOfRegionB = loadRegion(this.region_b, CF, true);
- assertTrue(rowCountOfRegionA > 0 && rowCountOfRegionB > 0);
- assertEquals(rowCountOfRegionA, TEST_UTIL.countRows(this.region_a));
- assertEquals(rowCountOfRegionB, TEST_UTIL.countRows(this.region_b));
-
- // Start transaction.
- RegionMergeTransactionImpl mt = prepareOnGoodRegions();
- Mockito.doThrow(new MockedFailedMergedRegionOpen())
- .when(mt)
- .openMergedRegion((Server) Mockito.anyObject(),
- (RegionServerServices) Mockito.anyObject(),
- (HRegion) Mockito.anyObject());
-
- // Run the execute. Look at what it returns.
- boolean expectedException = false;
- TEST_UTIL.getConfiguration().setInt(HConstants.REGIONSERVER_PORT, 0);
- CoordinatedStateManager cp = CoordinatedStateManagerFactory.getCoordinatedStateManager(
- TEST_UTIL.getConfiguration());
- Server mockServer = new HRegionServer(TEST_UTIL.getConfiguration(), cp);
- try {
- mt.execute(mockServer, null);
- } catch (MockedFailedMergedRegionOpen e) {
- expectedException = true;
- }
- assertTrue(expectedException);
- // Run rollback returns false that we should restart.
- assertFalse(mt.rollback(null, null));
- // Make sure that merged region is still in the filesystem, that
- // they have not been removed; this is supposed to be the case if we go
- // past point of no return.
- Path tableDir = this.region_a.getRegionFileSystem().getRegionDir()
- .getParent();
- Path mergedRegionDir = new Path(tableDir, mt.getMergedRegionInfo()
- .getEncodedName());
- assertTrue(TEST_UTIL.getTestFileSystem().exists(mergedRegionDir));
- }
-
- @Test
- public void testMergedRegionBoundary() {
- TableName tableName =
- TableName.valueOf("testMergedRegionBoundary");
- byte[] a = Bytes.toBytes("a");
- byte[] b = Bytes.toBytes("b");
- byte[] z = Bytes.toBytes("z");
- HRegionInfo r1 = new HRegionInfo(tableName);
- HRegionInfo r2 = new HRegionInfo(tableName, a, z);
- HRegionInfo m = RegionMergeTransactionImpl.getMergedRegionInfo(r1, r2);
- assertTrue(Bytes.equals(m.getStartKey(), r1.getStartKey())
- && Bytes.equals(m.getEndKey(), r1.getEndKey()));
-
- r1 = new HRegionInfo(tableName, null, a);
- r2 = new HRegionInfo(tableName, a, z);
- m = RegionMergeTransactionImpl.getMergedRegionInfo(r1, r2);
- assertTrue(Bytes.equals(m.getStartKey(), r1.getStartKey())
- && Bytes.equals(m.getEndKey(), r2.getEndKey()));
-
- r1 = new HRegionInfo(tableName, null, a);
- r2 = new HRegionInfo(tableName, z, null);
- m = RegionMergeTransactionImpl.getMergedRegionInfo(r1, r2);
- assertTrue(Bytes.equals(m.getStartKey(), r1.getStartKey())
- && Bytes.equals(m.getEndKey(), r2.getEndKey()));
-
- r1 = new HRegionInfo(tableName, a, z);
- r2 = new HRegionInfo(tableName, z, null);
- m = RegionMergeTransactionImpl.getMergedRegionInfo(r1, r2);
- assertTrue(Bytes.equals(m.getStartKey(), r1.getStartKey())
- && Bytes.equals(m.getEndKey(), r2.getEndKey()));
-
- r1 = new HRegionInfo(tableName, a, b);
- r2 = new HRegionInfo(tableName, b, z);
- m = RegionMergeTransactionImpl.getMergedRegionInfo(r1, r2);
- assertTrue(Bytes.equals(m.getStartKey(), r1.getStartKey())
- && Bytes.equals(m.getEndKey(), r2.getEndKey()));
- }
-
- /**
- * Exception used in this class only.
- */
- @SuppressWarnings("serial")
- private class MockedFailedMergedRegionCreation extends IOException {
- }
-
- @SuppressWarnings("serial")
- private class MockedFailedMergedRegionOpen extends IOException {
- }
-
- private HRegion createRegion(final Path testdir, final WALFactory wals,
- final byte[] startrow, final byte[] endrow)
- throws IOException {
- // Make a region with start and end keys.
- HTableDescriptor htd = new HTableDescriptor(TableName.valueOf("table"));
- HColumnDescriptor hcd = new HColumnDescriptor(CF);
- htd.addFamily(hcd);
- HRegionInfo hri = new HRegionInfo(htd.getTableName(), startrow, endrow);
- HRegion a = HBaseTestingUtility.createRegionAndWAL(hri, testdir,
- TEST_UTIL.getConfiguration(), htd);
- HBaseTestingUtility.closeRegionAndWAL(a);
- return HRegion.openHRegion(testdir, hri, htd,
- wals.getWAL(hri.getEncodedNameAsBytes(), hri.getTable().getNamespace()),
- TEST_UTIL.getConfiguration());
- }
-
- /**
- * Load region with rows from 'aaa' to 'zzz', skip the rows which are out of
- * range of the region
- * @param r Region
- * @param f Family
- * @param flush flush the cache if true
- * @return Count of rows loaded.
- * @throws IOException
- */
- private int loadRegion(final HRegion r, final byte[] f, final boolean flush)
- throws IOException {
- byte[] k = new byte[3];
- int rowCount = 0;
- for (byte b1 = 'a'; b1 <= 'z'; b1++) {
- for (byte b2 = 'a'; b2 <= 'z'; b2++) {
- for (byte b3 = 'a'; b3 <= 'z'; b3++) {
- k[0] = b1;
- k[1] = b2;
- k[2] = b3;
- if (!HRegion.rowIsInRange(r.getRegionInfo(), k)) {
- continue;
- }
- Put put = new Put(k);
- put.addColumn(f, null, k);
- if (r.getWAL() == null)
- put.setDurability(Durability.SKIP_WAL);
- r.put(put);
- rowCount++;
- }
- }
- if (flush) {
- r.flush(true);
- }
- }
- return rowCount;
- }
-
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMergeTable.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMergeTable.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMergeTable.java
deleted file mode 100644
index 661af14..0000000
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMergeTable.java
+++ /dev/null
@@ -1,172 +0,0 @@
-/**
- *
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.hadoop.hbase.util;
-
-import static org.junit.Assert.assertTrue;
-
-import java.io.IOException;
-import java.util.List;
-
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
-import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.fs.FileSystem;
-import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.hbase.*;
-import org.apache.hadoop.hbase.client.Admin;
-import org.apache.hadoop.hbase.client.Connection;
-import org.apache.hadoop.hbase.client.Put;
-import org.apache.hadoop.hbase.client.Durability;
-import org.apache.hadoop.hbase.regionserver.HRegion;
-import org.apache.hadoop.hbase.testclassification.MediumTests;
-import org.apache.hadoop.hbase.testclassification.MiscTests;
-import org.junit.Test;
-import org.junit.experimental.categories.Category;
-
-/**
- * Tests merging a normal table's regions
- */
-@Category({MiscTests.class, MediumTests.class})
-public class TestMergeTable {
- private static final Log LOG = LogFactory.getLog(TestMergeTable.class);
- private final HBaseTestingUtility UTIL = new HBaseTestingUtility();
- private static final byte [] COLUMN_NAME = Bytes.toBytes("contents");
- private static final byte [] VALUE;
- static {
- // We will use the same value for the rows as that is not really important here
- String partialValue = String.valueOf(System.currentTimeMillis());
- StringBuilder val = new StringBuilder();
- while (val.length() < 1024) {
- val.append(partialValue);
- }
- VALUE = Bytes.toBytes(val.toString());
- }
-
- /**
- * Test merge.
- * Hand-makes regions of a mergeable size and adds the hand-made regions to
- * hand-made meta. The hand-made regions are created offline. We then start
- * up mini cluster, disables the hand-made table and starts in on merging.
- * @throws Exception
- */
- @Test (timeout=300000) public void testMergeTable() throws Exception {
- // Table we are manually creating offline.
- HTableDescriptor desc = new HTableDescriptor(org.apache.hadoop.hbase.TableName.valueOf(Bytes.toBytes("test")));
- desc.addFamily(new HColumnDescriptor(COLUMN_NAME));
-
- // Set maximum regionsize down.
- UTIL.getConfiguration().setLong(HConstants.HREGION_MAX_FILESIZE, 64L * 1024L * 1024L);
- // Make it so we don't split.
- UTIL.getConfiguration().setInt("hbase.regionserver.regionSplitLimit", 0);
- // Startup hdfs. Its in here we'll be putting our manually made regions.
- UTIL.startMiniDFSCluster(1);
- // Create hdfs hbase rootdir.
- Path rootdir = UTIL.createRootDir();
- FileSystem fs = FileSystem.get(UTIL.getConfiguration());
- if (fs.exists(rootdir)) {
- if (fs.delete(rootdir, true)) {
- LOG.info("Cleaned up existing " + rootdir);
- }
- }
-
- // Now create three data regions: The first is too large to merge since it
- // will be > 64 MB in size. The second two will be smaller and will be
- // selected for merging.
-
- // To ensure that the first region is larger than 64MB we need to write at
- // least 65536 rows. We will make certain by writing 70000
- byte [] row_70001 = Bytes.toBytes("row_70001");
- byte [] row_80001 = Bytes.toBytes("row_80001");
-
- // Create regions and populate them at same time. Create the tabledir
- // for them first.
- new FSTableDescriptors(UTIL.getConfiguration(), fs, rootdir).createTableDescriptor(desc);
- HRegion [] regions = {
- createRegion(desc, null, row_70001, 1, 70000, rootdir),
- createRegion(desc, row_70001, row_80001, 70001, 10000, rootdir),
- createRegion(desc, row_80001, null, 80001, 11000, rootdir)
- };
-
- // Now create the root and meta regions and insert the data regions
- // created above into hbase:meta
- setupMeta(rootdir, regions);
- try {
- LOG.info("Starting mini zk cluster");
- UTIL.startMiniZKCluster();
- LOG.info("Starting mini hbase cluster");
- UTIL.startMiniHBaseCluster(1, 1);
- Configuration c = new Configuration(UTIL.getConfiguration());
- Connection connection = UTIL.getConnection();
-
- List<HRegionInfo> originalTableRegions =
- MetaTableAccessor.getTableRegions(connection, desc.getTableName());
- LOG.info("originalTableRegions size=" + originalTableRegions.size() +
- "; " + originalTableRegions);
- Admin admin = connection.getAdmin();
- admin.disableTable(desc.getTableName());
- admin.close();
- HMerge.merge(c, FileSystem.get(c), desc.getTableName());
- List<HRegionInfo> postMergeTableRegions =
- MetaTableAccessor.getTableRegions(connection, desc.getTableName());
- LOG.info("postMergeTableRegions size=" + postMergeTableRegions.size() +
- "; " + postMergeTableRegions);
- assertTrue("originalTableRegions=" + originalTableRegions.size() +
- ", postMergeTableRegions=" + postMergeTableRegions.size(),
- postMergeTableRegions.size() < originalTableRegions.size());
- LOG.info("Done with merge");
- } finally {
- UTIL.shutdownMiniCluster();
- LOG.info("After cluster shutdown");
- }
- }
-
- private HRegion createRegion(final HTableDescriptor desc,
- byte [] startKey, byte [] endKey, int firstRow, int nrows, Path rootdir)
- throws IOException {
- HRegionInfo hri = new HRegionInfo(desc.getTableName(), startKey, endKey);
- HRegion region = HBaseTestingUtility.createRegionAndWAL(hri, rootdir, UTIL.getConfiguration(),
- desc);
- LOG.info("Created region " + region.getRegionInfo().getRegionNameAsString());
- for(int i = firstRow; i < firstRow + nrows; i++) {
- Put put = new Put(Bytes.toBytes("row_" + String.format("%1$05d", i)));
- put.setDurability(Durability.SKIP_WAL);
- put.addColumn(COLUMN_NAME, null, VALUE);
- region.put(put);
- if (i % 10000 == 0) {
- LOG.info("Flushing write #" + i);
- region.flush(true);
- }
- }
- HBaseTestingUtility.closeRegionAndWAL(region);
- return region;
- }
-
- protected void setupMeta(Path rootdir, final HRegion [] regions)
- throws IOException {
- HRegion meta =
- HBaseTestingUtility.createRegionAndWAL(HRegionInfo.FIRST_META_REGIONINFO, rootdir,
- UTIL.getConfiguration(), UTIL.getMetaTableDescriptor());
- for (HRegion r: regions) {
- HRegion.addRegionToMETA(meta, r);
- }
- HBaseTestingUtility.closeRegionAndWAL(meta);
- }
-
-}
-
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMergeTool.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMergeTool.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMergeTool.java
deleted file mode 100644
index 1924c9e..0000000
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMergeTool.java
+++ /dev/null
@@ -1,301 +0,0 @@
-/**
- *
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.hadoop.hbase.util;
-
-import java.io.IOException;
-import java.util.ArrayList;
-import java.util.List;
-
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
-import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.hbase.Cell;
-import org.apache.hadoop.hbase.CellUtil;
-import org.apache.hadoop.hbase.HBaseTestCase;
-import org.apache.hadoop.hbase.HBaseTestingUtility;
-import org.apache.hadoop.hbase.HColumnDescriptor;
-import org.apache.hadoop.hbase.HConstants;
-import org.apache.hadoop.hbase.HRegionInfo;
-import org.apache.hadoop.hbase.HTableDescriptor;
-import org.apache.hadoop.hbase.client.Get;
-import org.apache.hadoop.hbase.client.Put;
-import org.apache.hadoop.hbase.client.Result;
-import org.apache.hadoop.hbase.client.Scan;
-import org.apache.hadoop.hbase.regionserver.HRegion;
-import org.apache.hadoop.hbase.regionserver.InternalScanner;
-import org.apache.hadoop.hbase.testclassification.LargeTests;
-import org.apache.hadoop.hbase.testclassification.MiscTests;
-import org.apache.hadoop.hbase.wal.WAL;
-import org.apache.hadoop.hbase.wal.WALFactory;
-import org.apache.hadoop.hdfs.MiniDFSCluster;
-import org.apache.hadoop.util.ToolRunner;
-import org.junit.After;
-import org.junit.Before;
-import org.junit.Test;
-import org.junit.experimental.categories.Category;
-
-/** Test stand alone merge tool that can merge arbitrary regions */
-@Category({MiscTests.class, LargeTests.class})
-public class TestMergeTool extends HBaseTestCase {
- private static final Log LOG = LogFactory.getLog(TestMergeTool.class);
- HBaseTestingUtility TEST_UTIL;
-// static final byte [] COLUMN_NAME = Bytes.toBytes("contents:");
- static final byte [] FAMILY = Bytes.toBytes("contents");
- static final byte [] QUALIFIER = Bytes.toBytes("dc");
-
- private final HRegionInfo[] sourceRegions = new HRegionInfo[5];
- private final HRegion[] regions = new HRegion[5];
- private HTableDescriptor desc;
- private byte [][][] rows;
- private MiniDFSCluster dfsCluster = null;
- private WALFactory wals;
-
- @Override
- @Before
- public void setUp() throws Exception {
- // Set the timeout down else this test will take a while to complete.
- this.conf.setLong("hbase.zookeeper.recoverable.waittime", 10);
- // Make it so we try and connect to a zk that is not there (else we might
- // find a zk ensemble put up by another concurrent test and this will
- // mess up this test. Choose unlikely port. Default test port is 21818.
- // Default zk port is 2181.
- this.conf.setInt(HConstants.ZOOKEEPER_CLIENT_PORT, 10001);
-
- this.conf.set("hbase.hstore.compactionThreshold", "2");
-
- // Create table description
- this.desc = new HTableDescriptor(org.apache.hadoop.hbase.TableName.valueOf("TestMergeTool"));
- this.desc.addFamily(new HColumnDescriptor(FAMILY));
-
- /*
- * Create the HRegionInfos for the regions.
- */
- // Region 0 will contain the key range [row_0200,row_0300)
- sourceRegions[0] = new HRegionInfo(this.desc.getTableName(),
- Bytes.toBytes("row_0200"),
- Bytes.toBytes("row_0300"));
-
- // Region 1 will contain the key range [row_0250,row_0400) and overlaps
- // with Region 0
- sourceRegions[1] =
- new HRegionInfo(this.desc.getTableName(),
- Bytes.toBytes("row_0250"),
- Bytes.toBytes("row_0400"));
-
- // Region 2 will contain the key range [row_0100,row_0200) and is adjacent
- // to Region 0 or the region resulting from the merge of Regions 0 and 1
- sourceRegions[2] =
- new HRegionInfo(this.desc.getTableName(),
- Bytes.toBytes("row_0100"),
- Bytes.toBytes("row_0200"));
-
- // Region 3 will contain the key range [row_0500,row_0600) and is not
- // adjacent to any of Regions 0, 1, 2 or the merged result of any or all
- // of those regions
- sourceRegions[3] =
- new HRegionInfo(this.desc.getTableName(),
- Bytes.toBytes("row_0500"),
- Bytes.toBytes("row_0600"));
-
- // Region 4 will have empty start and end keys and overlaps all regions.
- sourceRegions[4] =
- new HRegionInfo(this.desc.getTableName(),
- HConstants.EMPTY_BYTE_ARRAY,
- HConstants.EMPTY_BYTE_ARRAY);
-
- /*
- * Now create some row keys
- */
- this.rows = new byte [5][][];
- this.rows[0] = Bytes.toByteArrays(new String[] { "row_0210", "row_0280" });
- this.rows[1] = Bytes.toByteArrays(new String[] { "row_0260", "row_0350",
- "row_035" });
- this.rows[2] = Bytes.toByteArrays(new String[] { "row_0110", "row_0175",
- "row_0175", "row_0175"});
- this.rows[3] = Bytes.toByteArrays(new String[] { "row_0525", "row_0560",
- "row_0560", "row_0560", "row_0560"});
- this.rows[4] = Bytes.toByteArrays(new String[] { "row_0050", "row_1000",
- "row_1000", "row_1000", "row_1000", "row_1000" });
-
- // Start up dfs
- TEST_UTIL = new HBaseTestingUtility(conf);
- this.dfsCluster = TEST_UTIL.startMiniDFSCluster(2);
- this.fs = this.dfsCluster.getFileSystem();
- System.out.println("fs=" + this.fs);
- FSUtils.setFsDefault(this.conf, new Path(fs.getUri()));
- TEST_UTIL.createRootDir();
-
- // Note: we must call super.setUp after starting the mini cluster or
- // we will end up with a local file system
-
- super.setUp();
- wals = new WALFactory(conf, null, "TestMergeTool");
- try {
- // Create meta region
- createMetaRegion();
- new FSTableDescriptors(this.conf, this.fs, testDir).createTableDescriptor(this.desc);
- /*
- * Create the regions we will merge
- */
- for (int i = 0; i < sourceRegions.length; i++) {
- regions[i] =
- HBaseTestingUtility.createRegionAndWAL(this.sourceRegions[i], testDir, this.conf,
- this.desc);
- /*
- * Insert data
- */
- for (int j = 0; j < rows[i].length; j++) {
- byte [] row = rows[i][j];
- Put put = new Put(row);
- put.addColumn(FAMILY, QUALIFIER, row);
- regions[i].put(put);
- }
- HRegion.addRegionToMETA(meta, regions[i]);
- }
- // Close root and meta regions
- closeRootAndMeta();
-
- } catch (Exception e) {
- TEST_UTIL.shutdownMiniCluster();
- throw e;
- }
- }
-
- @Override
- @After
- public void tearDown() throws Exception {
- super.tearDown();
- for (int i = 0; i < sourceRegions.length; i++) {
- HRegion r = regions[i];
- if (r != null) {
- HBaseTestingUtility.closeRegionAndWAL(r);
- }
- }
- wals.close();
- TEST_UTIL.shutdownMiniCluster();
- }
-
- /*
- * @param msg Message that describes this merge
- * @param regionName1
- * @param regionName2
- * @param log Log to use merging.
- * @param upperbound Verifying, how high up in this.rows to go.
- * @return Merged region.
- * @throws Exception
- */
- private HRegion mergeAndVerify(final String msg, final String regionName1,
- final String regionName2, final WAL log, final int upperbound)
- throws Exception {
- Merge merger = new Merge();
- LOG.info(msg);
- LOG.info("fs2=" + this.conf.get("fs.defaultFS"));
- int errCode = ToolRunner.run(this.conf, merger,
- new String[] {this.desc.getTableName().getNameAsString(), regionName1, regionName2}
- );
- assertTrue("'" + msg + "' failed with errCode " + errCode, errCode == 0);
- HRegionInfo mergedInfo = merger.getMergedHRegionInfo();
-
- // Now verify that we can read all the rows from regions 0, 1
- // in the new merged region.
- HRegion merged = HRegion.openHRegion(mergedInfo, this.desc, log, this.conf);
- verifyMerge(merged, upperbound);
- merged.close();
- LOG.info("Verified " + msg);
- return merged;
- }
-
- private void verifyMerge(final HRegion merged, final int upperbound)
- throws IOException {
- //Test
- Scan scan = new Scan();
- scan.addFamily(FAMILY);
- InternalScanner scanner = merged.getScanner(scan);
- try {
- List<Cell> testRes = null;
- while (true) {
- testRes = new ArrayList<Cell>();
- boolean hasNext = scanner.next(testRes);
- if (!hasNext) {
- break;
- }
- }
- } finally {
- scanner.close();
- }
-
- //!Test
-
- for (int i = 0; i < upperbound; i++) {
- for (int j = 0; j < rows[i].length; j++) {
- Get get = new Get(rows[i][j]);
- get.addFamily(FAMILY);
- Result result = merged.get(get);
- assertEquals(1, result.size());
- byte [] bytes = CellUtil.cloneValue(result.rawCells()[0]);
- assertNotNull(Bytes.toStringBinary(rows[i][j]), bytes);
- assertTrue(Bytes.equals(bytes, rows[i][j]));
- }
- }
- }
-
- /**
- * Test merge tool.
- * @throws Exception
- */
- @Test
- public void testMergeTool() throws Exception {
- // First verify we can read the rows from the source regions and that they
- // contain the right data.
- for (int i = 0; i < regions.length; i++) {
- for (int j = 0; j < rows[i].length; j++) {
- Get get = new Get(rows[i][j]);
- get.addFamily(FAMILY);
- Result result = regions[i].get(get);
- byte [] bytes = CellUtil.cloneValue(result.rawCells()[0]);
- assertNotNull(bytes);
- assertTrue(Bytes.equals(bytes, rows[i][j]));
- }
- // Close the region and delete the log
- HBaseTestingUtility.closeRegionAndWAL(regions[i]);
- }
- WAL log = wals.getWAL(new byte[]{}, null);
- // Merge Region 0 and Region 1
- HRegion merged = mergeAndVerify("merging regions 0 and 1 ",
- this.sourceRegions[0].getRegionNameAsString(),
- this.sourceRegions[1].getRegionNameAsString(), log, 2);
-
- // Merge the result of merging regions 0 and 1 with region 2
- merged = mergeAndVerify("merging regions 0+1 and 2",
- merged.getRegionInfo().getRegionNameAsString(),
- this.sourceRegions[2].getRegionNameAsString(), log, 3);
-
- // Merge the result of merging regions 0, 1 and 2 with region 3
- merged = mergeAndVerify("merging regions 0+1+2 and 3",
- merged.getRegionInfo().getRegionNameAsString(),
- this.sourceRegions[3].getRegionNameAsString(), log, 4);
-
- // Merge the result of merging regions 0, 1, 2 and 3 with region 4
- merged = mergeAndVerify("merging regions 0+1+2+3 and 4",
- merged.getRegionInfo().getRegionNameAsString(),
- this.sourceRegions[4].getRegionNameAsString(), log, rows.length);
- }
-
-}
-
[6/7] hbase git commit: HBASE-17470 Remove merge region code from
region server (Stephen Yuan Jiang)
Posted by sy...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AdminProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AdminProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AdminProtos.java
index 654d152..b4bd84d 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AdminProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AdminProtos.java
@@ -16538,1406 +16538,6 @@ public final class AdminProtos {
}
- public interface MergeRegionsRequestOrBuilder extends
- // @@protoc_insertion_point(interface_extends:hbase.pb.MergeRegionsRequest)
- org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
- /**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
- */
- boolean hasRegionA();
- /**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
- */
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA();
- /**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
- */
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder();
-
- /**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
- */
- boolean hasRegionB();
- /**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
- */
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB();
- /**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
- */
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder();
-
- /**
- * <code>optional bool forcible = 3 [default = false];</code>
- */
- boolean hasForcible();
- /**
- * <code>optional bool forcible = 3 [default = false];</code>
- */
- boolean getForcible();
-
- /**
- * <pre>
- * wall clock time from master
- * </pre>
- *
- * <code>optional uint64 master_system_time = 4;</code>
- */
- boolean hasMasterSystemTime();
- /**
- * <pre>
- * wall clock time from master
- * </pre>
- *
- * <code>optional uint64 master_system_time = 4;</code>
- */
- long getMasterSystemTime();
- }
- /**
- * <pre>
- **
- * Merges the specified regions.
- * <p>
- * This method currently closes the regions and then merges them
- * </pre>
- *
- * Protobuf type {@code hbase.pb.MergeRegionsRequest}
- */
- public static final class MergeRegionsRequest extends
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
- // @@protoc_insertion_point(message_implements:hbase.pb.MergeRegionsRequest)
- MergeRegionsRequestOrBuilder {
- // Use MergeRegionsRequest.newBuilder() to construct.
- private MergeRegionsRequest(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
- super(builder);
- }
- private MergeRegionsRequest() {
- forcible_ = false;
- masterSystemTime_ = 0L;
- }
-
- @java.lang.Override
- public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
- return this.unknownFields;
- }
- private MergeRegionsRequest(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- this();
- int mutable_bitField0_ = 0;
- org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
- try {
- boolean done = false;
- while (!done) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
- done = true;
- break;
- default: {
- if (!parseUnknownField(input, unknownFields,
- extensionRegistry, tag)) {
- done = true;
- }
- break;
- }
- case 10: {
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = null;
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
- subBuilder = regionA_.toBuilder();
- }
- regionA_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry);
- if (subBuilder != null) {
- subBuilder.mergeFrom(regionA_);
- regionA_ = subBuilder.buildPartial();
- }
- bitField0_ |= 0x00000001;
- break;
- }
- case 18: {
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = null;
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
- subBuilder = regionB_.toBuilder();
- }
- regionB_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry);
- if (subBuilder != null) {
- subBuilder.mergeFrom(regionB_);
- regionB_ = subBuilder.buildPartial();
- }
- bitField0_ |= 0x00000002;
- break;
- }
- case 24: {
- bitField0_ |= 0x00000004;
- forcible_ = input.readBool();
- break;
- }
- case 32: {
- bitField0_ |= 0x00000008;
- masterSystemTime_ = input.readUInt64();
- break;
- }
- }
- }
- } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
- throw e.setUnfinishedMessage(this);
- } catch (java.io.IOException e) {
- throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
- e).setUnfinishedMessage(this);
- } finally {
- this.unknownFields = unknownFields.build();
- makeExtensionsImmutable();
- }
- }
- public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
- getDescriptor() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_MergeRegionsRequest_descriptor;
- }
-
- protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
- internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_MergeRegionsRequest_fieldAccessorTable
- .ensureFieldAccessorsInitialized(
- org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest.Builder.class);
- }
-
- private int bitField0_;
- public static final int REGION_A_FIELD_NUMBER = 1;
- private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier regionA_;
- /**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
- */
- public boolean hasRegionA() {
- return ((bitField0_ & 0x00000001) == 0x00000001);
- }
- /**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA() {
- return regionA_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionA_;
- }
- /**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder() {
- return regionA_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionA_;
- }
-
- public static final int REGION_B_FIELD_NUMBER = 2;
- private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier regionB_;
- /**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
- */
- public boolean hasRegionB() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
- }
- /**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB() {
- return regionB_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionB_;
- }
- /**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder() {
- return regionB_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionB_;
- }
-
- public static final int FORCIBLE_FIELD_NUMBER = 3;
- private boolean forcible_;
- /**
- * <code>optional bool forcible = 3 [default = false];</code>
- */
- public boolean hasForcible() {
- return ((bitField0_ & 0x00000004) == 0x00000004);
- }
- /**
- * <code>optional bool forcible = 3 [default = false];</code>
- */
- public boolean getForcible() {
- return forcible_;
- }
-
- public static final int MASTER_SYSTEM_TIME_FIELD_NUMBER = 4;
- private long masterSystemTime_;
- /**
- * <pre>
- * wall clock time from master
- * </pre>
- *
- * <code>optional uint64 master_system_time = 4;</code>
- */
- public boolean hasMasterSystemTime() {
- return ((bitField0_ & 0x00000008) == 0x00000008);
- }
- /**
- * <pre>
- * wall clock time from master
- * </pre>
- *
- * <code>optional uint64 master_system_time = 4;</code>
- */
- public long getMasterSystemTime() {
- return masterSystemTime_;
- }
-
- private byte memoizedIsInitialized = -1;
- public final boolean isInitialized() {
- byte isInitialized = memoizedIsInitialized;
- if (isInitialized == 1) return true;
- if (isInitialized == 0) return false;
-
- if (!hasRegionA()) {
- memoizedIsInitialized = 0;
- return false;
- }
- if (!hasRegionB()) {
- memoizedIsInitialized = 0;
- return false;
- }
- if (!getRegionA().isInitialized()) {
- memoizedIsInitialized = 0;
- return false;
- }
- if (!getRegionB().isInitialized()) {
- memoizedIsInitialized = 0;
- return false;
- }
- memoizedIsInitialized = 1;
- return true;
- }
-
- public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
- throws java.io.IOException {
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
- output.writeMessage(1, getRegionA());
- }
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
- output.writeMessage(2, getRegionB());
- }
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
- output.writeBool(3, forcible_);
- }
- if (((bitField0_ & 0x00000008) == 0x00000008)) {
- output.writeUInt64(4, masterSystemTime_);
- }
- unknownFields.writeTo(output);
- }
-
- public int getSerializedSize() {
- int size = memoizedSize;
- if (size != -1) return size;
-
- size = 0;
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
- .computeMessageSize(1, getRegionA());
- }
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
- .computeMessageSize(2, getRegionB());
- }
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
- .computeBoolSize(3, forcible_);
- }
- if (((bitField0_ & 0x00000008) == 0x00000008)) {
- size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
- .computeUInt64Size(4, masterSystemTime_);
- }
- size += unknownFields.getSerializedSize();
- memoizedSize = size;
- return size;
- }
-
- private static final long serialVersionUID = 0L;
- @java.lang.Override
- public boolean equals(final java.lang.Object obj) {
- if (obj == this) {
- return true;
- }
- if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest)) {
- return super.equals(obj);
- }
- org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest) obj;
-
- boolean result = true;
- result = result && (hasRegionA() == other.hasRegionA());
- if (hasRegionA()) {
- result = result && getRegionA()
- .equals(other.getRegionA());
- }
- result = result && (hasRegionB() == other.hasRegionB());
- if (hasRegionB()) {
- result = result && getRegionB()
- .equals(other.getRegionB());
- }
- result = result && (hasForcible() == other.hasForcible());
- if (hasForcible()) {
- result = result && (getForcible()
- == other.getForcible());
- }
- result = result && (hasMasterSystemTime() == other.hasMasterSystemTime());
- if (hasMasterSystemTime()) {
- result = result && (getMasterSystemTime()
- == other.getMasterSystemTime());
- }
- result = result && unknownFields.equals(other.unknownFields);
- return result;
- }
-
- @java.lang.Override
- public int hashCode() {
- if (memoizedHashCode != 0) {
- return memoizedHashCode;
- }
- int hash = 41;
- hash = (19 * hash) + getDescriptorForType().hashCode();
- if (hasRegionA()) {
- hash = (37 * hash) + REGION_A_FIELD_NUMBER;
- hash = (53 * hash) + getRegionA().hashCode();
- }
- if (hasRegionB()) {
- hash = (37 * hash) + REGION_B_FIELD_NUMBER;
- hash = (53 * hash) + getRegionB().hashCode();
- }
- if (hasForcible()) {
- hash = (37 * hash) + FORCIBLE_FIELD_NUMBER;
- hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean(
- getForcible());
- }
- if (hasMasterSystemTime()) {
- hash = (37 * hash) + MASTER_SYSTEM_TIME_FIELD_NUMBER;
- hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
- getMasterSystemTime());
- }
- hash = (29 * hash) + unknownFields.hashCode();
- memoizedHashCode = hash;
- return hash;
- }
-
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest parseFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest parseFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest parseFrom(byte[] data)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest parseFrom(
- byte[] data,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest parseFrom(java.io.InputStream input)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseWithIOException(PARSER, input);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest parseFrom(
- java.io.InputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseWithIOException(PARSER, input, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest parseDelimitedFrom(java.io.InputStream input)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseDelimitedWithIOException(PARSER, input);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest parseDelimitedFrom(
- java.io.InputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest parseFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseWithIOException(PARSER, input);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest parseFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseWithIOException(PARSER, input, extensionRegistry);
- }
-
- public Builder newBuilderForType() { return newBuilder(); }
- public static Builder newBuilder() {
- return DEFAULT_INSTANCE.toBuilder();
- }
- public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest prototype) {
- return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
- }
- public Builder toBuilder() {
- return this == DEFAULT_INSTANCE
- ? new Builder() : new Builder().mergeFrom(this);
- }
-
- @java.lang.Override
- protected Builder newBuilderForType(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
- Builder builder = new Builder(parent);
- return builder;
- }
- /**
- * <pre>
- **
- * Merges the specified regions.
- * <p>
- * This method currently closes the regions and then merges them
- * </pre>
- *
- * Protobuf type {@code hbase.pb.MergeRegionsRequest}
- */
- public static final class Builder extends
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
- // @@protoc_insertion_point(builder_implements:hbase.pb.MergeRegionsRequest)
- org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequestOrBuilder {
- public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
- getDescriptor() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_MergeRegionsRequest_descriptor;
- }
-
- protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
- internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_MergeRegionsRequest_fieldAccessorTable
- .ensureFieldAccessorsInitialized(
- org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest.Builder.class);
- }
-
- // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest.newBuilder()
- private Builder() {
- maybeForceBuilderInitialization();
- }
-
- private Builder(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
- super(parent);
- maybeForceBuilderInitialization();
- }
- private void maybeForceBuilderInitialization() {
- if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .alwaysUseFieldBuilders) {
- getRegionAFieldBuilder();
- getRegionBFieldBuilder();
- }
- }
- public Builder clear() {
- super.clear();
- if (regionABuilder_ == null) {
- regionA_ = null;
- } else {
- regionABuilder_.clear();
- }
- bitField0_ = (bitField0_ & ~0x00000001);
- if (regionBBuilder_ == null) {
- regionB_ = null;
- } else {
- regionBBuilder_.clear();
- }
- bitField0_ = (bitField0_ & ~0x00000002);
- forcible_ = false;
- bitField0_ = (bitField0_ & ~0x00000004);
- masterSystemTime_ = 0L;
- bitField0_ = (bitField0_ & ~0x00000008);
- return this;
- }
-
- public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
- getDescriptorForType() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_MergeRegionsRequest_descriptor;
- }
-
- public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest getDefaultInstanceForType() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest.getDefaultInstance();
- }
-
- public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest build() {
- org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest result = buildPartial();
- if (!result.isInitialized()) {
- throw newUninitializedMessageException(result);
- }
- return result;
- }
-
- public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest buildPartial() {
- org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest(this);
- int from_bitField0_ = bitField0_;
- int to_bitField0_ = 0;
- if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
- to_bitField0_ |= 0x00000001;
- }
- if (regionABuilder_ == null) {
- result.regionA_ = regionA_;
- } else {
- result.regionA_ = regionABuilder_.build();
- }
- if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
- to_bitField0_ |= 0x00000002;
- }
- if (regionBBuilder_ == null) {
- result.regionB_ = regionB_;
- } else {
- result.regionB_ = regionBBuilder_.build();
- }
- if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
- to_bitField0_ |= 0x00000004;
- }
- result.forcible_ = forcible_;
- if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
- to_bitField0_ |= 0x00000008;
- }
- result.masterSystemTime_ = masterSystemTime_;
- result.bitField0_ = to_bitField0_;
- onBuilt();
- return result;
- }
-
- public Builder clone() {
- return (Builder) super.clone();
- }
- public Builder setField(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
- Object value) {
- return (Builder) super.setField(field, value);
- }
- public Builder clearField(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
- return (Builder) super.clearField(field);
- }
- public Builder clearOneof(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
- return (Builder) super.clearOneof(oneof);
- }
- public Builder setRepeatedField(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
- int index, Object value) {
- return (Builder) super.setRepeatedField(field, index, value);
- }
- public Builder addRepeatedField(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
- Object value) {
- return (Builder) super.addRepeatedField(field, value);
- }
- public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
- if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest) {
- return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest)other);
- } else {
- super.mergeFrom(other);
- return this;
- }
- }
-
- public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest other) {
- if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest.getDefaultInstance()) return this;
- if (other.hasRegionA()) {
- mergeRegionA(other.getRegionA());
- }
- if (other.hasRegionB()) {
- mergeRegionB(other.getRegionB());
- }
- if (other.hasForcible()) {
- setForcible(other.getForcible());
- }
- if (other.hasMasterSystemTime()) {
- setMasterSystemTime(other.getMasterSystemTime());
- }
- this.mergeUnknownFields(other.unknownFields);
- onChanged();
- return this;
- }
-
- public final boolean isInitialized() {
- if (!hasRegionA()) {
- return false;
- }
- if (!hasRegionB()) {
- return false;
- }
- if (!getRegionA().isInitialized()) {
- return false;
- }
- if (!getRegionB().isInitialized()) {
- return false;
- }
- return true;
- }
-
- public Builder mergeFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest parsedMessage = null;
- try {
- parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
- parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest) e.getUnfinishedMessage();
- throw e.unwrapIOException();
- } finally {
- if (parsedMessage != null) {
- mergeFrom(parsedMessage);
- }
- }
- return this;
- }
- private int bitField0_;
-
- private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier regionA_ = null;
- private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionABuilder_;
- /**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
- */
- public boolean hasRegionA() {
- return ((bitField0_ & 0x00000001) == 0x00000001);
- }
- /**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA() {
- if (regionABuilder_ == null) {
- return regionA_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionA_;
- } else {
- return regionABuilder_.getMessage();
- }
- }
- /**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
- */
- public Builder setRegionA(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) {
- if (regionABuilder_ == null) {
- if (value == null) {
- throw new NullPointerException();
- }
- regionA_ = value;
- onChanged();
- } else {
- regionABuilder_.setMessage(value);
- }
- bitField0_ |= 0x00000001;
- return this;
- }
- /**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
- */
- public Builder setRegionA(
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) {
- if (regionABuilder_ == null) {
- regionA_ = builderForValue.build();
- onChanged();
- } else {
- regionABuilder_.setMessage(builderForValue.build());
- }
- bitField0_ |= 0x00000001;
- return this;
- }
- /**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
- */
- public Builder mergeRegionA(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) {
- if (regionABuilder_ == null) {
- if (((bitField0_ & 0x00000001) == 0x00000001) &&
- regionA_ != null &&
- regionA_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) {
- regionA_ =
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(regionA_).mergeFrom(value).buildPartial();
- } else {
- regionA_ = value;
- }
- onChanged();
- } else {
- regionABuilder_.mergeFrom(value);
- }
- bitField0_ |= 0x00000001;
- return this;
- }
- /**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
- */
- public Builder clearRegionA() {
- if (regionABuilder_ == null) {
- regionA_ = null;
- onChanged();
- } else {
- regionABuilder_.clear();
- }
- bitField0_ = (bitField0_ & ~0x00000001);
- return this;
- }
- /**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionABuilder() {
- bitField0_ |= 0x00000001;
- onChanged();
- return getRegionAFieldBuilder().getBuilder();
- }
- /**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder() {
- if (regionABuilder_ != null) {
- return regionABuilder_.getMessageOrBuilder();
- } else {
- return regionA_ == null ?
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionA_;
- }
- }
- /**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
- */
- private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>
- getRegionAFieldBuilder() {
- if (regionABuilder_ == null) {
- regionABuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>(
- getRegionA(),
- getParentForChildren(),
- isClean());
- regionA_ = null;
- }
- return regionABuilder_;
- }
-
- private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier regionB_ = null;
- private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBBuilder_;
- /**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
- */
- public boolean hasRegionB() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
- }
- /**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB() {
- if (regionBBuilder_ == null) {
- return regionB_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionB_;
- } else {
- return regionBBuilder_.getMessage();
- }
- }
- /**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
- */
- public Builder setRegionB(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) {
- if (regionBBuilder_ == null) {
- if (value == null) {
- throw new NullPointerException();
- }
- regionB_ = value;
- onChanged();
- } else {
- regionBBuilder_.setMessage(value);
- }
- bitField0_ |= 0x00000002;
- return this;
- }
- /**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
- */
- public Builder setRegionB(
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) {
- if (regionBBuilder_ == null) {
- regionB_ = builderForValue.build();
- onChanged();
- } else {
- regionBBuilder_.setMessage(builderForValue.build());
- }
- bitField0_ |= 0x00000002;
- return this;
- }
- /**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
- */
- public Builder mergeRegionB(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) {
- if (regionBBuilder_ == null) {
- if (((bitField0_ & 0x00000002) == 0x00000002) &&
- regionB_ != null &&
- regionB_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) {
- regionB_ =
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(regionB_).mergeFrom(value).buildPartial();
- } else {
- regionB_ = value;
- }
- onChanged();
- } else {
- regionBBuilder_.mergeFrom(value);
- }
- bitField0_ |= 0x00000002;
- return this;
- }
- /**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
- */
- public Builder clearRegionB() {
- if (regionBBuilder_ == null) {
- regionB_ = null;
- onChanged();
- } else {
- regionBBuilder_.clear();
- }
- bitField0_ = (bitField0_ & ~0x00000002);
- return this;
- }
- /**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionBBuilder() {
- bitField0_ |= 0x00000002;
- onChanged();
- return getRegionBFieldBuilder().getBuilder();
- }
- /**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder() {
- if (regionBBuilder_ != null) {
- return regionBBuilder_.getMessageOrBuilder();
- } else {
- return regionB_ == null ?
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionB_;
- }
- }
- /**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
- */
- private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>
- getRegionBFieldBuilder() {
- if (regionBBuilder_ == null) {
- regionBBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>(
- getRegionB(),
- getParentForChildren(),
- isClean());
- regionB_ = null;
- }
- return regionBBuilder_;
- }
-
- private boolean forcible_ ;
- /**
- * <code>optional bool forcible = 3 [default = false];</code>
- */
- public boolean hasForcible() {
- return ((bitField0_ & 0x00000004) == 0x00000004);
- }
- /**
- * <code>optional bool forcible = 3 [default = false];</code>
- */
- public boolean getForcible() {
- return forcible_;
- }
- /**
- * <code>optional bool forcible = 3 [default = false];</code>
- */
- public Builder setForcible(boolean value) {
- bitField0_ |= 0x00000004;
- forcible_ = value;
- onChanged();
- return this;
- }
- /**
- * <code>optional bool forcible = 3 [default = false];</code>
- */
- public Builder clearForcible() {
- bitField0_ = (bitField0_ & ~0x00000004);
- forcible_ = false;
- onChanged();
- return this;
- }
-
- private long masterSystemTime_ ;
- /**
- * <pre>
- * wall clock time from master
- * </pre>
- *
- * <code>optional uint64 master_system_time = 4;</code>
- */
- public boolean hasMasterSystemTime() {
- return ((bitField0_ & 0x00000008) == 0x00000008);
- }
- /**
- * <pre>
- * wall clock time from master
- * </pre>
- *
- * <code>optional uint64 master_system_time = 4;</code>
- */
- public long getMasterSystemTime() {
- return masterSystemTime_;
- }
- /**
- * <pre>
- * wall clock time from master
- * </pre>
- *
- * <code>optional uint64 master_system_time = 4;</code>
- */
- public Builder setMasterSystemTime(long value) {
- bitField0_ |= 0x00000008;
- masterSystemTime_ = value;
- onChanged();
- return this;
- }
- /**
- * <pre>
- * wall clock time from master
- * </pre>
- *
- * <code>optional uint64 master_system_time = 4;</code>
- */
- public Builder clearMasterSystemTime() {
- bitField0_ = (bitField0_ & ~0x00000008);
- masterSystemTime_ = 0L;
- onChanged();
- return this;
- }
- public final Builder setUnknownFields(
- final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
- return super.setUnknownFields(unknownFields);
- }
-
- public final Builder mergeUnknownFields(
- final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
- return super.mergeUnknownFields(unknownFields);
- }
-
-
- // @@protoc_insertion_point(builder_scope:hbase.pb.MergeRegionsRequest)
- }
-
- // @@protoc_insertion_point(class_scope:hbase.pb.MergeRegionsRequest)
- private static final org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest DEFAULT_INSTANCE;
- static {
- DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest();
- }
-
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest getDefaultInstance() {
- return DEFAULT_INSTANCE;
- }
-
- @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<MergeRegionsRequest>
- PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<MergeRegionsRequest>() {
- public MergeRegionsRequest parsePartialFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return new MergeRegionsRequest(input, extensionRegistry);
- }
- };
-
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<MergeRegionsRequest> parser() {
- return PARSER;
- }
-
- @java.lang.Override
- public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<MergeRegionsRequest> getParserForType() {
- return PARSER;
- }
-
- public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest getDefaultInstanceForType() {
- return DEFAULT_INSTANCE;
- }
-
- }
-
- public interface MergeRegionsResponseOrBuilder extends
- // @@protoc_insertion_point(interface_extends:hbase.pb.MergeRegionsResponse)
- org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
- }
- /**
- * Protobuf type {@code hbase.pb.MergeRegionsResponse}
- */
- public static final class MergeRegionsResponse extends
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
- // @@protoc_insertion_point(message_implements:hbase.pb.MergeRegionsResponse)
- MergeRegionsResponseOrBuilder {
- // Use MergeRegionsResponse.newBuilder() to construct.
- private MergeRegionsResponse(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
- super(builder);
- }
- private MergeRegionsResponse() {
- }
-
- @java.lang.Override
- public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
- return this.unknownFields;
- }
- private MergeRegionsResponse(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- this();
- org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
- try {
- boolean done = false;
- while (!done) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
- done = true;
- break;
- default: {
- if (!parseUnknownField(input, unknownFields,
- extensionRegistry, tag)) {
- done = true;
- }
- break;
- }
- }
- }
- } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
- throw e.setUnfinishedMessage(this);
- } catch (java.io.IOException e) {
- throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
- e).setUnfinishedMessage(this);
- } finally {
- this.unknownFields = unknownFields.build();
- makeExtensionsImmutable();
- }
- }
- public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
- getDescriptor() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_MergeRegionsResponse_descriptor;
- }
-
- protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
- internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_MergeRegionsResponse_fieldAccessorTable
- .ensureFieldAccessorsInitialized(
- org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.Builder.class);
- }
-
- private byte memoizedIsInitialized = -1;
- public final boolean isInitialized() {
- byte isInitialized = memoizedIsInitialized;
- if (isInitialized == 1) return true;
- if (isInitialized == 0) return false;
-
- memoizedIsInitialized = 1;
- return true;
- }
-
- public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
- throws java.io.IOException {
- unknownFields.writeTo(output);
- }
-
- public int getSerializedSize() {
- int size = memoizedSize;
- if (size != -1) return size;
-
- size = 0;
- size += unknownFields.getSerializedSize();
- memoizedSize = size;
- return size;
- }
-
- private static final long serialVersionUID = 0L;
- @java.lang.Override
- public boolean equals(final java.lang.Object obj) {
- if (obj == this) {
- return true;
- }
- if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse)) {
- return super.equals(obj);
- }
- org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse) obj;
-
- boolean result = true;
- result = result && unknownFields.equals(other.unknownFields);
- return result;
- }
-
- @java.lang.Override
- public int hashCode() {
- if (memoizedHashCode != 0) {
- return memoizedHashCode;
- }
- int hash = 41;
- hash = (19 * hash) + getDescriptorForType().hashCode();
- hash = (29 * hash) + unknownFields.hashCode();
- memoizedHashCode = hash;
- return hash;
- }
-
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse parseFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse parseFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse parseFrom(byte[] data)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse parseFrom(
- byte[] data,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse parseFrom(java.io.InputStream input)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseWithIOException(PARSER, input);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse parseFrom(
- java.io.InputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseWithIOException(PARSER, input, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse parseDelimitedFrom(java.io.InputStream input)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseDelimitedWithIOException(PARSER, input);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse parseDelimitedFrom(
- java.io.InputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse parseFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseWithIOException(PARSER, input);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse parseFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseWithIOException(PARSER, input, extensionRegistry);
- }
-
- public Builder newBuilderForType() { return newBuilder(); }
- public static Builder newBuilder() {
- return DEFAULT_INSTANCE.toBuilder();
- }
- public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse prototype) {
- return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
- }
- public Builder toBuilder() {
- return this == DEFAULT_INSTANCE
- ? new Builder() : new Builder().mergeFrom(this);
- }
-
- @java.lang.Override
- protected Builder newBuilderForType(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
- Builder builder = new Builder(parent);
- return builder;
- }
- /**
- * Protobuf type {@code hbase.pb.MergeRegionsResponse}
- */
- public static final class Builder extends
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
- // @@protoc_insertion_point(builder_implements:hbase.pb.MergeRegionsResponse)
- org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponseOrBuilder {
- public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
- getDescriptor() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_MergeRegionsResponse_descriptor;
- }
-
- protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
- internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_MergeRegionsResponse_fieldAccessorTable
- .ensureFieldAccessorsInitialized(
- org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.Builder.class);
- }
-
- // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.newBuilder()
- private Builder() {
- maybeForceBuilderInitialization();
- }
-
- private Builder(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
- super(parent);
- maybeForceBuilderInitialization();
- }
- private void maybeForceBuilderInitialization() {
- if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .alwaysUseFieldBuilders) {
- }
- }
- public Builder clear() {
- super.clear();
- return this;
- }
-
- public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
- getDescriptorForType() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_MergeRegionsResponse_descriptor;
- }
-
- public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse getDefaultInstanceForType() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.getDefaultInstance();
- }
-
- public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse build() {
- org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse result = buildPartial();
- if (!result.isInitialized()) {
- throw newUninitializedMessageException(result);
- }
- return result;
- }
-
- public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse buildPartial() {
- org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse(this);
- onBuilt();
- return result;
- }
-
- public Builder clone() {
- return (Builder) super.clone();
- }
- public Builder setField(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
- Object value) {
- return (Builder) super.setField(field, value);
- }
- public Builder clearField(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
- return (Builder) super.clearField(field);
- }
- public Builder clearOneof(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
- return (Builder) super.clearOneof(oneof);
- }
- public Builder setRepeatedField(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
- int index, Object value) {
- return (Builder) super.setRepeatedField(field, index, value);
- }
- public Builder addRepeatedField(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
- Object value) {
- return (Builder) super.addRepeatedField(field, value);
- }
- public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
- if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse) {
- return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse)other);
- } else {
- super.mergeFrom(other);
- return this;
- }
- }
-
- public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse other) {
- if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.getDefaultInstance()) return this;
- this.mergeUnknownFields(other.unknownFields);
- onChanged();
- return this;
- }
-
- public final boolean isInitialized() {
- return true;
- }
-
- public Builder mergeFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse parsedMessage = null;
- try {
- parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
- parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse) e.getUnfinishedMessage();
- throw e.unwrapIOException();
- } finally {
- if (parsedMessage != null) {
- mergeFrom(parsedMessage);
- }
- }
- return this;
- }
- public final Builder setUnknownFields(
- final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
- return super.setUnknownFields(unknownFields);
- }
-
- public final Builder mergeUnknownFields(
- final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
- return super.mergeUnknownFields(unknownFields);
- }
-
-
- // @@protoc_insertion_point(builder_scope:hbase.pb.MergeRegionsResponse)
- }
-
- // @@protoc_insertion_point(class_scope:hbase.pb.MergeRegionsResponse)
- private static final org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse DEFAULT_INSTANCE;
- static {
- DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse();
- }
-
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse getDefaultInstance() {
- return DEFAULT_INSTANCE;
- }
-
- @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<MergeRegionsResponse>
- PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<MergeRegionsResponse>() {
- public MergeRegionsResponse parsePartialFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return new MergeRegionsResponse(input, extensionRegistry);
- }
- };
-
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<MergeRegionsResponse> parser() {
- return PARSER;
- }
-
- @java.lang.Override
- public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<MergeRegionsResponse> getParserForType() {
- return PARSER;
- }
-
- public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse getDefaultInstanceForType() {
- return DEFAULT_INSTANCE;
- }
-
- }
-
public interface WALEntryOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.WALEntry)
org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
@@ -26239,14 +24839,6 @@ public final class AdminProtos {
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionResponse> done);
/**
- * <code>rpc MergeRegions(.hbase.pb.MergeRegionsRequest) returns (.hbase.pb.MergeRegionsResponse);</code>
- */
- public abstract void mergeRegions(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
- org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest request,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse> done);
-
- /**
* <code>rpc ReplicateWALEntry(.hbase.pb.ReplicateWALEntryRequest) returns (.hbase.pb.ReplicateWALEntryResponse);</code>
*/
public abstract void replicateWALEntry(
@@ -26396,14 +24988,6 @@ public final class AdminProtos {
}
@java.lang.Override
- public void mergeRegions(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
- org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest request,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse> done) {
- impl.mergeRegions(controller, request, done);
- }
-
- @java.lang.Override
public void replicateWALEntry(
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryRequest request,
@@ -26510,22 +25094,20 @@ public final class AdminProtos {
case 9:
return impl.compactRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest)request);
case 10:
- return impl.mergeRegions(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest)request);
- case 11:
return impl.replicateWALEntry(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryRequest)request);
- case 12:
+ case 11:
return impl.replay(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryRequest)request);
- case 13:
+ case 12:
return impl.rollWALWriter(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterRequest)request);
- case 14:
+ case 13:
return impl.getServerInfo(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetServerInfoRequest)request);
- case 15:
+ case 14:
return impl.stopServer(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerRequest)request);
- case 16:
+ case 15:
return impl.updateFavoredNodes(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateFavoredNodesRequest)request);
- case 17:
+ case 16:
return impl.updateConfiguration(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationRequest)request);
- case 18:
+ case 17:
return impl.getRegionLoad(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadRequest)request);
default:
throw new java.lang.AssertionError("Can't get here.");
@@ -26562,22 +25144,20 @@ public final class AdminProtos {
case 9:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest.getDefaultInstance();
case 10:
- return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest.getDefaultInstance();
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryRequest.getDefaultInstance();
case 11:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryRequest.getDefaultInstance();
case 12:
- return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryRequest.getDefaultInstance();
- case 13:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterRequest.getDefaultInstance();
- case 14:
+ case 13:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetServerInfoRequest.getDefaultInstance();
- case 15:
+ case 14:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerRequest.getDefaultInstance();
- case 16:
+ case 15:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateFavoredNodesRequest.getDefaultInstance();
- case 17:
+ case 16:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationRequest.getDefaultInstance();
- case 18:
+ case 17:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadRequest.getDefaultInstance();
default:
throw new java.lang.AssertionError("Can't get here.");
@@ -26614,22 +25194,20 @@ public final class AdminProtos {
case 9:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionResponse.getDefaultInstance();
case 10:
- return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.getDefaultInstance();
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryResponse.getDefaultInstance();
case 11:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryResponse.getDefaultInstance();
case 12:
- return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryResponse.getDefaultInstance();
- case 13:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterResponse.getDefaultInstance();
- case 14:
+ case 13:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetServerInfoResponse.getDefaultInstance();
- case 15:
+ case 14:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerResponse.getDefaultInstance();
- case 16:
+ case 15:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateFavoredNodesResponse.getDefaultInstance();
- case 17:
+ case 16:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationResponse.getDefaultInstance();
- case 18:
+ case 17:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadResponse.getDefaultInstance();
default:
throw new java.lang.AssertionError("Can't get here.");
@@ -26720,14 +25298,6 @@ public final class AdminProtos {
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionResponse> done);
/**
- * <code>rpc MergeRegions(.hbase.pb.MergeRegionsRequest) returns (.hbase.pb.MergeRegionsResponse);</code>
- */
- public abstract void mergeRegions(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
- org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest request,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse> done);
-
- /**
* <code>rpc ReplicateWALEntry(.hbase.pb.ReplicateWALEntryRequest) returns (.hbase.pb.ReplicateWALEntryResponse);</code>
*/
public abstract void replicateWALEntry(
@@ -26864,46 +25434,41 @@ public final class AdminProtos {
done));
return;
case 10:
- this.mergeRegions(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest)request,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse>specializeCallback(
- done));
- return;
- case 11:
this.replicateWALEntry(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryRequest)request,
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryResponse>specializeCallback(
done));
return;
- case 12:
+ case 11:
this.replay(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryRequest)request,
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryResponse>specializeCallback(
done));
return;
- case 13:
+ case 12:
this.rollWALWriter(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterRequest)request,
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterResponse>specializeCallback(
done));
return;
- case 14:
+ case 13:
this.getServerInfo(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetServerInfoRequest)request,
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetServerInfoResponse>specializeCallback(
done));
return;
- case 15:
+ case 14:
this.stopServer(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerRequest)request,
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerResponse>specializeCallback(
done));
return;
- case 16:
+ case 15:
this.updateFavoredNodes(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateFavoredNodesRequest)request,
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateFavoredNodesResponse>specializeCallback(
done));
return;
- case 17:
+ case 16:
this.updateConfiguration(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationRequest)request,
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationResponse>specializeCallback(
done));
return;
- case 18:
+ case 17:
this.getRegionLoad(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadRequest)request,
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadResponse>specializeCallback(
done));
@@ -26943,22 +25508,20 @@ public final class AdminProtos {
case 9:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest.getDefaultInstance();
case 10:
- return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest.getDefaultInstance();
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryRequest.getDefaultInstance();
case 11:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryRequest.getDefaultInstance();
case 12:
- return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryRequest.getDefaultInstance();
- case 13:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterRequest.getDefaultInstance();
- case 14:
+ case 13:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetServerInfoRequest.getDefaultInstance();
- case 15:
+ case 14:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerRequest.getDefaultInstance();
- case 16:
+ case 15:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateFavoredNodesRequest.getDefaultInstance();
- case 17:
+ case 16:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationRequest.getDefaultInstance();
- case 18:
+ case 17:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadRequest.getDefaultInstance();
default:
throw new java.lang.AssertionError("Can't get here.");
@@ -26995,22 +25558,20 @@ public final class AdminProtos {
case 9:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionResponse.getDefaultInstance();
case 10:
- return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.getDefaultInstance();
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryResponse.getDefaultInstance();
case 11:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryResponse.getDefaultInstance();
case 12:
- return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryResponse.getDefaultInstance();
- case 13:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterResponse.getDefaultInstance();
- case 14:
+ case 13:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetServerInfoResponse.getDefaultInstance();
- case 15:
+ case 14:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerResponse.getDefaultInstance();
- case 16:
+ case 15:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateFavoredNodesResponse.getDefaultInstance();
- case 17:
+ case 16:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationResponse.getDefaultInstance();
- case 18:
+ case 17:
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadResponse.getDefaultInstance();
default:
throw new java.lang.AssertionError("Can't get here.");
@@ -27183,27 +25744,12 @@ public final class AdminProtos {
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionResponse.getDefaultInstance()));
}
- public void mergeRegions(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
- org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest request,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse> done) {
- channel.callMethod(
- getDescriptor().getMethods().get(10),
- controller,
- request,
- org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.getDefaultInstance(),
- org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback(
- done,
- org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.class,
- org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.getDefaultInstance()));
- }
-
public void replicateWALEntry(
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryRequest request,
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryResponse> done) {
channel.callMethod(
- getDescriptor().getMethods().get(11),
+ getDescriptor().getMethods().get(10),
controller,
request,
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryResponse.getDefaultInstance(),
@@ -27218,7 +25764,7 @@ public final class AdminProtos {
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryRequest request,
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryResponse> done) {
channel.callMethod(
- getDescriptor().getMethods().get(12),
+ getDescriptor().getMethods().get(11),
controller,
request,
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryResponse.getDefaultInstance(),
@@ -27233,7 +25779,7 @@ public final class AdminProtos {
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterRequest request,
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterResponse> done) {
channel.callMethod(
- getDescriptor().getMethods().get(13),
+ getDescriptor().getMethods().get(12),
controller,
request,
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterResponse.getDefaultInstance(),
@@ -27248,7 +25794,7 @@ public final class AdminProtos {
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetServerInfoRequest request,
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetServerInfoResponse> done) {
channel.callMethod(
- getDescriptor().getMethods().get(14),
+ getDescriptor().getMethods().get(13),
controller,
request,
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetServerInfoResponse.getDefaultInstance(),
@@ -27263,7 +25809,7 @@ public final class AdminProtos {
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerRequest request,
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerResponse> done) {
channel.callMethod(
- getDescriptor().getMethods().get(15),
+ getDescriptor().getMethods().get(14),
controller,
request,
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerResponse.getDefaultInstance(),
@@ -27278,7 +25824,7 @@ public final class AdminProtos {
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateFavoredNodesRequest request,
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateFavoredNodesResponse> done) {
channel.callMethod(
- getDescriptor().getMethods().get(16),
+ getDescriptor().getMethods().get(15),
controller,
request,
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateFavoredNodesResponse.getDefaultInstance(),
@@ -27293,7 +25839,7 @@ public final class AdminProtos {
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationRequest request,
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationResponse> done) {
channel.callMethod(
- getDescriptor().getMethods().get(17),
+ getDescriptor().getMethods().get(16),
controller,
request,
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationResponse.getDefaultInstance(),
@@ -27308,7 +25854,7 @@ public final class AdminProtos {
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadRequest request,
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadResponse> done) {
channel.callMethod(
- getDescriptor().getMethods().get(18),
+ getDescriptor().getMethods().get(17),
controller,
request,
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadResponse.getDefaultInstance(),
@@ -27375,11 +25921,6 @@ public final class AdminProtos {
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest request)
throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
- public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse mergeRegions(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
- org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest request)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
-
public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryResponse replicateWALEntry(
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryRequest request)
@@ -27548,24 +26089,12 @@ public final class AdminProtos {
}
- public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse mergeRegions(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
- org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest request)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException {
- return (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse) channel.callBlockingMethod(
- getDescriptor().getMethods().get(10),
- controller,
- request,
- org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.getDefaultInstance());
- }
-
-
public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryResponse replicateWALEntry(
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryRequest request)
throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException {
return (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryResponse) channel.callBlockingMethod(
- getDescriptor().getMethods().get(11),
+ getDescriptor().getMethods().get(10),
controller,
request,
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryResponse.getDefaultInstance());
@@ -27577,7 +26106,7 @@ public final class AdminProtos {
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryRequest request)
throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException {
return (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryResponse) channel.callBlockingMethod(
- getDescriptor().getMethods().get(12),
+ getDescriptor().getMethods().get(11),
controller,
request,
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryResponse.getDefaultInstance());
@@ -27589,7 +26118,7 @@ public final class AdminProtos {
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterRequest request)
throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException {
return (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterResponse) channel.callBlockingMethod(
- getDescriptor().getMethods().get(13),
+ getDescriptor().getMethods().get(12),
controller,
request,
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterResponse.getDefaultInstance());
@@ -27601,7 +26130,7 @@ public final class AdminProtos {
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetServerInfoRequest request)
throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException {
return (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetServerInfoResponse) channel.callBlockingMethod(
- getDescriptor().getMethods().get(14),
+ getDescriptor().getMethods().get(13),
controller,
request,
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetServerInfoResponse.getDefaultInstance());
@@ -27613,7 +26142,7 @@ public final class AdminProtos {
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerRequest request)
throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException {
return (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerResponse) channel.callBlockingMethod(
- getDescriptor().getMethods().get(15),
+ getDescriptor().getMethods().get(14),
controller,
request,
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerResponse.getDefaultInstance());
@@ -27625,7 +26154,7 @@ public final class AdminProtos {
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateFavoredNodesRequest request)
<TRUNCATED>
[2/7] hbase git commit: HBASE-17470 Remove merge region code from
region server (Stephen Yuan Jiang)
Posted by sy...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionMergeTransactionImpl.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionMergeTransactionImpl.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionMergeTransactionImpl.java
deleted file mode 100644
index 5c86429..0000000
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionMergeTransactionImpl.java
+++ /dev/null
@@ -1,742 +0,0 @@
-/**
- * Copyright The Apache Software Foundation
- *
- * Licensed to the Apache Software Foundation (ASF) under one or more
- * contributor license agreements. See the NOTICE file distributed with this
- * work for additional information regarding copyright ownership. The ASF
- * licenses this file to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance with the License.
- * You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
- * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
- * License for the specific language governing permissions and limitationsME
- * under the License.
- */
-package org.apache.hadoop.hbase.regionserver;
-
-import java.io.IOException;
-import java.io.InterruptedIOException;
-import java.security.PrivilegedExceptionAction;
-import java.util.ArrayList;
-import java.util.List;
-import java.util.ListIterator;
-import java.util.Map;
-
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
-import org.apache.hadoop.hbase.classification.InterfaceAudience;
-import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.hbase.HConstants;
-import org.apache.hadoop.hbase.HRegionInfo;
-import org.apache.hadoop.hbase.MetaMutationAnnotation;
-import org.apache.hadoop.hbase.MetaTableAccessor;
-import org.apache.hadoop.hbase.Server;
-import org.apache.hadoop.hbase.ServerName;
-import org.apache.hadoop.hbase.client.Delete;
-import org.apache.hadoop.hbase.client.Mutation;
-import org.apache.hadoop.hbase.client.Put;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionStateTransition.TransitionCode;
-import org.apache.hadoop.hbase.regionserver.SplitTransactionImpl.LoggingProgressable;
-import org.apache.hadoop.hbase.security.User;
-import org.apache.hadoop.hbase.util.Bytes;
-import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-import org.apache.hadoop.hbase.util.Pair;
-
-import com.google.common.annotations.VisibleForTesting;
-
-@InterfaceAudience.Private
-public class RegionMergeTransactionImpl implements RegionMergeTransaction {
- private static final Log LOG = LogFactory.getLog(RegionMergeTransactionImpl.class);
-
- // Merged region info
- private HRegionInfo mergedRegionInfo;
- // region_a sorts before region_b
- private final HRegion region_a;
- private final HRegion region_b;
- // merges dir is under region_a
- private final Path mergesdir;
- // We only merge adjacent regions if forcible is false
- private final boolean forcible;
- private final long masterSystemTime;
-
- /*
- * Transaction state for listener, only valid during execute and
- * rollback
- */
- private RegionMergeTransactionPhase currentPhase = RegionMergeTransactionPhase.STARTED;
- private Server server;
- private RegionServerServices rsServices;
-
- public static class JournalEntryImpl implements JournalEntry {
- private RegionMergeTransactionPhase type;
- private long timestamp;
-
- public JournalEntryImpl(RegionMergeTransactionPhase type) {
- this(type, EnvironmentEdgeManager.currentTime());
- }
-
- public JournalEntryImpl(RegionMergeTransactionPhase type, long timestamp) {
- this.type = type;
- this.timestamp = timestamp;
- }
-
- @Override
- public String toString() {
- StringBuilder sb = new StringBuilder();
- sb.append(type);
- sb.append(" at ");
- sb.append(timestamp);
- return sb.toString();
- }
-
- @Override
- public RegionMergeTransactionPhase getPhase() {
- return type;
- }
-
- @Override
- public long getTimeStamp() {
- return timestamp;
- }
- }
-
- /*
- * Journal of how far the merge transaction has progressed.
- */
- private final List<JournalEntry> journal = new ArrayList<JournalEntry>();
-
- /**
- * Listeners
- */
- private final ArrayList<TransactionListener> listeners = new ArrayList<TransactionListener>();
-
- private static IOException closedByOtherException = new IOException(
- "Failed to close region: already closed by another thread");
-
- private RegionServerCoprocessorHost rsCoprocessorHost = null;
-
- /**
- * Constructor
- * @param a region a to merge
- * @param b region b to merge
- * @param forcible if false, we will only merge adjacent regions
- */
- public RegionMergeTransactionImpl(final Region a, final Region b,
- final boolean forcible) {
- this(a, b, forcible, EnvironmentEdgeManager.currentTime());
- }
- /**
- * Constructor
- * @param a region a to merge
- * @param b region b to merge
- * @param forcible if false, we will only merge adjacent regions
- * @param masterSystemTime the time at the master side
- */
- public RegionMergeTransactionImpl(final Region a, final Region b,
- final boolean forcible, long masterSystemTime) {
- if (a.getRegionInfo().compareTo(b.getRegionInfo()) <= 0) {
- this.region_a = (HRegion)a;
- this.region_b = (HRegion)b;
- } else {
- this.region_a = (HRegion)b;
- this.region_b = (HRegion)a;
- }
- this.forcible = forcible;
- this.masterSystemTime = masterSystemTime;
- this.mergesdir = region_a.getRegionFileSystem().getMergesDir();
- }
-
- private void transition(RegionMergeTransactionPhase nextPhase) throws IOException {
- transition(nextPhase, false);
- }
-
- private void transition(RegionMergeTransactionPhase nextPhase, boolean isRollback)
- throws IOException {
- if (!isRollback) {
- // Add to the journal first, because if the listener throws an exception
- // we need to roll back starting at 'nextPhase'
- this.journal.add(new JournalEntryImpl(nextPhase));
- }
- for (int i = 0; i < listeners.size(); i++) {
- TransactionListener listener = listeners.get(i);
- if (!isRollback) {
- listener.transition(this, currentPhase, nextPhase);
- } else {
- listener.rollback(this, currentPhase, nextPhase);
- }
- }
- currentPhase = nextPhase;
- }
-
- @Override
- public boolean prepare(final RegionServerServices services) throws IOException {
- if (!region_a.getTableDesc().getTableName()
- .equals(region_b.getTableDesc().getTableName())) {
- LOG.info("Can't merge regions " + region_a + "," + region_b
- + " because they do not belong to the same table");
- return false;
- }
- if (region_a.getRegionInfo().equals(region_b.getRegionInfo())) {
- LOG.info("Can't merge the same region " + region_a);
- return false;
- }
- if (!forcible && !HRegionInfo.areAdjacent(region_a.getRegionInfo(),
- region_b.getRegionInfo())) {
- String msg = "Skip merging " + region_a.getRegionInfo().getRegionNameAsString()
- + " and " + region_b.getRegionInfo().getRegionNameAsString()
- + ", because they are not adjacent.";
- LOG.info(msg);
- return false;
- }
- if (!this.region_a.isMergeable() || !this.region_b.isMergeable()) {
- return false;
- }
- try {
- boolean regionAHasMergeQualifier = hasMergeQualifierInMeta(services,
- region_a.getRegionInfo().getRegionName());
- if (regionAHasMergeQualifier ||
- hasMergeQualifierInMeta(services, region_b.getRegionInfo().getRegionName())) {
- LOG.debug("Region " + (regionAHasMergeQualifier ?
- region_a.getRegionInfo().getRegionNameAsString()
- : region_b.getRegionInfo().getRegionNameAsString())
- + " is not mergeable because it has merge qualifier in META");
- return false;
- }
- } catch (IOException e) {
- LOG.warn("Failed judging whether merge transaction is available for "
- + region_a.getRegionInfo().getRegionNameAsString() + " and "
- + region_b.getRegionInfo().getRegionNameAsString(), e);
- return false;
- }
-
- // WARN: make sure there is no parent region of the two merging regions in
- // hbase:meta If exists, fixing up daughters would cause daughter regions(we
- // have merged one) online again when we restart master, so we should clear
- // the parent region to prevent the above case
- // Since HBASE-7721, we don't need fix up daughters any more. so here do
- // nothing
-
- this.mergedRegionInfo = getMergedRegionInfo(region_a.getRegionInfo(),
- region_b.getRegionInfo());
-
- transition(RegionMergeTransactionPhase.PREPARED);
- return true;
- }
-
- @Override
- public Region execute(final Server server, final RegionServerServices services)
- throws IOException {
- if (User.isHBaseSecurityEnabled(region_a.getBaseConf())) {
- LOG.warn("Should use execute(Server, RegionServerServices, User)");
- }
- return execute(server, services, null);
- }
-
- @Override
- public Region execute(final Server server, final RegionServerServices services, User user)
- throws IOException {
- this.server = server;
- this.rsServices = services;
- if (rsCoprocessorHost == null) {
- rsCoprocessorHost = server != null ?
- ((HRegionServer) server).getRegionServerCoprocessorHost() : null;
- }
- final HRegion mergedRegion = createMergedRegion(server, services, user);
- if (rsCoprocessorHost != null) {
- rsCoprocessorHost.postMergeCommit(this.region_a, this.region_b, mergedRegion, user);
- }
- stepsAfterPONR(server, services, mergedRegion, user);
-
- transition(RegionMergeTransactionPhase.COMPLETED);
-
- return mergedRegion;
- }
-
- @VisibleForTesting
- public void stepsAfterPONR(final Server server, final RegionServerServices services,
- final HRegion mergedRegion, User user) throws IOException {
- openMergedRegion(server, services, mergedRegion);
- if (rsCoprocessorHost != null) {
- rsCoprocessorHost.postMerge(region_a, region_b, mergedRegion, user);
- }
- }
-
- /**
- * Prepare the merged region and region files.
- * @param server Hosting server instance. Can be null when testing
- * @param services Used to online/offline regions.
- * @return merged region
- * @throws IOException If thrown, transaction failed. Call
- * {@link #rollback(Server, RegionServerServices)}
- */
- private HRegion createMergedRegion(final Server server, final RegionServerServices services,
- User user) throws IOException {
- LOG.info("Starting merge of " + region_a + " and "
- + region_b.getRegionInfo().getRegionNameAsString() + ", forcible=" + forcible);
- if ((server != null && server.isStopped())
- || (services != null && services.isStopping())) {
- throw new IOException("Server is stopped or stopping");
- }
-
- if (rsCoprocessorHost != null) {
- boolean ret = rsCoprocessorHost.preMerge(region_a, region_b, user);
- if (ret) {
- throw new IOException("Coprocessor bypassing regions " + this.region_a + " "
- + this.region_b + " merge.");
- }
- }
-
- // If true, no cluster to write meta edits to or to use coordination.
- boolean testing = server == null ? true : server.getConfiguration()
- .getBoolean("hbase.testing.nocluster", false);
-
- HRegion mergedRegion = stepsBeforePONR(server, services, testing);
-
- @MetaMutationAnnotation
- final List<Mutation> metaEntries = new ArrayList<Mutation>();
- if (rsCoprocessorHost != null) {
- boolean ret = rsCoprocessorHost.preMergeCommit(region_a, region_b, metaEntries, user);
-
- if (ret) {
- throw new IOException("Coprocessor bypassing regions " + this.region_a + " "
- + this.region_b + " merge.");
- }
- try {
- for (Mutation p : metaEntries) {
- HRegionInfo.parseRegionName(p.getRow());
- }
- } catch (IOException e) {
- LOG.error("Row key of mutation from coprocessor is not parsable as region name."
- + "Mutations from coprocessor should only be for hbase:meta table.", e);
- throw e;
- }
- }
-
- // This is the point of no return. Similar with SplitTransaction.
- // IF we reach the PONR then subsequent failures need to crash out this
- // regionserver
- transition(RegionMergeTransactionPhase.PONR);
-
- // Add merged region and delete region_a and region_b
- // as an atomic update. See HBASE-7721. This update to hbase:meta makes the region
- // will determine whether the region is merged or not in case of failures.
- // If it is successful, master will roll-forward, if not, master will
- // rollback
- if (services != null && !services.reportRegionStateTransition(TransitionCode.MERGE_PONR,
- mergedRegionInfo, region_a.getRegionInfo(), region_b.getRegionInfo())) {
- // Passed PONR, let SSH clean it up
- throw new IOException("Failed to notify master that merge passed PONR: "
- + region_a.getRegionInfo().getRegionNameAsString() + " and "
- + region_b.getRegionInfo().getRegionNameAsString());
- }
- return mergedRegion;
- }
-
- @VisibleForTesting
- public void prepareMutationsForMerge(HRegionInfo mergedRegion, HRegionInfo regionA,
- HRegionInfo regionB, ServerName serverName, List<Mutation> mutations) throws IOException {
- HRegionInfo copyOfMerged = new HRegionInfo(mergedRegion);
-
- // use the maximum of what master passed us vs local time.
- long time = Math.max(EnvironmentEdgeManager.currentTime(), masterSystemTime);
-
- // Put for parent
- Put putOfMerged = MetaTableAccessor.makePutFromRegionInfo(copyOfMerged, time);
- putOfMerged.addColumn(HConstants.CATALOG_FAMILY, HConstants.MERGEA_QUALIFIER,
- regionA.toByteArray());
- putOfMerged.addColumn(HConstants.CATALOG_FAMILY, HConstants.MERGEB_QUALIFIER,
- regionB.toByteArray());
- mutations.add(putOfMerged);
- // Deletes for merging regions
- Delete deleteA = MetaTableAccessor.makeDeleteFromRegionInfo(regionA, time);
- Delete deleteB = MetaTableAccessor.makeDeleteFromRegionInfo(regionB, time);
- mutations.add(deleteA);
- mutations.add(deleteB);
- // The merged is a new region, openSeqNum = 1 is fine.
- addLocation(putOfMerged, serverName, 1);
- }
-
- @VisibleForTesting
- Put addLocation(final Put p, final ServerName sn, long openSeqNum) {
- p.addColumn(HConstants.CATALOG_FAMILY, HConstants.SERVER_QUALIFIER, Bytes
- .toBytes(sn.getHostAndPort()));
- p.addColumn(HConstants.CATALOG_FAMILY, HConstants.STARTCODE_QUALIFIER, Bytes.toBytes(sn
- .getStartcode()));
- p.addColumn(HConstants.CATALOG_FAMILY, HConstants.SEQNUM_QUALIFIER, Bytes.toBytes(openSeqNum));
- return p;
- }
-
- @VisibleForTesting
- public HRegion stepsBeforePONR(final Server server, final RegionServerServices services,
- boolean testing) throws IOException {
- if (services != null && !services.reportRegionStateTransition(TransitionCode.READY_TO_MERGE,
- mergedRegionInfo, region_a.getRegionInfo(), region_b.getRegionInfo())) {
- throw new IOException("Failed to get ok from master to merge "
- + region_a.getRegionInfo().getRegionNameAsString() + " and "
- + region_b.getRegionInfo().getRegionNameAsString());
- }
-
- transition(RegionMergeTransactionPhase.SET_MERGING);
-
- this.region_a.getRegionFileSystem().createMergesDir();
-
- transition(RegionMergeTransactionPhase.CREATED_MERGE_DIR);
-
- Map<byte[], List<StoreFile>> hstoreFilesOfRegionA = closeAndOfflineRegion(
- services, this.region_a, true, testing);
- Map<byte[], List<StoreFile>> hstoreFilesOfRegionB = closeAndOfflineRegion(
- services, this.region_b, false, testing);
-
- assert hstoreFilesOfRegionA != null && hstoreFilesOfRegionB != null;
-
- // mergeStoreFiles creates merged region dirs under the region_a merges dir
- // Nothing to unroll here if failure -- clean up of CREATE_MERGE_DIR will
- // clean this up.
- mergeStoreFiles(hstoreFilesOfRegionA, hstoreFilesOfRegionB);
-
- // Log to the journal that we are creating merged region. We could fail
- // halfway through. If we do, we could have left
- // stuff in fs that needs cleanup -- a storefile or two. Thats why we
- // add entry to journal BEFORE rather than AFTER the change.
-
- transition(RegionMergeTransactionPhase.STARTED_MERGED_REGION_CREATION);
-
- HRegion mergedRegion = createMergedRegionFromMerges(this.region_a,
- this.region_b, this.mergedRegionInfo);
- return mergedRegion;
- }
-
- /**
- * Create a merged region from the merges directory under region a. In order
- * to mock it for tests, place it with a new method.
- * @param a hri of region a
- * @param b hri of region b
- * @param mergedRegion hri of merged region
- * @return merged HRegion.
- * @throws IOException
- */
- @VisibleForTesting
- HRegion createMergedRegionFromMerges(final HRegion a, final HRegion b,
- final HRegionInfo mergedRegion) throws IOException {
- return a.createMergedRegionFromMerges(mergedRegion, b);
- }
-
- /**
- * Close the merging region and offline it in regionserver
- * @param services
- * @param region
- * @param isRegionA true if it is merging region a, false if it is region b
- * @param testing true if it is testing
- * @return a map of family name to list of store files
- * @throws IOException
- */
- private Map<byte[], List<StoreFile>> closeAndOfflineRegion(
- final RegionServerServices services, final HRegion region,
- final boolean isRegionA, final boolean testing) throws IOException {
- Map<byte[], List<StoreFile>> hstoreFilesToMerge = null;
- Exception exceptionToThrow = null;
- try {
- hstoreFilesToMerge = region.close(false);
- } catch (Exception e) {
- exceptionToThrow = e;
- }
- if (exceptionToThrow == null && hstoreFilesToMerge == null) {
- // The region was closed by a concurrent thread. We can't continue
- // with the merge, instead we must just abandon the merge. If we
- // reopen or merge this could cause problems because the region has
- // probably already been moved to a different server, or is in the
- // process of moving to a different server.
- exceptionToThrow = closedByOtherException;
- }
- if (exceptionToThrow != closedByOtherException) {
- transition(isRegionA ? RegionMergeTransactionPhase.CLOSED_REGION_A
- : RegionMergeTransactionPhase.CLOSED_REGION_B);
- }
- if (exceptionToThrow != null) {
- if (exceptionToThrow instanceof IOException)
- throw (IOException) exceptionToThrow;
- throw new IOException(exceptionToThrow);
- }
- if (!testing) {
- services.removeFromOnlineRegions(region, null);
- }
-
- transition(isRegionA ? RegionMergeTransactionPhase.OFFLINED_REGION_A
- : RegionMergeTransactionPhase.OFFLINED_REGION_B);
-
- return hstoreFilesToMerge;
- }
-
- /**
- * Get merged region info through the specified two regions
- * @param a merging region A
- * @param b merging region B
- * @return the merged region info
- */
- @VisibleForTesting
- static HRegionInfo getMergedRegionInfo(final HRegionInfo a, final HRegionInfo b) {
- long rid = EnvironmentEdgeManager.currentTime();
- // Regionid is timestamp. Merged region's id can't be less than that of
- // merging regions else will insert at wrong location in hbase:meta
- if (rid < a.getRegionId() || rid < b.getRegionId()) {
- LOG.warn("Clock skew; merging regions id are " + a.getRegionId()
- + " and " + b.getRegionId() + ", but current time here is " + rid);
- rid = Math.max(a.getRegionId(), b.getRegionId()) + 1;
- }
-
- byte[] startKey = null;
- byte[] endKey = null;
- // Choose the smaller as start key
- if (a.compareTo(b) <= 0) {
- startKey = a.getStartKey();
- } else {
- startKey = b.getStartKey();
- }
- // Choose the bigger as end key
- if (Bytes.equals(a.getEndKey(), HConstants.EMPTY_BYTE_ARRAY)
- || (!Bytes.equals(b.getEndKey(), HConstants.EMPTY_BYTE_ARRAY)
- && Bytes.compareTo(a.getEndKey(), b.getEndKey()) > 0)) {
- endKey = a.getEndKey();
- } else {
- endKey = b.getEndKey();
- }
-
- // Merged region is sorted between two merging regions in META
- HRegionInfo mergedRegionInfo = new HRegionInfo(a.getTable(), startKey,
- endKey, false, rid);
- return mergedRegionInfo;
- }
-
- /**
- * Perform time consuming opening of the merged region.
- * @param server Hosting server instance. Can be null when testing
- * @param services Used to online/offline regions.
- * @param merged the merged region
- * @throws IOException If thrown, transaction failed. Call
- * {@link #rollback(Server, RegionServerServices)}
- */
- @VisibleForTesting
- void openMergedRegion(final Server server, final RegionServerServices services,
- HRegion merged) throws IOException {
- boolean stopped = server != null && server.isStopped();
- boolean stopping = services != null && services.isStopping();
- if (stopped || stopping) {
- LOG.info("Not opening merged region " + merged.getRegionInfo().getRegionNameAsString()
- + " because stopping=" + stopping + ", stopped=" + stopped);
- return;
- }
- HRegionInfo hri = merged.getRegionInfo();
- LoggingProgressable reporter = server == null ? null
- : new LoggingProgressable(hri, server.getConfiguration().getLong(
- "hbase.regionserver.regionmerge.open.log.interval", 10000));
- merged.openHRegion(reporter);
-
- if (services != null) {
- if (!services.reportRegionStateTransition(TransitionCode.MERGED,
- mergedRegionInfo, region_a.getRegionInfo(), region_b.getRegionInfo())) {
- throw new IOException("Failed to report merged region to master: "
- + mergedRegionInfo.getShortNameToLog());
- }
- services.addToOnlineRegions(merged);
- }
- }
-
- /**
- * Create reference file(s) of merging regions under the region_a merges dir
- * @param hstoreFilesOfRegionA
- * @param hstoreFilesOfRegionB
- * @throws IOException
- */
- private void mergeStoreFiles(
- Map<byte[], List<StoreFile>> hstoreFilesOfRegionA,
- Map<byte[], List<StoreFile>> hstoreFilesOfRegionB)
- throws IOException {
- // Create reference file(s) of region A in mergdir
- HRegionFileSystem fs_a = this.region_a.getRegionFileSystem();
- for (Map.Entry<byte[], List<StoreFile>> entry : hstoreFilesOfRegionA
- .entrySet()) {
- String familyName = Bytes.toString(entry.getKey());
- for (StoreFile storeFile : entry.getValue()) {
- fs_a.mergeStoreFile(this.mergedRegionInfo, familyName, storeFile,
- this.mergesdir);
- }
- }
- // Create reference file(s) of region B in mergedir
- HRegionFileSystem fs_b = this.region_b.getRegionFileSystem();
- for (Map.Entry<byte[], List<StoreFile>> entry : hstoreFilesOfRegionB
- .entrySet()) {
- String familyName = Bytes.toString(entry.getKey());
- for (StoreFile storeFile : entry.getValue()) {
- fs_b.mergeStoreFile(this.mergedRegionInfo, familyName, storeFile,
- this.mergesdir);
- }
- }
- }
-
- @Override
- public boolean rollback(final Server server,
- final RegionServerServices services) throws IOException {
- if (User.isHBaseSecurityEnabled(region_a.getBaseConf())) {
- LOG.warn("Should use execute(Server, RegionServerServices, User)");
- }
- return rollback(server, services, null);
- }
-
- @Override
- public boolean rollback(final Server server,
- final RegionServerServices services, User user) throws IOException {
- assert this.mergedRegionInfo != null;
- this.server = server;
- this.rsServices = services;
- // Coprocessor callback
- if (rsCoprocessorHost != null) {
- rsCoprocessorHost.preRollBackMerge(region_a, region_b, user);
- }
-
- boolean result = true;
- ListIterator<JournalEntry> iterator = this.journal
- .listIterator(this.journal.size());
- // Iterate in reverse.
- while (iterator.hasPrevious()) {
- JournalEntry je = iterator.previous();
-
- transition(je.getPhase(), true);
-
- switch (je.getPhase()) {
-
- case SET_MERGING:
- if (services != null
- && !services.reportRegionStateTransition(TransitionCode.MERGE_REVERTED,
- mergedRegionInfo, region_a.getRegionInfo(), region_b.getRegionInfo())) {
- return false;
- }
- break;
-
- case CREATED_MERGE_DIR:
- this.region_a.writestate.writesEnabled = true;
- this.region_b.writestate.writesEnabled = true;
- this.region_a.getRegionFileSystem().cleanupMergesDir();
- break;
-
- case CLOSED_REGION_A:
- try {
- // So, this returns a seqid but if we just closed and then reopened,
- // we should be ok. On close, we flushed using sequenceid obtained
- // from hosting regionserver so no need to propagate the sequenceid
- // returned out of initialize below up into regionserver as we
- // normally do.
- this.region_a.initialize();
- } catch (IOException e) {
- LOG.error("Failed rollbacking CLOSED_REGION_A of region "
- + region_a.getRegionInfo().getRegionNameAsString(), e);
- throw new RuntimeException(e);
- }
- break;
-
- case OFFLINED_REGION_A:
- if (services != null)
- services.addToOnlineRegions(this.region_a);
- break;
-
- case CLOSED_REGION_B:
- try {
- this.region_b.initialize();
- } catch (IOException e) {
- LOG.error("Failed rollbacking CLOSED_REGION_A of region "
- + region_b.getRegionInfo().getRegionNameAsString(), e);
- throw new RuntimeException(e);
- }
- break;
-
- case OFFLINED_REGION_B:
- if (services != null)
- services.addToOnlineRegions(this.region_b);
- break;
-
- case STARTED_MERGED_REGION_CREATION:
- this.region_a.getRegionFileSystem().cleanupMergedRegion(
- this.mergedRegionInfo);
- break;
-
- case PONR:
- // We got to the point-of-no-return so we need to just abort. Return
- // immediately. Do not clean up created merged regions.
- return false;
-
- // Informational states only
- case STARTED:
- case PREPARED:
- case COMPLETED:
- break;
-
- default:
- throw new RuntimeException("Unhandled journal entry: " + je);
- }
- }
- // Coprocessor callback
- if (rsCoprocessorHost != null) {
- rsCoprocessorHost.postRollBackMerge(region_a, region_b, user);
- }
-
- return result;
- }
-
- @Override
- public HRegionInfo getMergedRegionInfo() {
- return this.mergedRegionInfo;
- }
-
- @VisibleForTesting
- Path getMergesDir() {
- return this.mergesdir;
- }
-
- /**
- * Checks if the given region has merge qualifier in hbase:meta
- * @param services
- * @param regionName name of specified region
- * @return true if the given region has merge qualifier in META.(It will be
- * cleaned by CatalogJanitor)
- * @throws IOException
- */
- @VisibleForTesting
- boolean hasMergeQualifierInMeta(final RegionServerServices services, final byte[] regionName)
- throws IOException {
- if (services == null) return false;
- // Get merge regions if it is a merged region and already has merge
- // qualifier
- Pair<HRegionInfo, HRegionInfo> mergeRegions = MetaTableAccessor
- .getRegionsFromMergeQualifier(services.getConnection(), regionName);
- if (mergeRegions != null &&
- (mergeRegions.getFirst() != null || mergeRegions.getSecond() != null)) {
- // It has merge qualifier
- return true;
- }
- return false;
- }
-
- @Override
- public List<JournalEntry> getJournal() {
- return journal;
- }
-
- @Override
- public RegionMergeTransaction registerTransactionListener(TransactionListener listener) {
- listeners.add(listener);
- return this;
- }
-
- @Override
- public Server getServer() {
- return server;
- }
-
- @Override
- public RegionServerServices getRegionServerServices() {
- return rsServices;
- }
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HMerge.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HMerge.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HMerge.java
deleted file mode 100644
index 307568c..0000000
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HMerge.java
+++ /dev/null
@@ -1,348 +0,0 @@
-/**
- *
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.hadoop.hbase.util;
-
-import java.io.IOException;
-import java.util.ArrayList;
-import java.util.List;
-import java.util.NoSuchElementException;
-import java.util.Random;
-
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
-import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.fs.FileSystem;
-import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.hbase.HConstants;
-import org.apache.hadoop.hbase.HRegionInfo;
-import org.apache.hadoop.hbase.HTableDescriptor;
-import org.apache.hadoop.hbase.MetaTableAccessor;
-import org.apache.hadoop.hbase.TableName;
-import org.apache.hadoop.hbase.TableNotDisabledException;
-import org.apache.hadoop.hbase.classification.InterfaceAudience;
-import org.apache.hadoop.hbase.client.Admin;
-import org.apache.hadoop.hbase.client.ClusterConnection;
-import org.apache.hadoop.hbase.client.Connection;
-import org.apache.hadoop.hbase.client.ConnectionFactory;
-import org.apache.hadoop.hbase.client.Delete;
-import org.apache.hadoop.hbase.client.Result;
-import org.apache.hadoop.hbase.client.ResultScanner;
-import org.apache.hadoop.hbase.client.Table;
-import org.apache.hadoop.hbase.regionserver.HRegion;
-import org.apache.hadoop.hbase.wal.WALFactory;
-import org.apache.hadoop.ipc.RemoteException;
-
-/**
- * A non-instantiable class that has a static method capable of compacting
- * a table by merging adjacent regions.
- */
-@InterfaceAudience.Private
-class HMerge {
- // TODO: Where is this class used? How does it relate to Merge in same package?
- private static final Log LOG = LogFactory.getLog(HMerge.class);
- static final Random rand = new Random();
-
- /*
- * Not instantiable
- */
- private HMerge() {
- super();
- }
-
- /**
- * Scans the table and merges two adjacent regions if they are small. This
- * only happens when a lot of rows are deleted.
- *
- * When merging the hbase:meta region, the HBase instance must be offline.
- * When merging a normal table, the HBase instance must be online, but the
- * table must be disabled.
- *
- * @param conf - configuration object for HBase
- * @param fs - FileSystem where regions reside
- * @param tableName - Table to be compacted
- * @throws IOException
- */
- public static void merge(Configuration conf, FileSystem fs,
- final TableName tableName)
- throws IOException {
- merge(conf, fs, tableName, true);
- }
-
- /**
- * Scans the table and merges two adjacent regions if they are small. This
- * only happens when a lot of rows are deleted.
- *
- * When merging the hbase:meta region, the HBase instance must be offline.
- * When merging a normal table, the HBase instance must be online, but the
- * table must be disabled.
- *
- * @param conf - configuration object for HBase
- * @param fs - FileSystem where regions reside
- * @param tableName - Table to be compacted
- * @param testMasterRunning True if we are to verify master is down before
- * running merge
- * @throws IOException
- */
- public static void merge(Configuration conf, FileSystem fs,
- final TableName tableName, final boolean testMasterRunning)
- throws IOException {
- boolean masterIsRunning = false;
- ClusterConnection hConnection = null;
- if (testMasterRunning) {
- try {
- hConnection = (ClusterConnection) ConnectionFactory.createConnection(conf);
- masterIsRunning = hConnection.isMasterRunning();
- } finally {
- if (hConnection != null) {
- hConnection.close();
- }
- }
- }
- if (tableName.equals(TableName.META_TABLE_NAME)) {
- if (masterIsRunning) {
- throw new IllegalStateException(
- "Can not compact hbase:meta table if instance is on-line");
- }
- // TODO reenable new OfflineMerger(conf, fs).process();
- } else {
- if(!masterIsRunning) {
- throw new IllegalStateException(
- "HBase instance must be running to merge a normal table");
- }
- try (Connection conn = ConnectionFactory.createConnection(conf);
- Admin admin = conn.getAdmin()) {
- if (!admin.isTableDisabled(tableName)) {
- throw new TableNotDisabledException(tableName);
- }
- }
- new OnlineMerger(conf, fs, tableName).process();
- }
- }
-
- private static abstract class Merger {
- protected final Configuration conf;
- protected final FileSystem fs;
- protected final Path rootDir;
- protected final HTableDescriptor htd;
- protected final WALFactory walFactory;
- private final long maxFilesize;
-
-
- protected Merger(Configuration conf, FileSystem fs, final TableName tableName)
- throws IOException {
- this.conf = conf;
- this.fs = fs;
- this.maxFilesize = conf.getLong(HConstants.HREGION_MAX_FILESIZE,
- HConstants.DEFAULT_MAX_FILE_SIZE);
-
- this.rootDir = FSUtils.getRootDir(conf);
- Path tabledir = FSUtils.getTableDir(this.rootDir, tableName);
- this.htd = FSTableDescriptors.getTableDescriptorFromFs(this.fs, tabledir);
- String logname = "merge_" + System.currentTimeMillis() + HConstants.HREGION_LOGDIR_NAME;
-
- final Configuration walConf = new Configuration(conf);
- FSUtils.setRootDir(walConf, tabledir);
- this.walFactory = new WALFactory(walConf, null, logname);
- }
-
- void process() throws IOException {
- try {
- for (HRegionInfo[] regionsToMerge = next();
- regionsToMerge != null;
- regionsToMerge = next()) {
- if (!merge(regionsToMerge)) {
- return;
- }
- }
- } finally {
- try {
- walFactory.close();
- } catch(IOException e) {
- LOG.error(e);
- }
- }
- }
-
- protected boolean merge(final HRegionInfo[] info) throws IOException {
- if (info.length < 2) {
- LOG.info("only one region - nothing to merge");
- return false;
- }
-
- HRegion currentRegion = null;
- long currentSize = 0;
- HRegion nextRegion = null;
- long nextSize = 0;
- for (int i = 0; i < info.length - 1; i++) {
- if (currentRegion == null) {
- currentRegion = HRegion.openHRegion(conf, fs, this.rootDir, info[i], this.htd,
- walFactory.getWAL(info[i].getEncodedNameAsBytes(),
- info[i].getTable().getNamespace()));
- currentSize = currentRegion.getLargestHStoreSize();
- }
- nextRegion = HRegion.openHRegion(conf, fs, this.rootDir, info[i + 1], this.htd,
- walFactory.getWAL(info[i + 1].getEncodedNameAsBytes(),
- info[i + 1].getTable().getNamespace()));
- nextSize = nextRegion.getLargestHStoreSize();
-
- if ((currentSize + nextSize) <= (maxFilesize / 2)) {
- // We merge two adjacent regions if their total size is less than
- // one half of the desired maximum size
- LOG.info("Merging regions " + currentRegion.getRegionInfo().getRegionNameAsString() +
- " and " + nextRegion.getRegionInfo().getRegionNameAsString());
- HRegion mergedRegion =
- HRegion.mergeAdjacent(currentRegion, nextRegion);
- updateMeta(currentRegion.getRegionInfo().getRegionName(),
- nextRegion.getRegionInfo().getRegionName(), mergedRegion);
- break;
- }
- LOG.info("not merging regions " +
- Bytes.toStringBinary(currentRegion.getRegionInfo().getRegionName()) +
- " and " + Bytes.toStringBinary(nextRegion.getRegionInfo().getRegionName()));
- currentRegion.close();
- currentRegion = nextRegion;
- currentSize = nextSize;
- }
- if(currentRegion != null) {
- currentRegion.close();
- }
- return true;
- }
-
- protected abstract HRegionInfo[] next() throws IOException;
-
- protected abstract void updateMeta(final byte [] oldRegion1,
- final byte [] oldRegion2, HRegion newRegion)
- throws IOException;
-
- }
-
- /** Instantiated to compact a normal user table */
- private static class OnlineMerger extends Merger {
- private final TableName tableName;
- private final Table table;
- private final ResultScanner metaScanner;
- private HRegionInfo latestRegion;
-
- OnlineMerger(Configuration conf, FileSystem fs,
- final TableName tableName)
- throws IOException {
- super(conf, fs, tableName);
- this.tableName = tableName;
- Connection connection = ConnectionFactory.createConnection(conf);
- this.table = connection.getTable(TableName.META_TABLE_NAME);
- this.metaScanner = table.getScanner(HConstants.CATALOG_FAMILY,
- HConstants.REGIONINFO_QUALIFIER);
- this.latestRegion = null;
- }
-
- private HRegionInfo nextRegion() throws IOException {
- try {
- Result results = getMetaRow();
- if (results == null) {
- return null;
- }
- HRegionInfo region = MetaTableAccessor.getHRegionInfo(results);
- if (region == null) {
- throw new NoSuchElementException("meta region entry missing " +
- Bytes.toString(HConstants.CATALOG_FAMILY) + ":" +
- Bytes.toString(HConstants.REGIONINFO_QUALIFIER));
- }
- if (!region.getTable().equals(this.tableName)) {
- return null;
- }
- return region;
- } catch (IOException e) {
- e = e instanceof RemoteException ? ((RemoteException) e).unwrapRemoteException() : e;
- LOG.error("meta scanner error", e);
- metaScanner.close();
- throw e;
- }
- }
-
- /*
- * Check current row has a HRegionInfo. Skip to next row if HRI is empty.
- * @return A Map of the row content else null if we are off the end.
- * @throws IOException
- */
- private Result getMetaRow() throws IOException {
- Result currentRow = metaScanner.next();
- boolean foundResult = false;
- while (currentRow != null) {
- LOG.info("Row: <" + Bytes.toStringBinary(currentRow.getRow()) + ">");
- byte[] regionInfoValue = currentRow.getValue(HConstants.CATALOG_FAMILY,
- HConstants.REGIONINFO_QUALIFIER);
- if (regionInfoValue == null || regionInfoValue.length == 0) {
- currentRow = metaScanner.next();
- continue;
- }
- HRegionInfo region = MetaTableAccessor.getHRegionInfo(currentRow);
- if (!region.getTable().equals(this.tableName)) {
- currentRow = metaScanner.next();
- continue;
- }
- foundResult = true;
- break;
- }
- return foundResult ? currentRow : null;
- }
-
- @Override
- protected HRegionInfo[] next() throws IOException {
- List<HRegionInfo> regions = new ArrayList<HRegionInfo>();
- if(latestRegion == null) {
- latestRegion = nextRegion();
- }
- if(latestRegion != null) {
- regions.add(latestRegion);
- }
- latestRegion = nextRegion();
- if(latestRegion != null) {
- regions.add(latestRegion);
- }
- return regions.toArray(new HRegionInfo[regions.size()]);
- }
-
- @Override
- protected void updateMeta(final byte [] oldRegion1,
- final byte [] oldRegion2,
- HRegion newRegion)
- throws IOException {
- byte[][] regionsToDelete = {oldRegion1, oldRegion2};
- for (int r = 0; r < regionsToDelete.length; r++) {
- if(Bytes.equals(regionsToDelete[r], latestRegion.getRegionName())) {
- latestRegion = null;
- }
- Delete delete = new Delete(regionsToDelete[r]);
- table.delete(delete);
- if(LOG.isDebugEnabled()) {
- LOG.debug("updated columns in row: " + Bytes.toStringBinary(regionsToDelete[r]));
- }
- }
- newRegion.getRegionInfo().setOffline(true);
-
- MetaTableAccessor.addRegionToMeta(table, newRegion.getRegionInfo());
-
- if(LOG.isDebugEnabled()) {
- LOG.debug("updated columns in row: "
- + Bytes.toStringBinary(newRegion.getRegionInfo().getRegionName()));
- }
- }
- }
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/main/java/org/apache/hadoop/hbase/util/Merge.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/Merge.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/Merge.java
deleted file mode 100644
index 7b96660..0000000
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/Merge.java
+++ /dev/null
@@ -1,264 +0,0 @@
-/**
- *
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-package org.apache.hadoop.hbase.util;
-
-import java.io.IOException;
-
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
-import org.apache.hadoop.conf.Configured;
-import org.apache.hadoop.fs.FileSystem;
-import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.hbase.HBaseConfiguration;
-import org.apache.hadoop.hbase.HBaseInterfaceAudience;
-import org.apache.hadoop.hbase.HConstants;
-import org.apache.hadoop.hbase.HRegionInfo;
-import org.apache.hadoop.hbase.HTableDescriptor;
-import org.apache.hadoop.hbase.MetaTableAccessor;
-import org.apache.hadoop.hbase.TableName;
-import org.apache.hadoop.hbase.ZooKeeperConnectionException;
-import org.apache.hadoop.hbase.classification.InterfaceAudience;
-import org.apache.hadoop.hbase.client.Delete;
-import org.apache.hadoop.hbase.client.Get;
-import org.apache.hadoop.hbase.client.HBaseAdmin;
-import org.apache.hadoop.hbase.client.Result;
-import org.apache.hadoop.hbase.regionserver.HRegion;
-import org.apache.hadoop.io.WritableComparator;
-import org.apache.hadoop.util.Tool;
-import org.apache.hadoop.util.ToolRunner;
-
-import com.google.common.base.Preconditions;
-
-/**
- * Utility that can merge any two regions in the same table: adjacent,
- * overlapping or disjoint.
- */
-@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS)
-public class Merge extends Configured implements Tool {
- private static final Log LOG = LogFactory.getLog(Merge.class);
- private Path rootdir;
- private volatile MetaUtils utils;
- private TableName tableName; // Name of table
- private volatile byte [] region1; // Name of region 1
- private volatile byte [] region2; // Name of region 2
- private volatile HRegionInfo mergeInfo = null;
-
- @Override
- public int run(String[] args) throws Exception {
- if (parseArgs(args) != 0) {
- return -1;
- }
-
- // Verify file system is up.
- FileSystem fs = FileSystem.get(getConf()); // get DFS handle
- LOG.info("Verifying that file system is available...");
- try {
- FSUtils.checkFileSystemAvailable(fs);
- } catch (IOException e) {
- LOG.fatal("File system is not available", e);
- return -1;
- }
-
- // Verify HBase is down
- LOG.info("Verifying that HBase is not running...");
- try {
- HBaseAdmin.available(getConf());
- LOG.fatal("HBase cluster must be off-line, and is not. Aborting.");
- return -1;
- } catch (ZooKeeperConnectionException zkce) {
- // If no zk, presume no master.
- }
-
- // Initialize MetaUtils and and get the root of the HBase installation
-
- this.utils = new MetaUtils(getConf());
- this.rootdir = FSUtils.getRootDir(getConf());
- try {
- mergeTwoRegions();
- return 0;
- } catch (IOException e) {
- LOG.fatal("Merge failed", e);
- return -1;
-
- } finally {
- if (this.utils != null) {
- this.utils.shutdown();
- }
- }
- }
-
- /** @return HRegionInfo for merge result */
- HRegionInfo getMergedHRegionInfo() {
- return this.mergeInfo;
- }
-
- /*
- * Merges two regions from a user table.
- */
- private void mergeTwoRegions() throws IOException {
- LOG.info("Merging regions " + Bytes.toStringBinary(this.region1) + " and " +
- Bytes.toStringBinary(this.region2) + " in table " + this.tableName);
- HRegion meta = this.utils.getMetaRegion();
- Get get = new Get(region1);
- get.addColumn(HConstants.CATALOG_FAMILY, HConstants.REGIONINFO_QUALIFIER);
- Result result1 = meta.get(get);
- Preconditions.checkState(!result1.isEmpty(),
- "First region cells can not be null");
- HRegionInfo info1 = MetaTableAccessor.getHRegionInfo(result1);
- if (info1 == null) {
- throw new NullPointerException("info1 is null using key " +
- Bytes.toStringBinary(region1) + " in " + meta);
- }
- get = new Get(region2);
- get.addColumn(HConstants.CATALOG_FAMILY, HConstants.REGIONINFO_QUALIFIER);
- Result result2 = meta.get(get);
- Preconditions.checkState(!result2.isEmpty(),
- "Second region cells can not be null");
- HRegionInfo info2 = MetaTableAccessor.getHRegionInfo(result2);
- if (info2 == null) {
- throw new NullPointerException("info2 is null using key " + meta);
- }
- HTableDescriptor htd = FSTableDescriptors.getTableDescriptorFromFs(FileSystem.get(getConf()),
- this.rootdir, this.tableName);
- HRegion merged = merge(htd, meta, info1, info2);
-
- LOG.info("Adding " + merged.getRegionInfo() + " to " +
- meta.getRegionInfo());
-
- HRegion.addRegionToMETA(meta, merged);
- merged.close();
- }
-
- /*
- * Actually merge two regions and update their info in the meta region(s)
- * Returns HRegion object for newly merged region
- */
- private HRegion merge(final HTableDescriptor htd, HRegion meta,
- HRegionInfo info1, HRegionInfo info2)
- throws IOException {
- if (info1 == null) {
- throw new IOException("Could not find " + Bytes.toStringBinary(region1) + " in " +
- Bytes.toStringBinary(meta.getRegionInfo().getRegionName()));
- }
- if (info2 == null) {
- throw new IOException("Could not find " + Bytes.toStringBinary(region2) + " in " +
- Bytes.toStringBinary(meta.getRegionInfo().getRegionName()));
- }
- HRegion merged = null;
- HRegion r1 = HRegion.openHRegion(info1, htd, utils.getLog(info1), getConf());
- try {
- HRegion r2 = HRegion.openHRegion(info2, htd, utils.getLog(info2), getConf());
- try {
- merged = HRegion.merge(r1, r2);
- } finally {
- if (!r2.isClosed()) {
- r2.close();
- }
- }
- } finally {
- if (!r1.isClosed()) {
- r1.close();
- }
- }
-
- // Remove the old regions from meta.
- // HRegion.merge has already deleted their files
-
- removeRegionFromMeta(meta, info1);
- removeRegionFromMeta(meta, info2);
-
- this.mergeInfo = merged.getRegionInfo();
- return merged;
- }
-
- /*
- * Removes a region's meta information from the passed <code>meta</code>
- * region.
- *
- * @param meta hbase:meta HRegion to be updated
- * @param regioninfo HRegionInfo of region to remove from <code>meta</code>
- *
- * @throws IOException
- */
- private void removeRegionFromMeta(HRegion meta, HRegionInfo regioninfo)
- throws IOException {
- if (LOG.isDebugEnabled()) {
- LOG.debug("Removing region: " + regioninfo + " from " + meta);
- }
-
- Delete delete = new Delete(regioninfo.getRegionName(),
- System.currentTimeMillis());
- meta.delete(delete);
- }
-
- /**
- * Parse given arguments and assign table name and regions names.
- * (generic args are handled by ToolRunner.)
- *
- * @param args the arguments to parse
- *
- * @throws IOException
- */
- private int parseArgs(String[] args) throws IOException {
- if (args.length != 3) {
- usage();
- return -1;
- }
- tableName = TableName.valueOf(args[0]);
-
- region1 = Bytes.toBytesBinary(args[1]);
- region2 = Bytes.toBytesBinary(args[2]);
- int status = 0;
- if (notInTable(tableName, region1) || notInTable(tableName, region2)) {
- status = -1;
- } else if (Bytes.equals(region1, region2)) {
- LOG.error("Can't merge a region with itself");
- status = -1;
- }
- return status;
- }
-
- private boolean notInTable(final TableName tn, final byte [] rn) {
- if (WritableComparator.compareBytes(tn.getName(), 0, tn.getName().length,
- rn, 0, tn.getName().length) != 0) {
- LOG.error("Region " + Bytes.toStringBinary(rn) + " does not belong to table " +
- tn);
- return true;
- }
- return false;
- }
-
- private void usage() {
- System.err
- .println("For hadoop 0.21+, Usage: hbase org.apache.hadoop.hbase.util.Merge "
- + "[-Dfs.defaultFS=hdfs://nn:port] <table-name> <region-1> <region-2>\n");
- }
-
- public static void main(String[] args) {
- int status;
- try {
- status = ToolRunner.run(HBaseConfiguration.create(), new Merge(), args);
- } catch (Exception e) {
- LOG.error("exiting due to error", e);
- status = -1;
- }
- System.exit(status);
- }
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin1.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin1.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin1.java
index 65cedda..b138c7d 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin1.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin1.java
@@ -55,7 +55,7 @@ import org.apache.hadoop.hbase.master.HMaster;
import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
import org.apache.hadoop.hbase.shaded.protobuf.RequestConverter;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest;
import org.apache.hadoop.hbase.regionserver.HRegion;
import org.apache.hadoop.hbase.regionserver.Store;
import org.apache.hadoop.hbase.regionserver.StoreFile;
@@ -1189,15 +1189,17 @@ public class TestAdmin1 {
assertTrue(gotException);
// Try going to the master directly (that will skip the check in admin)
try {
- DispatchMergingRegionsRequest request = RequestConverter
- .buildDispatchMergingRegionsRequest(
- regions.get(1).getFirst().getEncodedNameAsBytes(),
- regions.get(2).getFirst().getEncodedNameAsBytes(),
+ byte[][] nameofRegionsToMerge = new byte[2][];
+ nameofRegionsToMerge[0] = regions.get(1).getFirst().getEncodedNameAsBytes();
+ nameofRegionsToMerge[1] = regions.get(2).getFirst().getEncodedNameAsBytes();
+ MergeTableRegionsRequest request = RequestConverter
+ .buildMergeTableRegionsRequest(
+ nameofRegionsToMerge,
true,
HConstants.NO_NONCE,
- HConstants.NO_NONCE);
+ HConstants.NO_NONCE);
((ClusterConnection) TEST_UTIL.getAdmin().getConnection()).getMaster()
- .dispatchMergingRegions(null, request);
+ .mergeTableRegions(null, request);
} catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException m) {
Throwable t = m.getCause();
do {
@@ -1209,40 +1211,6 @@ public class TestAdmin1 {
} while (t != null);
}
assertTrue(gotException);
- gotException = false;
- // Try going to the regionservers directly
- // first move the region to the same regionserver
- if (!regions.get(2).getSecond().equals(regions.get(1).getSecond())) {
- moveRegionAndWait(regions.get(2).getFirst(), regions.get(1).getSecond());
- }
- try {
- AdminService.BlockingInterface admin = ((ClusterConnection) TEST_UTIL.getAdmin()
- .getConnection()).getAdmin(regions.get(1).getSecond());
- ProtobufUtil.mergeRegions(null, admin, regions.get(1).getFirst(), regions.get(2).getFirst(),
- true, null);
- } catch (MergeRegionException mm) {
- gotException = true;
- }
- assertTrue(gotException);
- }
-
- private void moveRegionAndWait(HRegionInfo destRegion, ServerName destServer)
- throws InterruptedException, MasterNotRunningException,
- ZooKeeperConnectionException, IOException {
- HMaster master = TEST_UTIL.getMiniHBaseCluster().getMaster();
- TEST_UTIL.getAdmin().move(
- destRegion.getEncodedNameAsBytes(),
- Bytes.toBytes(destServer.getServerName()));
- while (true) {
- ServerName serverName = master.getAssignmentManager()
- .getRegionStates().getRegionServerOfRegion(destRegion);
- if (serverName != null && serverName.equals(destServer)) {
- TEST_UTIL.assertRegionOnServer(
- destRegion, serverName, 200);
- break;
- }
- Thread.sleep(10);
- }
}
@Test (expected=IllegalArgumentException.class, timeout=300000)
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerObserver.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerObserver.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerObserver.java
deleted file mode 100644
index 4a62bff..0000000
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerObserver.java
+++ /dev/null
@@ -1,223 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.hadoop.hbase.coprocessor;
-
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
-import static org.junit.Assert.assertTrue;
-
-import java.io.IOException;
-import java.util.List;
-
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
-import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.hbase.Coprocessor;
-import org.apache.hadoop.hbase.HBaseConfiguration;
-import org.apache.hadoop.hbase.HBaseTestingUtility;
-import org.apache.hadoop.hbase.HColumnDescriptor;
-import org.apache.hadoop.hbase.HTableDescriptor;
-import org.apache.hadoop.hbase.MiniHBaseCluster;
-import org.apache.hadoop.hbase.TableName;
-import org.apache.hadoop.hbase.MetaTableAccessor;
-import org.apache.hadoop.hbase.client.Admin;
-import org.apache.hadoop.hbase.client.Mutation;
-import org.apache.hadoop.hbase.regionserver.HRegion;
-import org.apache.hadoop.hbase.regionserver.HRegionServer;
-import org.apache.hadoop.hbase.regionserver.Region;
-import org.apache.hadoop.hbase.regionserver.RegionMergeTransactionFactory;
-import org.apache.hadoop.hbase.regionserver.RegionMergeTransactionImpl;
-import org.apache.hadoop.hbase.regionserver.RegionServerCoprocessorHost;
-import org.apache.hadoop.hbase.testclassification.CoprocessorTests;
-import org.apache.hadoop.hbase.testclassification.MediumTests;
-import org.apache.hadoop.hbase.util.Bytes;
-import org.junit.Ignore;
-import org.junit.Test;
-import org.junit.experimental.categories.Category;
-
-/**
- * Tests invocation of the {@link org.apache.hadoop.hbase.coprocessor.RegionServerObserver}
- * interface hooks at all appropriate times during normal HMaster operations.
- */
-@Category({CoprocessorTests.class, MediumTests.class})
-public class TestRegionServerObserver {
- private static final Log LOG = LogFactory.getLog(TestRegionServerObserver.class);
-
- /**
- * Test verifies the hooks in regions merge.
- * @throws Exception
- */
- @Ignore
- @Test
- public void testCoprocessorHooksInRegionsMerge() throws Exception {
- final int NUM_MASTERS = 1;
- final int NUM_RS = 1;
- final String TABLENAME = "testRegionServerObserver";
- final String TABLENAME2 = "testRegionServerObserver_2";
- final byte[] FAM = Bytes.toBytes("fam");
-
- // Create config to use for this cluster
- Configuration conf = HBaseConfiguration.create();
- conf.setClass("hbase.coprocessor.regionserver.classes", CPRegionServerObserver.class,
- RegionServerObserver.class);
-
- // Start the cluster
- HBaseTestingUtility TEST_UTIL = new HBaseTestingUtility(conf);
- TEST_UTIL.startMiniCluster(NUM_MASTERS, NUM_RS);
- Admin admin = TEST_UTIL.getHBaseAdmin();
- try {
- MiniHBaseCluster cluster = TEST_UTIL.getHBaseCluster();
- HRegionServer regionServer = cluster.getRegionServer(0);
- RegionServerCoprocessorHost cpHost = regionServer.getRegionServerCoprocessorHost();
- Coprocessor coprocessor = cpHost.findCoprocessor(CPRegionServerObserver.class.getName());
- CPRegionServerObserver regionServerObserver = (CPRegionServerObserver) coprocessor;
- HTableDescriptor desc = new HTableDescriptor(TableName.valueOf(TABLENAME));
- desc.addFamily(new HColumnDescriptor(FAM));
- admin.createTable(desc, new byte[][] { Bytes.toBytes("row") });
- desc = new HTableDescriptor(TableName.valueOf(TABLENAME2));
- desc.addFamily(new HColumnDescriptor(FAM));
- admin.createTable(desc, new byte[][] { Bytes.toBytes("row") });
- assertFalse(regionServerObserver.wasRegionMergeCalled());
- List<Region> regions = regionServer.getOnlineRegions(TableName.valueOf(TABLENAME));
- admin.mergeRegionsAsync(
- regions.get(0).getRegionInfo().getEncodedNameAsBytes(),
- regions.get(1).getRegionInfo().getEncodedNameAsBytes(),
- true);
- int regionsCount = regionServer.getOnlineRegions(TableName.valueOf(TABLENAME)).size();
- while (regionsCount != 1) {
- regionsCount = regionServer.getOnlineRegions(TableName.valueOf(TABLENAME)).size();
- Thread.sleep(1000);
- }
- assertTrue(regionServerObserver.wasRegionMergeCalled());
- assertTrue(regionServerObserver.wasPreMergeCommit());
- assertTrue(regionServerObserver.wasPostMergeCommit());
- assertEquals(regionsCount, 1);
- assertEquals(regionServer.getOnlineRegions(TableName.valueOf(TABLENAME2)).size(), 1);
- } finally {
- if (admin != null) admin.close();
- TEST_UTIL.shutdownMiniCluster();
- }
- }
-
- public static class CPRegionServerObserver extends BaseRegionServerObserver {
- private RegionMergeTransactionImpl rmt = null;
- private HRegion mergedRegion = null;
-
- private boolean preMergeCalled;
- private boolean preMergeBeforePONRCalled;
- private boolean preMergeAfterPONRCalled;
- private boolean preRollBackMergeCalled;
- private boolean postRollBackMergeCalled;
- private boolean postMergeCalled;
-
- public void resetStates() {
- preMergeCalled = false;
- preMergeBeforePONRCalled = false;
- preMergeAfterPONRCalled = false;
- preRollBackMergeCalled = false;
- postRollBackMergeCalled = false;
- postMergeCalled = false;
- }
-
- @Override
- public void preMerge(ObserverContext<RegionServerCoprocessorEnvironment> ctx, Region regionA,
- Region regionB) throws IOException {
- preMergeCalled = true;
- }
-
- @Override
- public void preMergeCommit(ObserverContext<RegionServerCoprocessorEnvironment> ctx,
- Region regionA, Region regionB, List<Mutation> metaEntries) throws IOException {
- preMergeBeforePONRCalled = true;
- RegionServerCoprocessorEnvironment environment = ctx.getEnvironment();
- HRegionServer rs = (HRegionServer) environment.getRegionServerServices();
- List<Region> onlineRegions =
- rs.getOnlineRegions(TableName.valueOf("testRegionServerObserver_2"));
- rmt = (RegionMergeTransactionImpl) new RegionMergeTransactionFactory(rs.getConfiguration())
- .create(onlineRegions.get(0), onlineRegions.get(1), true);
- if (!rmt.prepare(rs)) {
- LOG.error("Prepare for the region merge of table "
- + onlineRegions.get(0).getTableDesc().getNameAsString()
- + " failed. So returning null. ");
- ctx.bypass();
- return;
- }
- mergedRegion = rmt.stepsBeforePONR(rs, rs, false);
- rmt.prepareMutationsForMerge(mergedRegion.getRegionInfo(), regionA.getRegionInfo(),
- regionB.getRegionInfo(), rs.getServerName(), metaEntries);
- MetaTableAccessor.mutateMetaTable(rs.getConnection(), metaEntries);
- }
-
- @Override
- public void postMergeCommit(ObserverContext<RegionServerCoprocessorEnvironment> ctx,
- Region regionA, Region regionB, Region mr) throws IOException {
- preMergeAfterPONRCalled = true;
- RegionServerCoprocessorEnvironment environment = ctx.getEnvironment();
- HRegionServer rs = (HRegionServer) environment.getRegionServerServices();
- rmt.stepsAfterPONR(rs, rs, this.mergedRegion, null);
- }
-
- @Override
- public void preRollBackMerge(ObserverContext<RegionServerCoprocessorEnvironment> ctx,
- Region regionA, Region regionB) throws IOException {
- preRollBackMergeCalled = true;
- }
-
- @Override
- public void postRollBackMerge(ObserverContext<RegionServerCoprocessorEnvironment> ctx,
- Region regionA, Region regionB) throws IOException {
- postRollBackMergeCalled = true;
- }
-
- @Override
- public void postMerge(ObserverContext<RegionServerCoprocessorEnvironment> c, Region regionA,
- Region regionB, Region mergedRegion) throws IOException {
- postMergeCalled = true;
- }
-
- public boolean wasPreMergeCalled() {
- return this.preMergeCalled;
- }
-
- public boolean wasPostMergeCalled() {
- return this.postMergeCalled;
- }
-
- public boolean wasPreMergeCommit() {
- return this.preMergeBeforePONRCalled;
- }
-
- public boolean wasPostMergeCommit() {
- return this.preMergeAfterPONRCalled;
- }
-
- public boolean wasPreRollBackMerge() {
- return this.preRollBackMergeCalled;
- }
-
- public boolean wasPostRollBackMerge() {
- return this.postRollBackMergeCalled;
- }
-
- public boolean wasRegionMergeCalled() {
- return this.preMergeCalled && this.postMergeCalled;
- }
-
- }
-
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockNoopMasterServices.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockNoopMasterServices.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockNoopMasterServices.java
index 28bf14a..48cf8a5 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockNoopMasterServices.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockNoopMasterServices.java
@@ -320,16 +320,6 @@ public class MockNoopMasterServices implements MasterServices, Server {
}
@Override
- public long dispatchMergingRegions(
- final HRegionInfo region_a,
- final HRegionInfo region_b,
- final boolean forcible,
- final long nonceGroup,
- final long nonce) throws IOException {
- return -1;
- }
-
- @Override
public boolean isActiveMaster() {
return true;
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockRegionServer.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockRegionServer.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockRegionServer.java
index 950ec92..c5f294a 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockRegionServer.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockRegionServer.java
@@ -68,8 +68,6 @@ import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetServerIn
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetServerInfoResponse;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileResponse;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.OpenRegionRequest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.OpenRegionResponse;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryRequest;
@@ -531,13 +529,6 @@ ClientProtos.ClientService.BlockingInterface, RegionServerServices {
}
@Override
- public MergeRegionsResponse mergeRegions(RpcController controller,
- MergeRegionsRequest request) throws ServiceException {
- // TODO Auto-generated method stub
- return null;
- }
-
- @Override
public CompactRegionResponse compactRegion(RpcController controller,
CompactRegionRequest request) throws ServiceException {
// TODO Auto-generated method stub
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestDispatchMergingRegionsProcedure.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestDispatchMergingRegionsProcedure.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestDispatchMergingRegionsProcedure.java
deleted file mode 100644
index c011321..0000000
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestDispatchMergingRegionsProcedure.java
+++ /dev/null
@@ -1,264 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-package org.apache.hadoop.hbase.master.procedure;
-
-import java.io.IOException;
-import java.util.List;
-import java.util.concurrent.TimeUnit;
-
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
-import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.hbase.HBaseTestingUtility;
-import org.apache.hadoop.hbase.HColumnDescriptor;
-import org.apache.hadoop.hbase.HConstants;
-import org.apache.hadoop.hbase.HRegionInfo;
-import org.apache.hadoop.hbase.HTableDescriptor;
-import org.apache.hadoop.hbase.TableName;
-import org.apache.hadoop.hbase.client.Admin;
-import org.apache.hadoop.hbase.procedure2.ProcedureExecutor;
-import org.apache.hadoop.hbase.procedure2.ProcedureTestingUtility;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsState;
-import org.apache.hadoop.hbase.testclassification.MasterTests;
-import org.apache.hadoop.hbase.testclassification.MediumTests;
-import org.apache.hadoop.hbase.util.Bytes;
-import org.apache.hadoop.hbase.util.JVMClusterUtil.RegionServerThread;
-import org.junit.After;
-import org.junit.AfterClass;
-import org.junit.Before;
-import org.junit.BeforeClass;
-import org.junit.Test;
-import org.junit.experimental.categories.Category;
-
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertTrue;
-
-@Category({MasterTests.class, MediumTests.class})
-public class TestDispatchMergingRegionsProcedure {
- private static final Log LOG = LogFactory.getLog(TestDispatchMergingRegionsProcedure.class);
-
- protected static final HBaseTestingUtility UTIL = new HBaseTestingUtility();
-
- private final static byte[] FAMILY = Bytes.toBytes("FAMILY");
- final static Configuration conf = UTIL.getConfiguration();
- private static Admin admin;
-
- private static void setupConf(Configuration conf) {
- // Reduce the maximum attempts to speed up the test
- conf.setInt("hbase.assignment.maximum.attempts", 3);
- conf.setInt("hbase.master.maximum.ping.server.attempts", 3);
- conf.setInt("hbase.master.ping.server.retry.sleep.interval", 1);
-
- conf.setInt(MasterProcedureConstants.MASTER_PROCEDURE_THREADS, 1);
- }
-
- @BeforeClass
- public static void setupCluster() throws Exception {
- setupConf(conf);
- UTIL.startMiniCluster(1);
- admin = UTIL.getHBaseAdmin();
- }
-
- @AfterClass
- public static void cleanupTest() throws Exception {
- try {
- UTIL.shutdownMiniCluster();
- } catch (Exception e) {
- LOG.warn("failure shutting down cluster", e);
- }
- }
-
- @Before
- public void setup() throws Exception {
- resetProcExecutorTestingKillFlag();
-
- // Turn off balancer so it doesn't cut in and mess up our placements.
- UTIL.getHBaseAdmin().setBalancerRunning(false, true);
- // Turn off the meta scanner so it don't remove parent on us.
- UTIL.getHBaseCluster().getMaster().setCatalogJanitorEnabled(false);
- resetProcExecutorTestingKillFlag();
- }
-
- @After
- public void tearDown() throws Exception {
- resetProcExecutorTestingKillFlag();
- for (HTableDescriptor htd: UTIL.getHBaseAdmin().listTables()) {
- LOG.info("Tear down, remove table=" + htd.getTableName());
- UTIL.deleteTable(htd.getTableName());
- }
- }
-
- private void resetProcExecutorTestingKillFlag() {
- final ProcedureExecutor<MasterProcedureEnv> procExec = getMasterProcedureExecutor();
- ProcedureTestingUtility.setKillAndToggleBeforeStoreUpdate(procExec, false);
- assertTrue("expected executor to be running", procExec.isRunning());
- }
-
- /**
- * This tests two region merges
- */
- @Test(timeout=60000)
- public void testMergeTwoRegions() throws Exception {
- final TableName tableName = TableName.valueOf("testMergeTwoRegions");
- final ProcedureExecutor<MasterProcedureEnv> procExec = getMasterProcedureExecutor();
-
- List<HRegionInfo> tableRegions = createTable(tableName, 3);
-
- HRegionInfo[] regionsToMerge = new HRegionInfo[2];
- regionsToMerge[0] = tableRegions.get(0);
- regionsToMerge[1] = tableRegions.get(1);
-
- final int initCompletedTaskCount = countOfCompletedMergeTaskCount();
- long procId = procExec.submitProcedure(new DispatchMergingRegionsProcedure(
- procExec.getEnvironment(), tableName, regionsToMerge, true));
- ProcedureTestingUtility.waitProcedure(procExec, procId);
- ProcedureTestingUtility.assertProcNotFailed(procExec, procId);
-
- assertRegionCount(tableName, 2, 1, initCompletedTaskCount);
- }
-
- /**
- * This tests two concurrent region merges
- */
- @Test(timeout=60000)
- public void testMergeRegionsConcurrently() throws Exception {
- final TableName tableName = TableName.valueOf("testMergeRegionsConcurrently");
- final ProcedureExecutor<MasterProcedureEnv> procExec = getMasterProcedureExecutor();
-
- List<HRegionInfo> tableRegions = createTable(tableName, 4);
-
- HRegionInfo[] regionsToMerge1 = new HRegionInfo[2];
- HRegionInfo[] regionsToMerge2 = new HRegionInfo[2];
- regionsToMerge1[0] = tableRegions.get(0);
- regionsToMerge1[1] = tableRegions.get(1);
- regionsToMerge2[0] = tableRegions.get(2);
- regionsToMerge2[1] = tableRegions.get(3);
-
- final int initCompletedTaskCount = countOfCompletedMergeTaskCount();
- long procId1 = procExec.submitProcedure(new DispatchMergingRegionsProcedure(
- procExec.getEnvironment(), tableName, regionsToMerge1, true));
- long procId2 = procExec.submitProcedure(new DispatchMergingRegionsProcedure(
- procExec.getEnvironment(), tableName, regionsToMerge2, true));
- ProcedureTestingUtility.waitProcedure(procExec, procId1);
- ProcedureTestingUtility.waitProcedure(procExec, procId2);
- ProcedureTestingUtility.assertProcNotFailed(procExec, procId1);
- ProcedureTestingUtility.assertProcNotFailed(procExec, procId2);
- assertRegionCount(tableName, 2, 2, initCompletedTaskCount);
- }
-
- private void waitForCompletedMergeTask(int expectedTaskCount, int initCompletedTaskCount)
- throws IOException, InterruptedException {
- while (true) {
- long currentCompletedTaskCount = countOfCompletedMergeTaskCount() - initCompletedTaskCount;
- if (currentCompletedTaskCount == expectedTaskCount) {
- return;
- }
- LOG.info("There are " + (expectedTaskCount - currentCompletedTaskCount) +
- " merge requests are not completed, wait 100 ms");
- TimeUnit.MILLISECONDS.sleep(100);
- }
- }
-
- private static int countOfCompletedMergeTaskCount() {
- int completedTaskCount = 0;
- for (RegionServerThread server : UTIL.getMiniHBaseCluster().getRegionServerThreads()) {
- completedTaskCount += server.getRegionServer().getCompactSplitThread().getCompletedMergeTaskCount();
- }
- return completedTaskCount;
- }
-
- @Test(timeout=60000)
- public void testRecoveryAndDoubleExecution() throws Exception {
- final TableName tableName = TableName.valueOf("testRecoveryAndDoubleExecution");
- final ProcedureExecutor<MasterProcedureEnv> procExec = getMasterProcedureExecutor();
-
- List<HRegionInfo> tableRegions = createTable(tableName, 3);
-
- ProcedureTestingUtility.waitNoProcedureRunning(procExec);
- ProcedureTestingUtility.setKillAndToggleBeforeStoreUpdate(procExec, true);
-
- HRegionInfo[] regionsToMerge = new HRegionInfo[2];
- regionsToMerge[0] = tableRegions.get(0);
- regionsToMerge[1] = tableRegions.get(1);
-
- final int initCompletedTaskCount = countOfCompletedMergeTaskCount();
- long procId = procExec.submitProcedure(
- new DispatchMergingRegionsProcedure(
- procExec.getEnvironment(), tableName, regionsToMerge, true));
-
- // Restart the executor and execute the step twice
- int numberOfSteps = DispatchMergingRegionsState.values().length;
- MasterProcedureTestingUtility.testRecoveryAndDoubleExecution(procExec, procId, numberOfSteps);
- ProcedureTestingUtility.assertProcNotFailed(procExec, procId);
-
- assertRegionCount(tableName, 2, 1, initCompletedTaskCount);
- }
-
- @Test(timeout = 60000)
- public void testRollbackAndDoubleExecution() throws Exception {
- final TableName tableName = TableName.valueOf("testRollbackAndDoubleExecution");
- final ProcedureExecutor<MasterProcedureEnv> procExec = getMasterProcedureExecutor();
-
- List<HRegionInfo> tableRegions = createTable(tableName, 3);
-
- ProcedureTestingUtility.waitNoProcedureRunning(procExec);
- ProcedureTestingUtility.setKillAndToggleBeforeStoreUpdate(procExec, true);
-
- HRegionInfo[] regionsToMerge = new HRegionInfo[2];
- regionsToMerge[0] = tableRegions.get(0);
- regionsToMerge[1] = tableRegions.get(1);
-
- long procId = procExec.submitProcedure(
- new DispatchMergingRegionsProcedure(
- procExec.getEnvironment(), tableName, regionsToMerge, true));
-
- int numberOfSteps = DispatchMergingRegionsState.values().length - 3;
- MasterProcedureTestingUtility.testRollbackAndDoubleExecution(procExec, procId, numberOfSteps);
- }
-
- private List<HRegionInfo> createTable(final TableName tableName, final int nregions)
- throws Exception {
- HTableDescriptor desc = new HTableDescriptor(tableName);
- desc.addFamily(new HColumnDescriptor(FAMILY));
- byte[][] splitRows = new byte[nregions - 1][];
- for (int i = 0; i < splitRows.length; ++i) {
- splitRows[i] = Bytes.toBytes(String.format("%d", i));
- }
- admin.createTable(desc, splitRows);
- return assertRegionCount(tableName, nregions);
- }
-
- public List<HRegionInfo> assertRegionCount(final TableName tableName, final int nregions)
- throws Exception {
- UTIL.waitUntilNoRegionsInTransition();
- List<HRegionInfo> tableRegions = admin.getTableRegions(tableName);
- assertEquals(nregions, tableRegions.size());
- return tableRegions;
- }
-
- public List<HRegionInfo> assertRegionCount(final TableName tableName, final int nregions,
- int expectedTaskCount, int initCompletedTaskCount) throws Exception {
- waitForCompletedMergeTask(expectedTaskCount, initCompletedTaskCount);
- return assertRegionCount(tableName, nregions);
- }
-
- private ProcedureExecutor<MasterProcedureEnv> getMasterProcedureExecutor() {
- return UTIL.getHBaseCluster().getMaster().getMasterProcedureExecutor();
- }
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java
index a63fec6..c973471 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java
@@ -2646,55 +2646,6 @@ public class TestHRegion {
}
}
- // ////////////////////////////////////////////////////////////////////////////
- // Merge test
- // ////////////////////////////////////////////////////////////////////////////
- @Test
- public void testMerge() throws IOException {
- byte[][] families = { fam1, fam2, fam3 };
- Configuration hc = initSplit();
- // Setting up region
- this.region = initHRegion(tableName, method, hc, families);
- try {
- LOG.info("" + HBaseTestCase.addContent(region, fam3));
- region.flush(true);
- region.compactStores();
- byte[] splitRow = region.checkSplit();
- assertNotNull(splitRow);
- LOG.info("SplitRow: " + Bytes.toString(splitRow));
- HRegion[] subregions = splitRegion(region, splitRow);
- try {
- // Need to open the regions.
- for (int i = 0; i < subregions.length; i++) {
- HRegion.openHRegion(subregions[i], null);
- subregions[i].compactStores();
- }
- Path oldRegionPath = region.getRegionFileSystem().getRegionDir();
- Path oldRegion1 = subregions[0].getRegionFileSystem().getRegionDir();
- Path oldRegion2 = subregions[1].getRegionFileSystem().getRegionDir();
- long startTime = System.currentTimeMillis();
- region = HRegion.mergeAdjacent(subregions[0], subregions[1]);
- LOG.info("Merge regions elapsed time: "
- + ((System.currentTimeMillis() - startTime) / 1000.0));
- FILESYSTEM.delete(oldRegion1, true);
- FILESYSTEM.delete(oldRegion2, true);
- FILESYSTEM.delete(oldRegionPath, true);
- LOG.info("splitAndMerge completed.");
- } finally {
- for (int i = 0; i < subregions.length; i++) {
- try {
- HBaseTestingUtility.closeRegionAndWAL(subregions[i]);
- } catch (IOException e) {
- // Ignore.
- }
- }
- }
- } finally {
- HBaseTestingUtility.closeRegionAndWAL(this.region);
- this.region = null;
- }
- }
-
/**
* @param parent
* Region to split.
[7/7] hbase git commit: HBASE-17470 Remove merge region code from
region server (Stephen Yuan Jiang)
Posted by sy...@apache.org.
HBASE-17470 Remove merge region code from region server (Stephen Yuan Jiang)
Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/805d39fc
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/805d39fc
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/805d39fc
Branch: refs/heads/master
Commit: 805d39fca6b6be004d2c554cc5d4f76bf48bc01a
Parents: 287f95a
Author: Stephen Yuan Jiang <sy...@gmail.com>
Authored: Tue Jan 17 15:39:51 2017 -0800
Committer: Stephen Yuan Jiang <sy...@gmail.com>
Committed: Tue Jan 17 15:39:51 2017 -0800
----------------------------------------------------------------------
.../hbase/client/ConnectionImplementation.java | 7 -
.../hbase/shaded/protobuf/ProtobufUtil.java | 63 -
.../hbase/shaded/protobuf/RequestConverter.java | 18 -
.../shaded/protobuf/generated/AdminProtos.java | 1759 +-------
.../generated/MasterProcedureProtos.java | 1864 +-------
.../shaded/protobuf/generated/MasterProtos.java | 4160 ++++++------------
.../src/main/protobuf/Admin.proto | 19 -
.../src/main/protobuf/Master.proto | 19 -
.../src/main/protobuf/MasterProcedure.proto | 15 -
.../BaseMasterAndRegionObserver.java | 12 +
.../hbase/coprocessor/BaseMasterObserver.java | 12 +
.../hbase/coprocessor/MasterObserver.java | 8 +
.../org/apache/hadoop/hbase/master/HMaster.java | 50 -
.../hbase/master/MasterCoprocessorHost.java | 22 -
.../hadoop/hbase/master/MasterRpcServices.java | 46 -
.../hadoop/hbase/master/MasterServices.java | 17 -
.../hadoop/hbase/master/ServerManager.java | 31 -
.../DispatchMergingRegionsProcedure.java | 579 ---
.../hbase/regionserver/CompactSplitThread.java | 14 -
.../hadoop/hbase/regionserver/HRegion.java | 106 -
.../hbase/regionserver/RSRpcServices.java | 42 -
.../hbase/regionserver/RegionMergeRequest.java | 153 -
.../RegionMergeTransactionFactory.java | 76 -
.../RegionMergeTransactionImpl.java | 742 ----
.../org/apache/hadoop/hbase/util/HMerge.java | 348 --
.../org/apache/hadoop/hbase/util/Merge.java | 264 --
.../apache/hadoop/hbase/client/TestAdmin1.java | 50 +-
.../coprocessor/TestRegionServerObserver.java | 223 -
.../hbase/master/MockNoopMasterServices.java | 10 -
.../hadoop/hbase/master/MockRegionServer.java | 9 -
.../TestDispatchMergingRegionsProcedure.java | 264 --
.../hadoop/hbase/regionserver/TestHRegion.java | 49 -
.../TestRegionMergeTransaction.java | 485 --
.../hadoop/hbase/util/TestMergeTable.java | 172 -
.../apache/hadoop/hbase/util/TestMergeTool.java | 301 --
35 files changed, 1650 insertions(+), 10359 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionImplementation.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionImplementation.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionImplementation.java
index cfed9f6..adbc7f9 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionImplementation.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionImplementation.java
@@ -1312,13 +1312,6 @@ class ConnectionImplementation implements ClusterConnection, Closeable {
}
@Override
- public MasterProtos.DispatchMergingRegionsResponse dispatchMergingRegions(
- RpcController controller, MasterProtos.DispatchMergingRegionsRequest request)
- throws ServiceException {
- return stub.dispatchMergingRegions(controller, request);
- }
-
- @Override
public MasterProtos.MergeTableRegionsResponse mergeTableRegions(
RpcController controller, MasterProtos.MergeTableRegionsRequest request)
throws ServiceException {
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java
index 585a5f8..d862d5f 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java
@@ -121,7 +121,6 @@ import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetServerIn
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetServerInfoResponse;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileResponse;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.OpenRegionRequest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ServerInfo;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionRequest;
@@ -1950,46 +1949,6 @@ public final class ProtobufUtil {
}
}
- /**
- * A helper to merge regions using admin protocol. Send request to
- * regionserver.
- * @param admin
- * @param region_a
- * @param region_b
- * @param forcible true if do a compulsory merge, otherwise we will only merge
- * two adjacent regions
- * @param user effective user
- * @throws IOException
- */
- public static void mergeRegions(final RpcController controller,
- final AdminService.BlockingInterface admin,
- final HRegionInfo region_a, final HRegionInfo region_b,
- final boolean forcible, final User user) throws IOException {
- final MergeRegionsRequest request = ProtobufUtil.buildMergeRegionsRequest(
- region_a.getRegionName(), region_b.getRegionName(),forcible);
- if (user != null) {
- try {
- user.runAs(new PrivilegedExceptionAction<Void>() {
- @Override
- public Void run() throws Exception {
- admin.mergeRegions(controller, request);
- return null;
- }
- });
- } catch (InterruptedException ie) {
- InterruptedIOException iioe = new InterruptedIOException();
- iioe.initCause(ie);
- throw iioe;
- }
- } else {
- try {
- admin.mergeRegions(controller, request);
- } catch (ServiceException se) {
- throw ProtobufUtil.getRemoteException(se);
- }
- }
- }
-
// End helpers for Admin
/*
@@ -3225,28 +3184,6 @@ public final class ProtobufUtil {
}
/**
- * Create a MergeRegionsRequest for the given regions
- * @param regionA name of region a
- * @param regionB name of region b
- * @param forcible true if it is a compulsory merge
- * @return a MergeRegionsRequest
- */
- public static MergeRegionsRequest buildMergeRegionsRequest(
- final byte[] regionA, final byte[] regionB, final boolean forcible) {
- MergeRegionsRequest.Builder builder = MergeRegionsRequest.newBuilder();
- RegionSpecifier regionASpecifier = RequestConverter.buildRegionSpecifier(
- RegionSpecifierType.REGION_NAME, regionA);
- RegionSpecifier regionBSpecifier = RequestConverter.buildRegionSpecifier(
- RegionSpecifierType.REGION_NAME, regionB);
- builder.setRegionA(regionASpecifier);
- builder.setRegionB(regionBSpecifier);
- builder.setForcible(forcible);
- // send the master's wall clock time as well, so that the RS can refer to it
- builder.setMasterSystemTime(EnvironmentEdgeManager.currentTime());
- return builder.build();
- }
-
- /**
* Get a ServerName from the passed in data bytes.
* @param data Data with a serialize server name in it; can handle the old style
* servername where servername was host and port. Works too with data that
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java
index 4acb525..fd08d98 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java
@@ -84,7 +84,6 @@ import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTabl
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest;
@@ -1089,23 +1088,6 @@ public final class RequestConverter {
return builder.build();
}
- public static DispatchMergingRegionsRequest buildDispatchMergingRegionsRequest(
- final byte[] encodedNameOfRegionA,
- final byte[] encodedNameOfRegionB,
- final boolean forcible,
- final long nonceGroup,
- final long nonce) throws DeserializationException {
- DispatchMergingRegionsRequest.Builder builder = DispatchMergingRegionsRequest.newBuilder();
- builder.setRegionA(buildRegionSpecifier(
- RegionSpecifierType.ENCODED_REGION_NAME, encodedNameOfRegionA));
- builder.setRegionB(buildRegionSpecifier(
- RegionSpecifierType.ENCODED_REGION_NAME, encodedNameOfRegionB));
- builder.setForcible(forcible);
- builder.setNonceGroup(nonceGroup);
- builder.setNonce(nonce);
- return builder.build();
- }
-
public static MergeTableRegionsRequest buildMergeTableRegionsRequest(
final byte[][] encodedNameOfdaughaterRegions,
final boolean forcible,
[4/7] hbase git commit: HBASE-17470 Remove merge region code from
region server (Stephen Yuan Jiang)
Posted by sy...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProtos.java
index 8f293f3..6a737b8 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProtos.java
@@ -5507,35 +5507,33 @@ public final class MasterProtos {
}
- public interface DispatchMergingRegionsRequestOrBuilder extends
- // @@protoc_insertion_point(interface_extends:hbase.pb.DispatchMergingRegionsRequest)
+ public interface MergeTableRegionsRequestOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:hbase.pb.MergeTableRegionsRequest)
org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
- */
- boolean hasRegionA();
- /**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA();
+ java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier>
+ getRegionList();
/**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder();
-
+ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(int index);
/**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- boolean hasRegionB();
+ int getRegionCount();
/**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB();
+ java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>
+ getRegionOrBuilderList();
/**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder();
+ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder(
+ int index);
/**
* <code>optional bool forcible = 3 [default = false];</code>
@@ -5567,20 +5565,21 @@ public final class MasterProtos {
/**
* <pre>
**
- * Dispatch merging the specified regions.
+ * Merging the specified regions in a table.
* </pre>
*
- * Protobuf type {@code hbase.pb.DispatchMergingRegionsRequest}
+ * Protobuf type {@code hbase.pb.MergeTableRegionsRequest}
*/
- public static final class DispatchMergingRegionsRequest extends
+ public static final class MergeTableRegionsRequest extends
org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
- // @@protoc_insertion_point(message_implements:hbase.pb.DispatchMergingRegionsRequest)
- DispatchMergingRegionsRequestOrBuilder {
- // Use DispatchMergingRegionsRequest.newBuilder() to construct.
- private DispatchMergingRegionsRequest(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ // @@protoc_insertion_point(message_implements:hbase.pb.MergeTableRegionsRequest)
+ MergeTableRegionsRequestOrBuilder {
+ // Use MergeTableRegionsRequest.newBuilder() to construct.
+ private MergeTableRegionsRequest(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
- private DispatchMergingRegionsRequest() {
+ private MergeTableRegionsRequest() {
+ region_ = java.util.Collections.emptyList();
forcible_ = false;
nonceGroup_ = 0L;
nonce_ = 0L;
@@ -5591,7 +5590,7 @@ public final class MasterProtos {
getUnknownFields() {
return this.unknownFields;
}
- private DispatchMergingRegionsRequest(
+ private MergeTableRegionsRequest(
org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
@@ -5615,43 +5614,26 @@ public final class MasterProtos {
break;
}
case 10: {
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = null;
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
- subBuilder = regionA_.toBuilder();
- }
- regionA_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry);
- if (subBuilder != null) {
- subBuilder.mergeFrom(regionA_);
- regionA_ = subBuilder.buildPartial();
- }
- bitField0_ |= 0x00000001;
- break;
- }
- case 18: {
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = null;
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
- subBuilder = regionB_.toBuilder();
- }
- regionB_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry);
- if (subBuilder != null) {
- subBuilder.mergeFrom(regionB_);
- regionB_ = subBuilder.buildPartial();
+ if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
+ region_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier>();
+ mutable_bitField0_ |= 0x00000001;
}
- bitField0_ |= 0x00000002;
+ region_.add(
+ input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry));
break;
}
case 24: {
- bitField0_ |= 0x00000004;
+ bitField0_ |= 0x00000001;
forcible_ = input.readBool();
break;
}
case 32: {
- bitField0_ |= 0x00000008;
+ bitField0_ |= 0x00000002;
nonceGroup_ = input.readUInt64();
break;
}
case 40: {
- bitField0_ |= 0x00000010;
+ bitField0_ |= 0x00000004;
nonce_ = input.readUInt64();
break;
}
@@ -5663,63 +5645,59 @@ public final class MasterProtos {
throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
+ if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
+ region_ = java.util.Collections.unmodifiableList(region_);
+ }
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DispatchMergingRegionsRequest_descriptor;
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsRequest_descriptor;
}
protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DispatchMergingRegionsRequest_fieldAccessorTable
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsRequest_fieldAccessorTable
.ensureFieldAccessorsInitialized(
- org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest.Builder.class);
+ org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.Builder.class);
}
private int bitField0_;
- public static final int REGION_A_FIELD_NUMBER = 1;
- private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier regionA_;
- /**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
- */
- public boolean hasRegionA() {
- return ((bitField0_ & 0x00000001) == 0x00000001);
- }
+ public static final int REGION_FIELD_NUMBER = 1;
+ private java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier> region_;
/**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA() {
- return regionA_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionA_;
+ public java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier> getRegionList() {
+ return region_;
}
/**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder() {
- return regionA_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionA_;
+ public java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>
+ getRegionOrBuilderList() {
+ return region_;
}
-
- public static final int REGION_B_FIELD_NUMBER = 2;
- private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier regionB_;
/**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- public boolean hasRegionB() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
+ public int getRegionCount() {
+ return region_.size();
}
/**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB() {
- return regionB_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionB_;
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(int index) {
+ return region_.get(index);
}
/**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder() {
- return regionB_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionB_;
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder(
+ int index) {
+ return region_.get(index);
}
public static final int FORCIBLE_FIELD_NUMBER = 3;
@@ -5728,7 +5706,7 @@ public final class MasterProtos {
* <code>optional bool forcible = 3 [default = false];</code>
*/
public boolean hasForcible() {
- return ((bitField0_ & 0x00000004) == 0x00000004);
+ return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* <code>optional bool forcible = 3 [default = false];</code>
@@ -5743,7 +5721,7 @@ public final class MasterProtos {
* <code>optional uint64 nonce_group = 4 [default = 0];</code>
*/
public boolean hasNonceGroup() {
- return ((bitField0_ & 0x00000008) == 0x00000008);
+ return ((bitField0_ & 0x00000002) == 0x00000002);
}
/**
* <code>optional uint64 nonce_group = 4 [default = 0];</code>
@@ -5758,7 +5736,7 @@ public final class MasterProtos {
* <code>optional uint64 nonce = 5 [default = 0];</code>
*/
public boolean hasNonce() {
- return ((bitField0_ & 0x00000010) == 0x00000010);
+ return ((bitField0_ & 0x00000004) == 0x00000004);
}
/**
* <code>optional uint64 nonce = 5 [default = 0];</code>
@@ -5773,21 +5751,11 @@ public final class MasterProtos {
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
- if (!hasRegionA()) {
- memoizedIsInitialized = 0;
- return false;
- }
- if (!hasRegionB()) {
- memoizedIsInitialized = 0;
- return false;
- }
- if (!getRegionA().isInitialized()) {
- memoizedIsInitialized = 0;
- return false;
- }
- if (!getRegionB().isInitialized()) {
- memoizedIsInitialized = 0;
- return false;
+ for (int i = 0; i < getRegionCount(); i++) {
+ if (!getRegion(i).isInitialized()) {
+ memoizedIsInitialized = 0;
+ return false;
+ }
}
memoizedIsInitialized = 1;
return true;
@@ -5795,19 +5763,16 @@ public final class MasterProtos {
public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
- output.writeMessage(1, getRegionA());
- }
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
- output.writeMessage(2, getRegionB());
+ for (int i = 0; i < region_.size(); i++) {
+ output.writeMessage(1, region_.get(i));
}
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBool(3, forcible_);
}
- if (((bitField0_ & 0x00000008) == 0x00000008)) {
+ if (((bitField0_ & 0x00000002) == 0x00000002)) {
output.writeUInt64(4, nonceGroup_);
}
- if (((bitField0_ & 0x00000010) == 0x00000010)) {
+ if (((bitField0_ & 0x00000004) == 0x00000004)) {
output.writeUInt64(5, nonce_);
}
unknownFields.writeTo(output);
@@ -5818,23 +5783,19 @@ public final class MasterProtos {
if (size != -1) return size;
size = 0;
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
- .computeMessageSize(1, getRegionA());
- }
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ for (int i = 0; i < region_.size(); i++) {
size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
- .computeMessageSize(2, getRegionB());
+ .computeMessageSize(1, region_.get(i));
}
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBoolSize(3, forcible_);
}
- if (((bitField0_ & 0x00000008) == 0x00000008)) {
+ if (((bitField0_ & 0x00000002) == 0x00000002)) {
size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(4, nonceGroup_);
}
- if (((bitField0_ & 0x00000010) == 0x00000010)) {
+ if (((bitField0_ & 0x00000004) == 0x00000004)) {
size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(5, nonce_);
}
@@ -5849,22 +5810,14 @@ public final class MasterProtos {
if (obj == this) {
return true;
}
- if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest)) {
+ if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest)) {
return super.equals(obj);
}
- org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest) obj;
+ org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest) obj;
boolean result = true;
- result = result && (hasRegionA() == other.hasRegionA());
- if (hasRegionA()) {
- result = result && getRegionA()
- .equals(other.getRegionA());
- }
- result = result && (hasRegionB() == other.hasRegionB());
- if (hasRegionB()) {
- result = result && getRegionB()
- .equals(other.getRegionB());
- }
+ result = result && getRegionList()
+ .equals(other.getRegionList());
result = result && (hasForcible() == other.hasForcible());
if (hasForcible()) {
result = result && (getForcible()
@@ -5891,13 +5844,9 @@ public final class MasterProtos {
}
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
- if (hasRegionA()) {
- hash = (37 * hash) + REGION_A_FIELD_NUMBER;
- hash = (53 * hash) + getRegionA().hashCode();
- }
- if (hasRegionB()) {
- hash = (37 * hash) + REGION_B_FIELD_NUMBER;
- hash = (53 * hash) + getRegionB().hashCode();
+ if (getRegionCount() > 0) {
+ hash = (37 * hash) + REGION_FIELD_NUMBER;
+ hash = (53 * hash) + getRegionList().hashCode();
}
if (hasForcible()) {
hash = (37 * hash) + FORCIBLE_FIELD_NUMBER;
@@ -5919,58 +5868,58 @@ public final class MasterProtos {
return hash;
}
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseFrom(
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom(
org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseFrom(
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom(
org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseFrom(byte[] data)
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom(byte[] data)
throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseFrom(
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom(
byte[] data,
org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseFrom(java.io.InputStream input)
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseFrom(
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom(
java.io.InputStream input,
org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseDelimitedFrom(java.io.InputStream input)
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseDelimitedFrom(
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseFrom(
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom(
org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseFrom(
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom(
org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
@@ -5982,7 +5931,7 @@ public final class MasterProtos {
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
- public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest prototype) {
+ public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
public Builder toBuilder() {
@@ -5999,28 +5948,28 @@ public final class MasterProtos {
/**
* <pre>
**
- * Dispatch merging the specified regions.
+ * Merging the specified regions in a table.
* </pre>
*
- * Protobuf type {@code hbase.pb.DispatchMergingRegionsRequest}
+ * Protobuf type {@code hbase.pb.MergeTableRegionsRequest}
*/
public static final class Builder extends
org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
- // @@protoc_insertion_point(builder_implements:hbase.pb.DispatchMergingRegionsRequest)
- org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequestOrBuilder {
+ // @@protoc_insertion_point(builder_implements:hbase.pb.MergeTableRegionsRequest)
+ org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequestOrBuilder {
public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DispatchMergingRegionsRequest_descriptor;
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsRequest_descriptor;
}
protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DispatchMergingRegionsRequest_fieldAccessorTable
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsRequest_fieldAccessorTable
.ensureFieldAccessorsInitialized(
- org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest.Builder.class);
+ org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.Builder.class);
}
- // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest.newBuilder()
+ // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
@@ -6033,80 +5982,66 @@ public final class MasterProtos {
private void maybeForceBuilderInitialization() {
if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
- getRegionAFieldBuilder();
- getRegionBFieldBuilder();
+ getRegionFieldBuilder();
}
}
public Builder clear() {
super.clear();
- if (regionABuilder_ == null) {
- regionA_ = null;
- } else {
- regionABuilder_.clear();
- }
- bitField0_ = (bitField0_ & ~0x00000001);
- if (regionBBuilder_ == null) {
- regionB_ = null;
+ if (regionBuilder_ == null) {
+ region_ = java.util.Collections.emptyList();
+ bitField0_ = (bitField0_ & ~0x00000001);
} else {
- regionBBuilder_.clear();
+ regionBuilder_.clear();
}
- bitField0_ = (bitField0_ & ~0x00000002);
forcible_ = false;
- bitField0_ = (bitField0_ & ~0x00000004);
+ bitField0_ = (bitField0_ & ~0x00000002);
nonceGroup_ = 0L;
- bitField0_ = (bitField0_ & ~0x00000008);
+ bitField0_ = (bitField0_ & ~0x00000004);
nonce_ = 0L;
- bitField0_ = (bitField0_ & ~0x00000010);
+ bitField0_ = (bitField0_ & ~0x00000008);
return this;
}
public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DispatchMergingRegionsRequest_descriptor;
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsRequest_descriptor;
}
- public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest getDefaultInstanceForType() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest.getDefaultInstance();
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest getDefaultInstanceForType() {
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.getDefaultInstance();
}
- public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest build() {
- org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest result = buildPartial();
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest build() {
+ org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
- public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest buildPartial() {
- org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest(this);
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest buildPartial() {
+ org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest(this);
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
- if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
- to_bitField0_ |= 0x00000001;
- }
- if (regionABuilder_ == null) {
- result.regionA_ = regionA_;
+ if (regionBuilder_ == null) {
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ region_ = java.util.Collections.unmodifiableList(region_);
+ bitField0_ = (bitField0_ & ~0x00000001);
+ }
+ result.region_ = region_;
} else {
- result.regionA_ = regionABuilder_.build();
+ result.region_ = regionBuilder_.build();
}
if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
- to_bitField0_ |= 0x00000002;
- }
- if (regionBBuilder_ == null) {
- result.regionB_ = regionB_;
- } else {
- result.regionB_ = regionBBuilder_.build();
- }
- if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
- to_bitField0_ |= 0x00000004;
+ to_bitField0_ |= 0x00000001;
}
result.forcible_ = forcible_;
- if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
- to_bitField0_ |= 0x00000008;
+ if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
+ to_bitField0_ |= 0x00000002;
}
result.nonceGroup_ = nonceGroup_;
- if (((from_bitField0_ & 0x00000010) == 0x00000010)) {
- to_bitField0_ |= 0x00000010;
+ if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
+ to_bitField0_ |= 0x00000004;
}
result.nonce_ = nonce_;
result.bitField0_ = to_bitField0_;
@@ -6141,21 +6076,41 @@ public final class MasterProtos {
return (Builder) super.addRepeatedField(field, value);
}
public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
- if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest) {
- return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest)other);
+ if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest) {
+ return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest)other);
} else {
super.mergeFrom(other);
return this;
}
}
- public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest other) {
- if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest.getDefaultInstance()) return this;
- if (other.hasRegionA()) {
- mergeRegionA(other.getRegionA());
- }
- if (other.hasRegionB()) {
- mergeRegionB(other.getRegionB());
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest other) {
+ if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.getDefaultInstance()) return this;
+ if (regionBuilder_ == null) {
+ if (!other.region_.isEmpty()) {
+ if (region_.isEmpty()) {
+ region_ = other.region_;
+ bitField0_ = (bitField0_ & ~0x00000001);
+ } else {
+ ensureRegionIsMutable();
+ region_.addAll(other.region_);
+ }
+ onChanged();
+ }
+ } else {
+ if (!other.region_.isEmpty()) {
+ if (regionBuilder_.isEmpty()) {
+ regionBuilder_.dispose();
+ regionBuilder_ = null;
+ region_ = other.region_;
+ bitField0_ = (bitField0_ & ~0x00000001);
+ regionBuilder_ =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
+ getRegionFieldBuilder() : null;
+ } else {
+ regionBuilder_.addAllMessages(other.region_);
+ }
+ }
}
if (other.hasForcible()) {
setForcible(other.getForcible());
@@ -6172,17 +6127,10 @@ public final class MasterProtos {
}
public final boolean isInitialized() {
- if (!hasRegionA()) {
- return false;
- }
- if (!hasRegionB()) {
- return false;
- }
- if (!getRegionA().isInitialized()) {
- return false;
- }
- if (!getRegionB().isInitialized()) {
- return false;
+ for (int i = 0; i < getRegionCount(); i++) {
+ if (!getRegion(i).isInitialized()) {
+ return false;
+ }
}
return true;
}
@@ -6191,11 +6139,11 @@ public final class MasterProtos {
org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parsedMessage = null;
+ org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
- parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest) e.getUnfinishedMessage();
+ parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
if (parsedMessage != null) {
@@ -6206,240 +6154,244 @@ public final class MasterProtos {
}
private int bitField0_;
- private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier regionA_ = null;
- private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionABuilder_;
+ private java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier> region_ =
+ java.util.Collections.emptyList();
+ private void ensureRegionIsMutable() {
+ if (!((bitField0_ & 0x00000001) == 0x00000001)) {
+ region_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier>(region_);
+ bitField0_ |= 0x00000001;
+ }
+ }
+
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
+ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBuilder_;
+
/**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- public boolean hasRegionA() {
- return ((bitField0_ & 0x00000001) == 0x00000001);
+ public java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier> getRegionList() {
+ if (regionBuilder_ == null) {
+ return java.util.Collections.unmodifiableList(region_);
+ } else {
+ return regionBuilder_.getMessageList();
+ }
}
/**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA() {
- if (regionABuilder_ == null) {
- return regionA_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionA_;
+ public int getRegionCount() {
+ if (regionBuilder_ == null) {
+ return region_.size();
} else {
- return regionABuilder_.getMessage();
+ return regionBuilder_.getCount();
}
}
/**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- public Builder setRegionA(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) {
- if (regionABuilder_ == null) {
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(int index) {
+ if (regionBuilder_ == null) {
+ return region_.get(index);
+ } else {
+ return regionBuilder_.getMessage(index);
+ }
+ }
+ /**
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
+ */
+ public Builder setRegion(
+ int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) {
+ if (regionBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
- regionA_ = value;
+ ensureRegionIsMutable();
+ region_.set(index, value);
onChanged();
} else {
- regionABuilder_.setMessage(value);
+ regionBuilder_.setMessage(index, value);
}
- bitField0_ |= 0x00000001;
return this;
}
/**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- public Builder setRegionA(
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) {
- if (regionABuilder_ == null) {
- regionA_ = builderForValue.build();
+ public Builder setRegion(
+ int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) {
+ if (regionBuilder_ == null) {
+ ensureRegionIsMutable();
+ region_.set(index, builderForValue.build());
onChanged();
} else {
- regionABuilder_.setMessage(builderForValue.build());
+ regionBuilder_.setMessage(index, builderForValue.build());
}
- bitField0_ |= 0x00000001;
return this;
}
/**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- public Builder mergeRegionA(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) {
- if (regionABuilder_ == null) {
- if (((bitField0_ & 0x00000001) == 0x00000001) &&
- regionA_ != null &&
- regionA_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) {
- regionA_ =
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(regionA_).mergeFrom(value).buildPartial();
- } else {
- regionA_ = value;
+ public Builder addRegion(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) {
+ if (regionBuilder_ == null) {
+ if (value == null) {
+ throw new NullPointerException();
}
+ ensureRegionIsMutable();
+ region_.add(value);
onChanged();
} else {
- regionABuilder_.mergeFrom(value);
+ regionBuilder_.addMessage(value);
}
- bitField0_ |= 0x00000001;
return this;
}
/**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- public Builder clearRegionA() {
- if (regionABuilder_ == null) {
- regionA_ = null;
+ public Builder addRegion(
+ int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) {
+ if (regionBuilder_ == null) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ ensureRegionIsMutable();
+ region_.add(index, value);
onChanged();
} else {
- regionABuilder_.clear();
+ regionBuilder_.addMessage(index, value);
}
- bitField0_ = (bitField0_ & ~0x00000001);
return this;
}
/**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionABuilder() {
- bitField0_ |= 0x00000001;
- onChanged();
- return getRegionAFieldBuilder().getBuilder();
- }
- /**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder() {
- if (regionABuilder_ != null) {
- return regionABuilder_.getMessageOrBuilder();
+ public Builder addRegion(
+ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) {
+ if (regionBuilder_ == null) {
+ ensureRegionIsMutable();
+ region_.add(builderForValue.build());
+ onChanged();
} else {
- return regionA_ == null ?
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionA_;
+ regionBuilder_.addMessage(builderForValue.build());
}
+ return this;
}
/**
- * <code>required .hbase.pb.RegionSpecifier region_a = 1;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>
- getRegionAFieldBuilder() {
- if (regionABuilder_ == null) {
- regionABuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>(
- getRegionA(),
- getParentForChildren(),
- isClean());
- regionA_ = null;
+ public Builder addRegion(
+ int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) {
+ if (regionBuilder_ == null) {
+ ensureRegionIsMutable();
+ region_.add(index, builderForValue.build());
+ onChanged();
+ } else {
+ regionBuilder_.addMessage(index, builderForValue.build());
}
- return regionABuilder_;
- }
-
- private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier regionB_ = null;
- private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBBuilder_;
- /**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
- */
- public boolean hasRegionB() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
+ return this;
}
/**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB() {
- if (regionBBuilder_ == null) {
- return regionB_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionB_;
+ public Builder addAllRegion(
+ java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier> values) {
+ if (regionBuilder_ == null) {
+ ensureRegionIsMutable();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ values, region_);
+ onChanged();
} else {
- return regionBBuilder_.getMessage();
+ regionBuilder_.addAllMessages(values);
}
+ return this;
}
/**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- public Builder setRegionB(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) {
- if (regionBBuilder_ == null) {
- if (value == null) {
- throw new NullPointerException();
- }
- regionB_ = value;
+ public Builder clearRegion() {
+ if (regionBuilder_ == null) {
+ region_ = java.util.Collections.emptyList();
+ bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
} else {
- regionBBuilder_.setMessage(value);
+ regionBuilder_.clear();
}
- bitField0_ |= 0x00000002;
return this;
}
/**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- public Builder setRegionB(
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) {
- if (regionBBuilder_ == null) {
- regionB_ = builderForValue.build();
+ public Builder removeRegion(int index) {
+ if (regionBuilder_ == null) {
+ ensureRegionIsMutable();
+ region_.remove(index);
onChanged();
} else {
- regionBBuilder_.setMessage(builderForValue.build());
+ regionBuilder_.remove(index);
}
- bitField0_ |= 0x00000002;
return this;
}
/**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- public Builder mergeRegionB(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) {
- if (regionBBuilder_ == null) {
- if (((bitField0_ & 0x00000002) == 0x00000002) &&
- regionB_ != null &&
- regionB_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) {
- regionB_ =
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(regionB_).mergeFrom(value).buildPartial();
- } else {
- regionB_ = value;
- }
- onChanged();
- } else {
- regionBBuilder_.mergeFrom(value);
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionBuilder(
+ int index) {
+ return getRegionFieldBuilder().getBuilder(index);
+ }
+ /**
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
+ */
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder(
+ int index) {
+ if (regionBuilder_ == null) {
+ return region_.get(index); } else {
+ return regionBuilder_.getMessageOrBuilder(index);
}
- bitField0_ |= 0x00000002;
- return this;
}
/**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- public Builder clearRegionB() {
- if (regionBBuilder_ == null) {
- regionB_ = null;
- onChanged();
+ public java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>
+ getRegionOrBuilderList() {
+ if (regionBuilder_ != null) {
+ return regionBuilder_.getMessageOrBuilderList();
} else {
- regionBBuilder_.clear();
+ return java.util.Collections.unmodifiableList(region_);
}
- bitField0_ = (bitField0_ & ~0x00000002);
- return this;
}
/**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionBBuilder() {
- bitField0_ |= 0x00000002;
- onChanged();
- return getRegionBFieldBuilder().getBuilder();
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder addRegionBuilder() {
+ return getRegionFieldBuilder().addBuilder(
+ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance());
}
/**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder() {
- if (regionBBuilder_ != null) {
- return regionBBuilder_.getMessageOrBuilder();
- } else {
- return regionB_ == null ?
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionB_;
- }
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder addRegionBuilder(
+ int index) {
+ return getRegionFieldBuilder().addBuilder(
+ index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance());
}
/**
- * <code>required .hbase.pb.RegionSpecifier region_b = 2;</code>
+ * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
*/
- private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
+ public java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder>
+ getRegionBuilderList() {
+ return getRegionFieldBuilder().getBuilderList();
+ }
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>
- getRegionBFieldBuilder() {
- if (regionBBuilder_ == null) {
- regionBBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
+ getRegionFieldBuilder() {
+ if (regionBuilder_ == null) {
+ regionBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>(
- getRegionB(),
+ region_,
+ ((bitField0_ & 0x00000001) == 0x00000001),
getParentForChildren(),
isClean());
- regionB_ = null;
+ region_ = null;
}
- return regionBBuilder_;
+ return regionBuilder_;
}
private boolean forcible_ ;
@@ -6447,7 +6399,7 @@ public final class MasterProtos {
* <code>optional bool forcible = 3 [default = false];</code>
*/
public boolean hasForcible() {
- return ((bitField0_ & 0x00000004) == 0x00000004);
+ return ((bitField0_ & 0x00000002) == 0x00000002);
}
/**
* <code>optional bool forcible = 3 [default = false];</code>
@@ -6459,7 +6411,7 @@ public final class MasterProtos {
* <code>optional bool forcible = 3 [default = false];</code>
*/
public Builder setForcible(boolean value) {
- bitField0_ |= 0x00000004;
+ bitField0_ |= 0x00000002;
forcible_ = value;
onChanged();
return this;
@@ -6468,7 +6420,7 @@ public final class MasterProtos {
* <code>optional bool forcible = 3 [default = false];</code>
*/
public Builder clearForcible() {
- bitField0_ = (bitField0_ & ~0x00000004);
+ bitField0_ = (bitField0_ & ~0x00000002);
forcible_ = false;
onChanged();
return this;
@@ -6479,7 +6431,7 @@ public final class MasterProtos {
* <code>optional uint64 nonce_group = 4 [default = 0];</code>
*/
public boolean hasNonceGroup() {
- return ((bitField0_ & 0x00000008) == 0x00000008);
+ return ((bitField0_ & 0x00000004) == 0x00000004);
}
/**
* <code>optional uint64 nonce_group = 4 [default = 0];</code>
@@ -6491,7 +6443,7 @@ public final class MasterProtos {
* <code>optional uint64 nonce_group = 4 [default = 0];</code>
*/
public Builder setNonceGroup(long value) {
- bitField0_ |= 0x00000008;
+ bitField0_ |= 0x00000004;
nonceGroup_ = value;
onChanged();
return this;
@@ -6500,7 +6452,7 @@ public final class MasterProtos {
* <code>optional uint64 nonce_group = 4 [default = 0];</code>
*/
public Builder clearNonceGroup() {
- bitField0_ = (bitField0_ & ~0x00000008);
+ bitField0_ = (bitField0_ & ~0x00000004);
nonceGroup_ = 0L;
onChanged();
return this;
@@ -6511,7 +6463,7 @@ public final class MasterProtos {
* <code>optional uint64 nonce = 5 [default = 0];</code>
*/
public boolean hasNonce() {
- return ((bitField0_ & 0x00000010) == 0x00000010);
+ return ((bitField0_ & 0x00000008) == 0x00000008);
}
/**
* <code>optional uint64 nonce = 5 [default = 0];</code>
@@ -6523,7 +6475,7 @@ public final class MasterProtos {
* <code>optional uint64 nonce = 5 [default = 0];</code>
*/
public Builder setNonce(long value) {
- bitField0_ |= 0x00000010;
+ bitField0_ |= 0x00000008;
nonce_ = value;
onChanged();
return this;
@@ -6532,7 +6484,7 @@ public final class MasterProtos {
* <code>optional uint64 nonce = 5 [default = 0];</code>
*/
public Builder clearNonce() {
- bitField0_ = (bitField0_ & ~0x00000010);
+ bitField0_ = (bitField0_ & ~0x00000008);
nonce_ = 0L;
onChanged();
return this;
@@ -6548,46 +6500,46 @@ public final class MasterProtos {
}
- // @@protoc_insertion_point(builder_scope:hbase.pb.DispatchMergingRegionsRequest)
+ // @@protoc_insertion_point(builder_scope:hbase.pb.MergeTableRegionsRequest)
}
- // @@protoc_insertion_point(class_scope:hbase.pb.DispatchMergingRegionsRequest)
- private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest DEFAULT_INSTANCE;
+ // @@protoc_insertion_point(class_scope:hbase.pb.MergeTableRegionsRequest)
+ private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest DEFAULT_INSTANCE;
static {
- DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest();
+ DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest();
}
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest getDefaultInstance() {
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest getDefaultInstance() {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<DispatchMergingRegionsRequest>
- PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<DispatchMergingRegionsRequest>() {
- public DispatchMergingRegionsRequest parsePartialFrom(
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<MergeTableRegionsRequest>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<MergeTableRegionsRequest>() {
+ public MergeTableRegionsRequest parsePartialFrom(
org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return new DispatchMergingRegionsRequest(input, extensionRegistry);
+ return new MergeTableRegionsRequest(input, extensionRegistry);
}
};
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<DispatchMergingRegionsRequest> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<MergeTableRegionsRequest> parser() {
return PARSER;
}
@java.lang.Override
- public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<DispatchMergingRegionsRequest> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<MergeTableRegionsRequest> getParserForType() {
return PARSER;
}
- public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest getDefaultInstanceForType() {
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
- public interface DispatchMergingRegionsResponseOrBuilder extends
- // @@protoc_insertion_point(interface_extends:hbase.pb.DispatchMergingRegionsResponse)
+ public interface MergeTableRegionsResponseOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:hbase.pb.MergeTableRegionsResponse)
org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
@@ -6600,17 +6552,17 @@ public final class MasterProtos {
long getProcId();
}
/**
- * Protobuf type {@code hbase.pb.DispatchMergingRegionsResponse}
+ * Protobuf type {@code hbase.pb.MergeTableRegionsResponse}
*/
- public static final class DispatchMergingRegionsResponse extends
+ public static final class MergeTableRegionsResponse extends
org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
- // @@protoc_insertion_point(message_implements:hbase.pb.DispatchMergingRegionsResponse)
- DispatchMergingRegionsResponseOrBuilder {
- // Use DispatchMergingRegionsResponse.newBuilder() to construct.
- private DispatchMergingRegionsResponse(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ // @@protoc_insertion_point(message_implements:hbase.pb.MergeTableRegionsResponse)
+ MergeTableRegionsResponseOrBuilder {
+ // Use MergeTableRegionsResponse.newBuilder() to construct.
+ private MergeTableRegionsResponse(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
- private DispatchMergingRegionsResponse() {
+ private MergeTableRegionsResponse() {
procId_ = 0L;
}
@@ -6619,7 +6571,7 @@ public final class MasterProtos {
getUnknownFields() {
return this.unknownFields;
}
- private DispatchMergingRegionsResponse(
+ private MergeTableRegionsResponse(
org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
@@ -6661,14 +6613,14 @@ public final class MasterProtos {
}
public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DispatchMergingRegionsResponse_descriptor;
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsResponse_descriptor;
}
protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DispatchMergingRegionsResponse_fieldAccessorTable
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsResponse_fieldAccessorTable
.ensureFieldAccessorsInitialized(
- org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.Builder.class);
+ org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse.Builder.class);
}
private int bitField0_;
@@ -6725,1508 +6677,10 @@ public final class MasterProtos {
if (obj == this) {
return true;
}
- if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse)) {
+ if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse)) {
return super.equals(obj);
}
- org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse) obj;
-
- boolean result = true;
- result = result && (hasProcId() == other.hasProcId());
- if (hasProcId()) {
- result = result && (getProcId()
- == other.getProcId());
- }
- result = result && unknownFields.equals(other.unknownFields);
- return result;
- }
-
- @java.lang.Override
- public int hashCode() {
- if (memoizedHashCode != 0) {
- return memoizedHashCode;
- }
- int hash = 41;
- hash = (19 * hash) + getDescriptorForType().hashCode();
- if (hasProcId()) {
- hash = (37 * hash) + PROC_ID_FIELD_NUMBER;
- hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
- getProcId());
- }
- hash = (29 * hash) + unknownFields.hashCode();
- memoizedHashCode = hash;
- return hash;
- }
-
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseFrom(byte[] data)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseFrom(
- byte[] data,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseFrom(java.io.InputStream input)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseWithIOException(PARSER, input);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseFrom(
- java.io.InputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseWithIOException(PARSER, input, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseDelimitedFrom(java.io.InputStream input)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseDelimitedWithIOException(PARSER, input);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseDelimitedFrom(
- java.io.InputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseWithIOException(PARSER, input);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseWithIOException(PARSER, input, extensionRegistry);
- }
-
- public Builder newBuilderForType() { return newBuilder(); }
- public static Builder newBuilder() {
- return DEFAULT_INSTANCE.toBuilder();
- }
- public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse prototype) {
- return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
- }
- public Builder toBuilder() {
- return this == DEFAULT_INSTANCE
- ? new Builder() : new Builder().mergeFrom(this);
- }
-
- @java.lang.Override
- protected Builder newBuilderForType(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
- Builder builder = new Builder(parent);
- return builder;
- }
- /**
- * Protobuf type {@code hbase.pb.DispatchMergingRegionsResponse}
- */
- public static final class Builder extends
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
- // @@protoc_insertion_point(builder_implements:hbase.pb.DispatchMergingRegionsResponse)
- org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponseOrBuilder {
- public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
- getDescriptor() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DispatchMergingRegionsResponse_descriptor;
- }
-
- protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
- internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DispatchMergingRegionsResponse_fieldAccessorTable
- .ensureFieldAccessorsInitialized(
- org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.Builder.class);
- }
-
- // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.newBuilder()
- private Builder() {
- maybeForceBuilderInitialization();
- }
-
- private Builder(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
- super(parent);
- maybeForceBuilderInitialization();
- }
- private void maybeForceBuilderInitialization() {
- if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .alwaysUseFieldBuilders) {
- }
- }
- public Builder clear() {
- super.clear();
- procId_ = 0L;
- bitField0_ = (bitField0_ & ~0x00000001);
- return this;
- }
-
- public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
- getDescriptorForType() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DispatchMergingRegionsResponse_descriptor;
- }
-
- public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse getDefaultInstanceForType() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.getDefaultInstance();
- }
-
- public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse build() {
- org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse result = buildPartial();
- if (!result.isInitialized()) {
- throw newUninitializedMessageException(result);
- }
- return result;
- }
-
- public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse buildPartial() {
- org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse(this);
- int from_bitField0_ = bitField0_;
- int to_bitField0_ = 0;
- if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
- to_bitField0_ |= 0x00000001;
- }
- result.procId_ = procId_;
- result.bitField0_ = to_bitField0_;
- onBuilt();
- return result;
- }
-
- public Builder clone() {
- return (Builder) super.clone();
- }
- public Builder setField(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
- Object value) {
- return (Builder) super.setField(field, value);
- }
- public Builder clearField(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
- return (Builder) super.clearField(field);
- }
- public Builder clearOneof(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
- return (Builder) super.clearOneof(oneof);
- }
- public Builder setRepeatedField(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
- int index, Object value) {
- return (Builder) super.setRepeatedField(field, index, value);
- }
- public Builder addRepeatedField(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
- Object value) {
- return (Builder) super.addRepeatedField(field, value);
- }
- public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
- if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse) {
- return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse)other);
- } else {
- super.mergeFrom(other);
- return this;
- }
- }
-
- public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse other) {
- if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.getDefaultInstance()) return this;
- if (other.hasProcId()) {
- setProcId(other.getProcId());
- }
- this.mergeUnknownFields(other.unknownFields);
- onChanged();
- return this;
- }
-
- public final boolean isInitialized() {
- return true;
- }
-
- public Builder mergeFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parsedMessage = null;
- try {
- parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
- parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse) e.getUnfinishedMessage();
- throw e.unwrapIOException();
- } finally {
- if (parsedMessage != null) {
- mergeFrom(parsedMessage);
- }
- }
- return this;
- }
- private int bitField0_;
-
- private long procId_ ;
- /**
- * <code>optional uint64 proc_id = 1;</code>
- */
- public boolean hasProcId() {
- return ((bitField0_ & 0x00000001) == 0x00000001);
- }
- /**
- * <code>optional uint64 proc_id = 1;</code>
- */
- public long getProcId() {
- return procId_;
- }
- /**
- * <code>optional uint64 proc_id = 1;</code>
- */
- public Builder setProcId(long value) {
- bitField0_ |= 0x00000001;
- procId_ = value;
- onChanged();
- return this;
- }
- /**
- * <code>optional uint64 proc_id = 1;</code>
- */
- public Builder clearProcId() {
- bitField0_ = (bitField0_ & ~0x00000001);
- procId_ = 0L;
- onChanged();
- return this;
- }
- public final Builder setUnknownFields(
- final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
- return super.setUnknownFields(unknownFields);
- }
-
- public final Builder mergeUnknownFields(
- final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
- return super.mergeUnknownFields(unknownFields);
- }
-
-
- // @@protoc_insertion_point(builder_scope:hbase.pb.DispatchMergingRegionsResponse)
- }
-
- // @@protoc_insertion_point(class_scope:hbase.pb.DispatchMergingRegionsResponse)
- private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse DEFAULT_INSTANCE;
- static {
- DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse();
- }
-
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse getDefaultInstance() {
- return DEFAULT_INSTANCE;
- }
-
- @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<DispatchMergingRegionsResponse>
- PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<DispatchMergingRegionsResponse>() {
- public DispatchMergingRegionsResponse parsePartialFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return new DispatchMergingRegionsResponse(input, extensionRegistry);
- }
- };
-
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<DispatchMergingRegionsResponse> parser() {
- return PARSER;
- }
-
- @java.lang.Override
- public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<DispatchMergingRegionsResponse> getParserForType() {
- return PARSER;
- }
-
- public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse getDefaultInstanceForType() {
- return DEFAULT_INSTANCE;
- }
-
- }
-
- public interface MergeTableRegionsRequestOrBuilder extends
- // @@protoc_insertion_point(interface_extends:hbase.pb.MergeTableRegionsRequest)
- org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
- /**
- * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
- */
- java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier>
- getRegionList();
- /**
- * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
- */
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(int index);
- /**
- * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
- */
- int getRegionCount();
- /**
- * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
- */
- java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>
- getRegionOrBuilderList();
- /**
- * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
- */
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder(
- int index);
-
- /**
- * <code>optional bool forcible = 3 [default = false];</code>
- */
- boolean hasForcible();
- /**
- * <code>optional bool forcible = 3 [default = false];</code>
- */
- boolean getForcible();
-
- /**
- * <code>optional uint64 nonce_group = 4 [default = 0];</code>
- */
- boolean hasNonceGroup();
- /**
- * <code>optional uint64 nonce_group = 4 [default = 0];</code>
- */
- long getNonceGroup();
-
- /**
- * <code>optional uint64 nonce = 5 [default = 0];</code>
- */
- boolean hasNonce();
- /**
- * <code>optional uint64 nonce = 5 [default = 0];</code>
- */
- long getNonce();
- }
- /**
- * <pre>
- **
- * Merging the specified regions in a table.
- * </pre>
- *
- * Protobuf type {@code hbase.pb.MergeTableRegionsRequest}
- */
- public static final class MergeTableRegionsRequest extends
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
- // @@protoc_insertion_point(message_implements:hbase.pb.MergeTableRegionsRequest)
- MergeTableRegionsRequestOrBuilder {
- // Use MergeTableRegionsRequest.newBuilder() to construct.
- private MergeTableRegionsRequest(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
- super(builder);
- }
- private MergeTableRegionsRequest() {
- region_ = java.util.Collections.emptyList();
- forcible_ = false;
- nonceGroup_ = 0L;
- nonce_ = 0L;
- }
-
- @java.lang.Override
- public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
- return this.unknownFields;
- }
- private MergeTableRegionsRequest(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- this();
- int mutable_bitField0_ = 0;
- org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
- try {
- boolean done = false;
- while (!done) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
- done = true;
- break;
- default: {
- if (!parseUnknownField(input, unknownFields,
- extensionRegistry, tag)) {
- done = true;
- }
- break;
- }
- case 10: {
- if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
- region_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier>();
- mutable_bitField0_ |= 0x00000001;
- }
- region_.add(
- input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry));
- break;
- }
- case 24: {
- bitField0_ |= 0x00000001;
- forcible_ = input.readBool();
- break;
- }
- case 32: {
- bitField0_ |= 0x00000002;
- nonceGroup_ = input.readUInt64();
- break;
- }
- case 40: {
- bitField0_ |= 0x00000004;
- nonce_ = input.readUInt64();
- break;
- }
- }
- }
- } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
- throw e.setUnfinishedMessage(this);
- } catch (java.io.IOException e) {
- throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
- e).setUnfinishedMessage(this);
- } finally {
- if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
- region_ = java.util.Collections.unmodifiableList(region_);
- }
- this.unknownFields = unknownFields.build();
- makeExtensionsImmutable();
- }
- }
- public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
- getDescriptor() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsRequest_descriptor;
- }
-
- protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
- internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsRequest_fieldAccessorTable
- .ensureFieldAccessorsInitialized(
- org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.Builder.class);
- }
-
- private int bitField0_;
- public static final int REGION_FIELD_NUMBER = 1;
- private java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier> region_;
- /**
- * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
- */
- public java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier> getRegionList() {
- return region_;
- }
- /**
- * <code>repeated .hbase.pb.RegionSpecifier region = 1;</code>
- */
- public java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>
- getRegionOrBuilderList() {
- return region
<TRUNCATED>
[5/7] hbase git commit: HBASE-17470 Remove merge region code from
region server (Stephen Yuan Jiang)
Posted by sy...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProcedureProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProcedureProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProcedureProtos.java
index 8f5992c..f4f6a8c 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProcedureProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProcedureProtos.java
@@ -1716,123 +1716,6 @@ public final class MasterProcedureProtos {
}
/**
- * Protobuf enum {@code hbase.pb.DispatchMergingRegionsState}
- */
- public enum DispatchMergingRegionsState
- implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
- /**
- * <code>DISPATCH_MERGING_REGIONS_PREPARE = 1;</code>
- */
- DISPATCH_MERGING_REGIONS_PREPARE(1),
- /**
- * <code>DISPATCH_MERGING_REGIONS_PRE_OPERATION = 2;</code>
- */
- DISPATCH_MERGING_REGIONS_PRE_OPERATION(2),
- /**
- * <code>DISPATCH_MERGING_REGIONS_MOVE_REGION_TO_SAME_RS = 3;</code>
- */
- DISPATCH_MERGING_REGIONS_MOVE_REGION_TO_SAME_RS(3),
- /**
- * <code>DISPATCH_MERGING_REGIONS_DO_MERGE_IN_RS = 4;</code>
- */
- DISPATCH_MERGING_REGIONS_DO_MERGE_IN_RS(4),
- /**
- * <code>DISPATCH_MERGING_REGIONS_POST_OPERATION = 5;</code>
- */
- DISPATCH_MERGING_REGIONS_POST_OPERATION(5),
- ;
-
- /**
- * <code>DISPATCH_MERGING_REGIONS_PREPARE = 1;</code>
- */
- public static final int DISPATCH_MERGING_REGIONS_PREPARE_VALUE = 1;
- /**
- * <code>DISPATCH_MERGING_REGIONS_PRE_OPERATION = 2;</code>
- */
- public static final int DISPATCH_MERGING_REGIONS_PRE_OPERATION_VALUE = 2;
- /**
- * <code>DISPATCH_MERGING_REGIONS_MOVE_REGION_TO_SAME_RS = 3;</code>
- */
- public static final int DISPATCH_MERGING_REGIONS_MOVE_REGION_TO_SAME_RS_VALUE = 3;
- /**
- * <code>DISPATCH_MERGING_REGIONS_DO_MERGE_IN_RS = 4;</code>
- */
- public static final int DISPATCH_MERGING_REGIONS_DO_MERGE_IN_RS_VALUE = 4;
- /**
- * <code>DISPATCH_MERGING_REGIONS_POST_OPERATION = 5;</code>
- */
- public static final int DISPATCH_MERGING_REGIONS_POST_OPERATION_VALUE = 5;
-
-
- public final int getNumber() {
- return value;
- }
-
- /**
- * @deprecated Use {@link #forNumber(int)} instead.
- */
- @java.lang.Deprecated
- public static DispatchMergingRegionsState valueOf(int value) {
- return forNumber(value);
- }
-
- public static DispatchMergingRegionsState forNumber(int value) {
- switch (value) {
- case 1: return DISPATCH_MERGING_REGIONS_PREPARE;
- case 2: return DISPATCH_MERGING_REGIONS_PRE_OPERATION;
- case 3: return DISPATCH_MERGING_REGIONS_MOVE_REGION_TO_SAME_RS;
- case 4: return DISPATCH_MERGING_REGIONS_DO_MERGE_IN_RS;
- case 5: return DISPATCH_MERGING_REGIONS_POST_OPERATION;
- default: return null;
- }
- }
-
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<DispatchMergingRegionsState>
- internalGetValueMap() {
- return internalValueMap;
- }
- private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
- DispatchMergingRegionsState> internalValueMap =
- new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<DispatchMergingRegionsState>() {
- public DispatchMergingRegionsState findValueByNumber(int number) {
- return DispatchMergingRegionsState.forNumber(number);
- }
- };
-
- public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
- getValueDescriptor() {
- return getDescriptor().getValues().get(ordinal());
- }
- public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
- getDescriptorForType() {
- return getDescriptor();
- }
- public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
- getDescriptor() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.getDescriptor().getEnumTypes().get(14);
- }
-
- private static final DispatchMergingRegionsState[] VALUES = values();
-
- public static DispatchMergingRegionsState valueOf(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
- if (desc.getType() != getDescriptor()) {
- throw new java.lang.IllegalArgumentException(
- "EnumValueDescriptor is not for this type.");
- }
- return VALUES[desc.getIndex()];
- }
-
- private final int value;
-
- private DispatchMergingRegionsState(int value) {
- this.value = value;
- }
-
- // @@protoc_insertion_point(enum_scope:hbase.pb.DispatchMergingRegionsState)
- }
-
- /**
* Protobuf enum {@code hbase.pb.MergeTableRegionsState}
*/
public enum MergeTableRegionsState
@@ -1980,7 +1863,7 @@ public final class MasterProcedureProtos {
}
public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.getDescriptor().getEnumTypes().get(15);
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.getDescriptor().getEnumTypes().get(14);
}
private static final MergeTableRegionsState[] VALUES = values();
@@ -2142,7 +2025,7 @@ public final class MasterProcedureProtos {
}
public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.getDescriptor().getEnumTypes().get(16);
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.getDescriptor().getEnumTypes().get(15);
}
private static final SplitTableRegionState[] VALUES = values();
@@ -2303,7 +2186,7 @@ public final class MasterProcedureProtos {
}
public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.getDescriptor().getEnumTypes().get(17);
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.getDescriptor().getEnumTypes().get(16);
}
private static final ServerCrashState[] VALUES = values();
@@ -20009,8 +19892,8 @@ public final class MasterProcedureProtos {
}
- public interface DispatchMergingRegionsStateDataOrBuilder extends
- // @@protoc_insertion_point(interface_extends:hbase.pb.DispatchMergingRegionsStateData)
+ public interface MergeTableRegionsStateDataOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:hbase.pb.MergeTableRegionsStateData)
org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
@@ -20027,63 +19910,63 @@ public final class MasterProcedureProtos {
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationOrBuilder getUserInfoOrBuilder();
/**
- * <code>required .hbase.pb.TableName table_name = 2;</code>
- */
- boolean hasTableName();
- /**
- * <code>required .hbase.pb.TableName table_name = 2;</code>
- */
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName();
- /**
- * <code>required .hbase.pb.TableName table_name = 2;</code>
- */
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder();
-
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
+ * <code>repeated .hbase.pb.RegionInfo region_info = 2;</code>
*/
java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo>
getRegionInfoList();
/**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
+ * <code>repeated .hbase.pb.RegionInfo region_info = 2;</code>
*/
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo(int index);
/**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
+ * <code>repeated .hbase.pb.RegionInfo region_info = 2;</code>
*/
int getRegionInfoCount();
/**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
+ * <code>repeated .hbase.pb.RegionInfo region_info = 2;</code>
*/
java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder>
getRegionInfoOrBuilderList();
/**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
+ * <code>repeated .hbase.pb.RegionInfo region_info = 2;</code>
*/
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionInfoOrBuilder(
int index);
/**
- * <code>optional bool forcible = 4;</code>
+ * <code>required .hbase.pb.RegionInfo merged_region_info = 3;</code>
+ */
+ boolean hasMergedRegionInfo();
+ /**
+ * <code>required .hbase.pb.RegionInfo merged_region_info = 3;</code>
+ */
+ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo getMergedRegionInfo();
+ /**
+ * <code>required .hbase.pb.RegionInfo merged_region_info = 3;</code>
+ */
+ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getMergedRegionInfoOrBuilder();
+
+ /**
+ * <code>optional bool forcible = 4 [default = false];</code>
*/
boolean hasForcible();
/**
- * <code>optional bool forcible = 4;</code>
+ * <code>optional bool forcible = 4 [default = false];</code>
*/
boolean getForcible();
}
/**
- * Protobuf type {@code hbase.pb.DispatchMergingRegionsStateData}
+ * Protobuf type {@code hbase.pb.MergeTableRegionsStateData}
*/
- public static final class DispatchMergingRegionsStateData extends
+ public static final class MergeTableRegionsStateData extends
org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
- // @@protoc_insertion_point(message_implements:hbase.pb.DispatchMergingRegionsStateData)
- DispatchMergingRegionsStateDataOrBuilder {
- // Use DispatchMergingRegionsStateData.newBuilder() to construct.
- private DispatchMergingRegionsStateData(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ // @@protoc_insertion_point(message_implements:hbase.pb.MergeTableRegionsStateData)
+ MergeTableRegionsStateDataOrBuilder {
+ // Use MergeTableRegionsStateData.newBuilder() to construct.
+ private MergeTableRegionsStateData(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
- private DispatchMergingRegionsStateData() {
+ private MergeTableRegionsStateData() {
regionInfo_ = java.util.Collections.emptyList();
forcible_ = false;
}
@@ -20093,7 +19976,7 @@ public final class MasterProcedureProtos {
getUnknownFields() {
return this.unknownFields;
}
- private DispatchMergingRegionsStateData(
+ private MergeTableRegionsStateData(
org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
@@ -20130,25 +20013,25 @@ public final class MasterProcedureProtos {
break;
}
case 18: {
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null;
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
- subBuilder = tableName_.toBuilder();
- }
- tableName_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry);
- if (subBuilder != null) {
- subBuilder.mergeFrom(tableName_);
- tableName_ = subBuilder.buildPartial();
+ if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
+ regionInfo_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo>();
+ mutable_bitField0_ |= 0x00000002;
}
- bitField0_ |= 0x00000002;
+ regionInfo_.add(
+ input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.PARSER, extensionRegistry));
break;
}
case 26: {
- if (!((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
- regionInfo_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo>();
- mutable_bitField0_ |= 0x00000004;
+ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder subBuilder = null;
+ if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ subBuilder = mergedRegionInfo_.toBuilder();
}
- regionInfo_.add(
- input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.PARSER, extensionRegistry));
+ mergedRegionInfo_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.PARSER, extensionRegistry);
+ if (subBuilder != null) {
+ subBuilder.mergeFrom(mergedRegionInfo_);
+ mergedRegionInfo_ = subBuilder.buildPartial();
+ }
+ bitField0_ |= 0x00000002;
break;
}
case 32: {
@@ -20164,7 +20047,7 @@ public final class MasterProcedureProtos {
throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
- if (((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
+ if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
regionInfo_ = java.util.Collections.unmodifiableList(regionInfo_);
}
this.unknownFields = unknownFields.build();
@@ -20173,14 +20056,14 @@ public final class MasterProcedureProtos {
}
public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.internal_static_hbase_pb_DispatchMergingRegionsStateData_descriptor;
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.internal_static_hbase_pb_MergeTableRegionsStateData_descriptor;
}
protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.internal_static_hbase_pb_DispatchMergingRegionsStateData_fieldAccessorTable
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.internal_static_hbase_pb_MergeTableRegionsStateData_fieldAccessorTable
.ensureFieldAccessorsInitialized(
- org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData.Builder.class);
+ org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.MergeTableRegionsStateData.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.MergeTableRegionsStateData.Builder.class);
}
private int bitField0_;
@@ -20205,72 +20088,72 @@ public final class MasterProcedureProtos {
return userInfo_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance() : userInfo_;
}
- public static final int TABLE_NAME_FIELD_NUMBER = 2;
- private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_;
- /**
- * <code>required .hbase.pb.TableName table_name = 2;</code>
- */
- public boolean hasTableName() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
- }
- /**
- * <code>required .hbase.pb.TableName table_name = 2;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() {
- return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
- }
- /**
- * <code>required .hbase.pb.TableName table_name = 2;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
- return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
- }
-
- public static final int REGION_INFO_FIELD_NUMBER = 3;
+ public static final int REGION_INFO_FIELD_NUMBER = 2;
private java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo> regionInfo_;
/**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
+ * <code>repeated .hbase.pb.RegionInfo region_info = 2;</code>
*/
public java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo> getRegionInfoList() {
return regionInfo_;
}
/**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
+ * <code>repeated .hbase.pb.RegionInfo region_info = 2;</code>
*/
public java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder>
getRegionInfoOrBuilderList() {
return regionInfo_;
}
/**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
+ * <code>repeated .hbase.pb.RegionInfo region_info = 2;</code>
*/
public int getRegionInfoCount() {
return regionInfo_.size();
}
/**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
+ * <code>repeated .hbase.pb.RegionInfo region_info = 2;</code>
*/
public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo(int index) {
return regionInfo_.get(index);
}
/**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
+ * <code>repeated .hbase.pb.RegionInfo region_info = 2;</code>
*/
public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionInfoOrBuilder(
int index) {
return regionInfo_.get(index);
}
+ public static final int MERGED_REGION_INFO_FIELD_NUMBER = 3;
+ private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo mergedRegionInfo_;
+ /**
+ * <code>required .hbase.pb.RegionInfo merged_region_info = 3;</code>
+ */
+ public boolean hasMergedRegionInfo() {
+ return ((bitField0_ & 0x00000002) == 0x00000002);
+ }
+ /**
+ * <code>required .hbase.pb.RegionInfo merged_region_info = 3;</code>
+ */
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo getMergedRegionInfo() {
+ return mergedRegionInfo_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance() : mergedRegionInfo_;
+ }
+ /**
+ * <code>required .hbase.pb.RegionInfo merged_region_info = 3;</code>
+ */
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getMergedRegionInfoOrBuilder() {
+ return mergedRegionInfo_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance() : mergedRegionInfo_;
+ }
+
public static final int FORCIBLE_FIELD_NUMBER = 4;
private boolean forcible_;
/**
- * <code>optional bool forcible = 4;</code>
+ * <code>optional bool forcible = 4 [default = false];</code>
*/
public boolean hasForcible() {
return ((bitField0_ & 0x00000004) == 0x00000004);
}
/**
- * <code>optional bool forcible = 4;</code>
+ * <code>optional bool forcible = 4 [default = false];</code>
*/
public boolean getForcible() {
return forcible_;
@@ -20286,7 +20169,7 @@ public final class MasterProcedureProtos {
memoizedIsInitialized = 0;
return false;
}
- if (!hasTableName()) {
+ if (!hasMergedRegionInfo()) {
memoizedIsInitialized = 0;
return false;
}
@@ -20294,16 +20177,16 @@ public final class MasterProcedureProtos {
memoizedIsInitialized = 0;
return false;
}
- if (!getTableName().isInitialized()) {
- memoizedIsInitialized = 0;
- return false;
- }
for (int i = 0; i < getRegionInfoCount(); i++) {
if (!getRegionInfo(i).isInitialized()) {
memoizedIsInitialized = 0;
return false;
}
}
+ if (!getMergedRegionInfo().isInitialized()) {
+ memoizedIsInitialized = 0;
+ return false;
+ }
memoizedIsInitialized = 1;
return true;
}
@@ -20313,11 +20196,11 @@ public final class MasterProcedureProtos {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeMessage(1, getUserInfo());
}
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
- output.writeMessage(2, getTableName());
- }
for (int i = 0; i < regionInfo_.size(); i++) {
- output.writeMessage(3, regionInfo_.get(i));
+ output.writeMessage(2, regionInfo_.get(i));
+ }
+ if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ output.writeMessage(3, getMergedRegionInfo());
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
output.writeBool(4, forcible_);
@@ -20334,13 +20217,13 @@ public final class MasterProcedureProtos {
size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(1, getUserInfo());
}
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ for (int i = 0; i < regionInfo_.size(); i++) {
size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
- .computeMessageSize(2, getTableName());
+ .computeMessageSize(2, regionInfo_.get(i));
}
- for (int i = 0; i < regionInfo_.size(); i++) {
+ if (((bitField0_ & 0x00000002) == 0x00000002)) {
size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
- .computeMessageSize(3, regionInfo_.get(i));
+ .computeMessageSize(3, getMergedRegionInfo());
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
@@ -20357,1281 +20240,10 @@ public final class MasterProcedureProtos {
if (obj == this) {
return true;
}
- if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData)) {
+ if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.MergeTableRegionsStateData)) {
return super.equals(obj);
}
- org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData) obj;
-
- boolean result = true;
- result = result && (hasUserInfo() == other.hasUserInfo());
- if (hasUserInfo()) {
- result = result && getUserInfo()
- .equals(other.getUserInfo());
- }
- result = result && (hasTableName() == other.hasTableName());
- if (hasTableName()) {
- result = result && getTableName()
- .equals(other.getTableName());
- }
- result = result && getRegionInfoList()
- .equals(other.getRegionInfoList());
- result = result && (hasForcible() == other.hasForcible());
- if (hasForcible()) {
- result = result && (getForcible()
- == other.getForcible());
- }
- result = result && unknownFields.equals(other.unknownFields);
- return result;
- }
-
- @java.lang.Override
- public int hashCode() {
- if (memoizedHashCode != 0) {
- return memoizedHashCode;
- }
- int hash = 41;
- hash = (19 * hash) + getDescriptorForType().hashCode();
- if (hasUserInfo()) {
- hash = (37 * hash) + USER_INFO_FIELD_NUMBER;
- hash = (53 * hash) + getUserInfo().hashCode();
- }
- if (hasTableName()) {
- hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
- hash = (53 * hash) + getTableName().hashCode();
- }
- if (getRegionInfoCount() > 0) {
- hash = (37 * hash) + REGION_INFO_FIELD_NUMBER;
- hash = (53 * hash) + getRegionInfoList().hashCode();
- }
- if (hasForcible()) {
- hash = (37 * hash) + FORCIBLE_FIELD_NUMBER;
- hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean(
- getForcible());
- }
- hash = (29 * hash) + unknownFields.hashCode();
- memoizedHashCode = hash;
- return hash;
- }
-
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData parseFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData parseFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData parseFrom(byte[] data)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData parseFrom(
- byte[] data,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData parseFrom(java.io.InputStream input)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseWithIOException(PARSER, input);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData parseFrom(
- java.io.InputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseWithIOException(PARSER, input, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData parseDelimitedFrom(java.io.InputStream input)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseDelimitedWithIOException(PARSER, input);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData parseDelimitedFrom(
- java.io.InputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData parseFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseWithIOException(PARSER, input);
- }
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData parseFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseWithIOException(PARSER, input, extensionRegistry);
- }
-
- public Builder newBuilderForType() { return newBuilder(); }
- public static Builder newBuilder() {
- return DEFAULT_INSTANCE.toBuilder();
- }
- public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData prototype) {
- return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
- }
- public Builder toBuilder() {
- return this == DEFAULT_INSTANCE
- ? new Builder() : new Builder().mergeFrom(this);
- }
-
- @java.lang.Override
- protected Builder newBuilderForType(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
- Builder builder = new Builder(parent);
- return builder;
- }
- /**
- * Protobuf type {@code hbase.pb.DispatchMergingRegionsStateData}
- */
- public static final class Builder extends
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
- // @@protoc_insertion_point(builder_implements:hbase.pb.DispatchMergingRegionsStateData)
- org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateDataOrBuilder {
- public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
- getDescriptor() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.internal_static_hbase_pb_DispatchMergingRegionsStateData_descriptor;
- }
-
- protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
- internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.internal_static_hbase_pb_DispatchMergingRegionsStateData_fieldAccessorTable
- .ensureFieldAccessorsInitialized(
- org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData.Builder.class);
- }
-
- // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData.newBuilder()
- private Builder() {
- maybeForceBuilderInitialization();
- }
-
- private Builder(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
- super(parent);
- maybeForceBuilderInitialization();
- }
- private void maybeForceBuilderInitialization() {
- if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .alwaysUseFieldBuilders) {
- getUserInfoFieldBuilder();
- getTableNameFieldBuilder();
- getRegionInfoFieldBuilder();
- }
- }
- public Builder clear() {
- super.clear();
- if (userInfoBuilder_ == null) {
- userInfo_ = null;
- } else {
- userInfoBuilder_.clear();
- }
- bitField0_ = (bitField0_ & ~0x00000001);
- if (tableNameBuilder_ == null) {
- tableName_ = null;
- } else {
- tableNameBuilder_.clear();
- }
- bitField0_ = (bitField0_ & ~0x00000002);
- if (regionInfoBuilder_ == null) {
- regionInfo_ = java.util.Collections.emptyList();
- bitField0_ = (bitField0_ & ~0x00000004);
- } else {
- regionInfoBuilder_.clear();
- }
- forcible_ = false;
- bitField0_ = (bitField0_ & ~0x00000008);
- return this;
- }
-
- public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
- getDescriptorForType() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.internal_static_hbase_pb_DispatchMergingRegionsStateData_descriptor;
- }
-
- public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData getDefaultInstanceForType() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData.getDefaultInstance();
- }
-
- public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData build() {
- org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData result = buildPartial();
- if (!result.isInitialized()) {
- throw newUninitializedMessageException(result);
- }
- return result;
- }
-
- public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData buildPartial() {
- org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData(this);
- int from_bitField0_ = bitField0_;
- int to_bitField0_ = 0;
- if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
- to_bitField0_ |= 0x00000001;
- }
- if (userInfoBuilder_ == null) {
- result.userInfo_ = userInfo_;
- } else {
- result.userInfo_ = userInfoBuilder_.build();
- }
- if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
- to_bitField0_ |= 0x00000002;
- }
- if (tableNameBuilder_ == null) {
- result.tableName_ = tableName_;
- } else {
- result.tableName_ = tableNameBuilder_.build();
- }
- if (regionInfoBuilder_ == null) {
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
- regionInfo_ = java.util.Collections.unmodifiableList(regionInfo_);
- bitField0_ = (bitField0_ & ~0x00000004);
- }
- result.regionInfo_ = regionInfo_;
- } else {
- result.regionInfo_ = regionInfoBuilder_.build();
- }
- if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
- to_bitField0_ |= 0x00000004;
- }
- result.forcible_ = forcible_;
- result.bitField0_ = to_bitField0_;
- onBuilt();
- return result;
- }
-
- public Builder clone() {
- return (Builder) super.clone();
- }
- public Builder setField(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
- Object value) {
- return (Builder) super.setField(field, value);
- }
- public Builder clearField(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
- return (Builder) super.clearField(field);
- }
- public Builder clearOneof(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
- return (Builder) super.clearOneof(oneof);
- }
- public Builder setRepeatedField(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
- int index, Object value) {
- return (Builder) super.setRepeatedField(field, index, value);
- }
- public Builder addRepeatedField(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
- Object value) {
- return (Builder) super.addRepeatedField(field, value);
- }
- public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
- if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData) {
- return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData)other);
- } else {
- super.mergeFrom(other);
- return this;
- }
- }
-
- public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData other) {
- if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData.getDefaultInstance()) return this;
- if (other.hasUserInfo()) {
- mergeUserInfo(other.getUserInfo());
- }
- if (other.hasTableName()) {
- mergeTableName(other.getTableName());
- }
- if (regionInfoBuilder_ == null) {
- if (!other.regionInfo_.isEmpty()) {
- if (regionInfo_.isEmpty()) {
- regionInfo_ = other.regionInfo_;
- bitField0_ = (bitField0_ & ~0x00000004);
- } else {
- ensureRegionInfoIsMutable();
- regionInfo_.addAll(other.regionInfo_);
- }
- onChanged();
- }
- } else {
- if (!other.regionInfo_.isEmpty()) {
- if (regionInfoBuilder_.isEmpty()) {
- regionInfoBuilder_.dispose();
- regionInfoBuilder_ = null;
- regionInfo_ = other.regionInfo_;
- bitField0_ = (bitField0_ & ~0x00000004);
- regionInfoBuilder_ =
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
- getRegionInfoFieldBuilder() : null;
- } else {
- regionInfoBuilder_.addAllMessages(other.regionInfo_);
- }
- }
- }
- if (other.hasForcible()) {
- setForcible(other.getForcible());
- }
- this.mergeUnknownFields(other.unknownFields);
- onChanged();
- return this;
- }
-
- public final boolean isInitialized() {
- if (!hasUserInfo()) {
- return false;
- }
- if (!hasTableName()) {
- return false;
- }
- if (!getUserInfo().isInitialized()) {
- return false;
- }
- if (!getTableName().isInitialized()) {
- return false;
- }
- for (int i = 0; i < getRegionInfoCount(); i++) {
- if (!getRegionInfo(i).isInitialized()) {
- return false;
- }
- }
- return true;
- }
-
- public Builder mergeFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData parsedMessage = null;
- try {
- parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
- parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData) e.getUnfinishedMessage();
- throw e.unwrapIOException();
- } finally {
- if (parsedMessage != null) {
- mergeFrom(parsedMessage);
- }
- }
- return this;
- }
- private int bitField0_;
-
- private org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation userInfo_ = null;
- private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
- org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationOrBuilder> userInfoBuilder_;
- /**
- * <code>required .hbase.pb.UserInformation user_info = 1;</code>
- */
- public boolean hasUserInfo() {
- return ((bitField0_ & 0x00000001) == 0x00000001);
- }
- /**
- * <code>required .hbase.pb.UserInformation user_info = 1;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation getUserInfo() {
- if (userInfoBuilder_ == null) {
- return userInfo_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance() : userInfo_;
- } else {
- return userInfoBuilder_.getMessage();
- }
- }
- /**
- * <code>required .hbase.pb.UserInformation user_info = 1;</code>
- */
- public Builder setUserInfo(org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation value) {
- if (userInfoBuilder_ == null) {
- if (value == null) {
- throw new NullPointerException();
- }
- userInfo_ = value;
- onChanged();
- } else {
- userInfoBuilder_.setMessage(value);
- }
- bitField0_ |= 0x00000001;
- return this;
- }
- /**
- * <code>required .hbase.pb.UserInformation user_info = 1;</code>
- */
- public Builder setUserInfo(
- org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.Builder builderForValue) {
- if (userInfoBuilder_ == null) {
- userInfo_ = builderForValue.build();
- onChanged();
- } else {
- userInfoBuilder_.setMessage(builderForValue.build());
- }
- bitField0_ |= 0x00000001;
- return this;
- }
- /**
- * <code>required .hbase.pb.UserInformation user_info = 1;</code>
- */
- public Builder mergeUserInfo(org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation value) {
- if (userInfoBuilder_ == null) {
- if (((bitField0_ & 0x00000001) == 0x00000001) &&
- userInfo_ != null &&
- userInfo_ != org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance()) {
- userInfo_ =
- org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.newBuilder(userInfo_).mergeFrom(value).buildPartial();
- } else {
- userInfo_ = value;
- }
- onChanged();
- } else {
- userInfoBuilder_.mergeFrom(value);
- }
- bitField0_ |= 0x00000001;
- return this;
- }
- /**
- * <code>required .hbase.pb.UserInformation user_info = 1;</code>
- */
- public Builder clearUserInfo() {
- if (userInfoBuilder_ == null) {
- userInfo_ = null;
- onChanged();
- } else {
- userInfoBuilder_.clear();
- }
- bitField0_ = (bitField0_ & ~0x00000001);
- return this;
- }
- /**
- * <code>required .hbase.pb.UserInformation user_info = 1;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.Builder getUserInfoBuilder() {
- bitField0_ |= 0x00000001;
- onChanged();
- return getUserInfoFieldBuilder().getBuilder();
- }
- /**
- * <code>required .hbase.pb.UserInformation user_info = 1;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationOrBuilder getUserInfoOrBuilder() {
- if (userInfoBuilder_ != null) {
- return userInfoBuilder_.getMessageOrBuilder();
- } else {
- return userInfo_ == null ?
- org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance() : userInfo_;
- }
- }
- /**
- * <code>required .hbase.pb.UserInformation user_info = 1;</code>
- */
- private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
- org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationOrBuilder>
- getUserInfoFieldBuilder() {
- if (userInfoBuilder_ == null) {
- userInfoBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
- org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationOrBuilder>(
- getUserInfo(),
- getParentForChildren(),
- isClean());
- userInfo_ = null;
- }
- return userInfoBuilder_;
- }
-
- private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_ = null;
- private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
- /**
- * <code>required .hbase.pb.TableName table_name = 2;</code>
- */
- public boolean hasTableName() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
- }
- /**
- * <code>required .hbase.pb.TableName table_name = 2;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() {
- if (tableNameBuilder_ == null) {
- return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
- } else {
- return tableNameBuilder_.getMessage();
- }
- }
- /**
- * <code>required .hbase.pb.TableName table_name = 2;</code>
- */
- public Builder setTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) {
- if (tableNameBuilder_ == null) {
- if (value == null) {
- throw new NullPointerException();
- }
- tableName_ = value;
- onChanged();
- } else {
- tableNameBuilder_.setMessage(value);
- }
- bitField0_ |= 0x00000002;
- return this;
- }
- /**
- * <code>required .hbase.pb.TableName table_name = 2;</code>
- */
- public Builder setTableName(
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
- if (tableNameBuilder_ == null) {
- tableName_ = builderForValue.build();
- onChanged();
- } else {
- tableNameBuilder_.setMessage(builderForValue.build());
- }
- bitField0_ |= 0x00000002;
- return this;
- }
- /**
- * <code>required .hbase.pb.TableName table_name = 2;</code>
- */
- public Builder mergeTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) {
- if (tableNameBuilder_ == null) {
- if (((bitField0_ & 0x00000002) == 0x00000002) &&
- tableName_ != null &&
- tableName_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) {
- tableName_ =
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
- } else {
- tableName_ = value;
- }
- onChanged();
- } else {
- tableNameBuilder_.mergeFrom(value);
- }
- bitField0_ |= 0x00000002;
- return this;
- }
- /**
- * <code>required .hbase.pb.TableName table_name = 2;</code>
- */
- public Builder clearTableName() {
- if (tableNameBuilder_ == null) {
- tableName_ = null;
- onChanged();
- } else {
- tableNameBuilder_.clear();
- }
- bitField0_ = (bitField0_ & ~0x00000002);
- return this;
- }
- /**
- * <code>required .hbase.pb.TableName table_name = 2;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() {
- bitField0_ |= 0x00000002;
- onChanged();
- return getTableNameFieldBuilder().getBuilder();
- }
- /**
- * <code>required .hbase.pb.TableName table_name = 2;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
- if (tableNameBuilder_ != null) {
- return tableNameBuilder_.getMessageOrBuilder();
- } else {
- return tableName_ == null ?
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
- }
- }
- /**
- * <code>required .hbase.pb.TableName table_name = 2;</code>
- */
- private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>
- getTableNameFieldBuilder() {
- if (tableNameBuilder_ == null) {
- tableNameBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
- getTableName(),
- getParentForChildren(),
- isClean());
- tableName_ = null;
- }
- return tableNameBuilder_;
- }
-
- private java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo> regionInfo_ =
- java.util.Collections.emptyList();
- private void ensureRegionInfoIsMutable() {
- if (!((bitField0_ & 0x00000004) == 0x00000004)) {
- regionInfo_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo>(regionInfo_);
- bitField0_ |= 0x00000004;
- }
- }
-
- private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder> regionInfoBuilder_;
-
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
- */
- public java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo> getRegionInfoList() {
- if (regionInfoBuilder_ == null) {
- return java.util.Collections.unmodifiableList(regionInfo_);
- } else {
- return regionInfoBuilder_.getMessageList();
- }
- }
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
- */
- public int getRegionInfoCount() {
- if (regionInfoBuilder_ == null) {
- return regionInfo_.size();
- } else {
- return regionInfoBuilder_.getCount();
- }
- }
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo(int index) {
- if (regionInfoBuilder_ == null) {
- return regionInfo_.get(index);
- } else {
- return regionInfoBuilder_.getMessage(index);
- }
- }
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
- */
- public Builder setRegionInfo(
- int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo value) {
- if (regionInfoBuilder_ == null) {
- if (value == null) {
- throw new NullPointerException();
- }
- ensureRegionInfoIsMutable();
- regionInfo_.set(index, value);
- onChanged();
- } else {
- regionInfoBuilder_.setMessage(index, value);
- }
- return this;
- }
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
- */
- public Builder setRegionInfo(
- int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder builderForValue) {
- if (regionInfoBuilder_ == null) {
- ensureRegionInfoIsMutable();
- regionInfo_.set(index, builderForValue.build());
- onChanged();
- } else {
- regionInfoBuilder_.setMessage(index, builderForValue.build());
- }
- return this;
- }
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
- */
- public Builder addRegionInfo(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo value) {
- if (regionInfoBuilder_ == null) {
- if (value == null) {
- throw new NullPointerException();
- }
- ensureRegionInfoIsMutable();
- regionInfo_.add(value);
- onChanged();
- } else {
- regionInfoBuilder_.addMessage(value);
- }
- return this;
- }
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
- */
- public Builder addRegionInfo(
- int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo value) {
- if (regionInfoBuilder_ == null) {
- if (value == null) {
- throw new NullPointerException();
- }
- ensureRegionInfoIsMutable();
- regionInfo_.add(index, value);
- onChanged();
- } else {
- regionInfoBuilder_.addMessage(index, value);
- }
- return this;
- }
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
- */
- public Builder addRegionInfo(
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder builderForValue) {
- if (regionInfoBuilder_ == null) {
- ensureRegionInfoIsMutable();
- regionInfo_.add(builderForValue.build());
- onChanged();
- } else {
- regionInfoBuilder_.addMessage(builderForValue.build());
- }
- return this;
- }
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
- */
- public Builder addRegionInfo(
- int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder builderForValue) {
- if (regionInfoBuilder_ == null) {
- ensureRegionInfoIsMutable();
- regionInfo_.add(index, builderForValue.build());
- onChanged();
- } else {
- regionInfoBuilder_.addMessage(index, builderForValue.build());
- }
- return this;
- }
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
- */
- public Builder addAllRegionInfo(
- java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo> values) {
- if (regionInfoBuilder_ == null) {
- ensureRegionInfoIsMutable();
- org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
- values, regionInfo_);
- onChanged();
- } else {
- regionInfoBuilder_.addAllMessages(values);
- }
- return this;
- }
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
- */
- public Builder clearRegionInfo() {
- if (regionInfoBuilder_ == null) {
- regionInfo_ = java.util.Collections.emptyList();
- bitField0_ = (bitField0_ & ~0x00000004);
- onChanged();
- } else {
- regionInfoBuilder_.clear();
- }
- return this;
- }
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
- */
- public Builder removeRegionInfo(int index) {
- if (regionInfoBuilder_ == null) {
- ensureRegionInfoIsMutable();
- regionInfo_.remove(index);
- onChanged();
- } else {
- regionInfoBuilder_.remove(index);
- }
- return this;
- }
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder getRegionInfoBuilder(
- int index) {
- return getRegionInfoFieldBuilder().getBuilder(index);
- }
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionInfoOrBuilder(
- int index) {
- if (regionInfoBuilder_ == null) {
- return regionInfo_.get(index); } else {
- return regionInfoBuilder_.getMessageOrBuilder(index);
- }
- }
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
- */
- public java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder>
- getRegionInfoOrBuilderList() {
- if (regionInfoBuilder_ != null) {
- return regionInfoBuilder_.getMessageOrBuilderList();
- } else {
- return java.util.Collections.unmodifiableList(regionInfo_);
- }
- }
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder addRegionInfoBuilder() {
- return getRegionInfoFieldBuilder().addBuilder(
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance());
- }
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder addRegionInfoBuilder(
- int index) {
- return getRegionInfoFieldBuilder().addBuilder(
- index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance());
- }
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 3;</code>
- */
- public java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder>
- getRegionInfoBuilderList() {
- return getRegionInfoFieldBuilder().getBuilderList();
- }
- private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder>
- getRegionInfoFieldBuilder() {
- if (regionInfoBuilder_ == null) {
- regionInfoBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder>(
- regionInfo_,
- ((bitField0_ & 0x00000004) == 0x00000004),
- getParentForChildren(),
- isClean());
- regionInfo_ = null;
- }
- return regionInfoBuilder_;
- }
-
- private boolean forcible_ ;
- /**
- * <code>optional bool forcible = 4;</code>
- */
- public boolean hasForcible() {
- return ((bitField0_ & 0x00000008) == 0x00000008);
- }
- /**
- * <code>optional bool forcible = 4;</code>
- */
- public boolean getForcible() {
- return forcible_;
- }
- /**
- * <code>optional bool forcible = 4;</code>
- */
- public Builder setForcible(boolean value) {
- bitField0_ |= 0x00000008;
- forcible_ = value;
- onChanged();
- return this;
- }
- /**
- * <code>optional bool forcible = 4;</code>
- */
- public Builder clearForcible() {
- bitField0_ = (bitField0_ & ~0x00000008);
- forcible_ = false;
- onChanged();
- return this;
- }
- public final Builder setUnknownFields(
- final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
- return super.setUnknownFields(unknownFields);
- }
-
- public final Builder mergeUnknownFields(
- final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
- return super.mergeUnknownFields(unknownFields);
- }
-
-
- // @@protoc_insertion_point(builder_scope:hbase.pb.DispatchMergingRegionsStateData)
- }
-
- // @@protoc_insertion_point(class_scope:hbase.pb.DispatchMergingRegionsStateData)
- private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData DEFAULT_INSTANCE;
- static {
- DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData();
- }
-
- public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData getDefaultInstance() {
- return DEFAULT_INSTANCE;
- }
-
- @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<DispatchMergingRegionsStateData>
- PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<DispatchMergingRegionsStateData>() {
- public DispatchMergingRegionsStateData parsePartialFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return new DispatchMergingRegionsStateData(input, extensionRegistry);
- }
- };
-
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<DispatchMergingRegionsStateData> parser() {
- return PARSER;
- }
-
- @java.lang.Override
- public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<DispatchMergingRegionsStateData> getParserForType() {
- return PARSER;
- }
-
- public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsStateData getDefaultInstanceForType() {
- return DEFAULT_INSTANCE;
- }
-
- }
-
- public interface MergeTableRegionsStateDataOrBuilder extends
- // @@protoc_insertion_point(interface_extends:hbase.pb.MergeTableRegionsStateData)
- org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
- /**
- * <code>required .hbase.pb.UserInformation user_info = 1;</code>
- */
- boolean hasUserInfo();
- /**
- * <code>required .hbase.pb.UserInformation user_info = 1;</code>
- */
- org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation getUserInfo();
- /**
- * <code>required .hbase.pb.UserInformation user_info = 1;</code>
- */
- org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationOrBuilder getUserInfoOrBuilder();
-
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 2;</code>
- */
- java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo>
- getRegionInfoList();
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 2;</code>
- */
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo(int index);
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 2;</code>
- */
- int getRegionInfoCount();
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 2;</code>
- */
- java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder>
- getRegionInfoOrBuilderList();
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 2;</code>
- */
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionInfoOrBuilder(
- int index);
-
- /**
- * <code>required .hbase.pb.RegionInfo merged_region_info = 3;</code>
- */
- boolean hasMergedRegionInfo();
- /**
- * <code>required .hbase.pb.RegionInfo merged_region_info = 3;</code>
- */
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo getMergedRegionInfo();
- /**
- * <code>required .hbase.pb.RegionInfo merged_region_info = 3;</code>
- */
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getMergedRegionInfoOrBuilder();
-
- /**
- * <code>optional bool forcible = 4 [default = false];</code>
- */
- boolean hasForcible();
- /**
- * <code>optional bool forcible = 4 [default = false];</code>
- */
- boolean getForcible();
- }
- /**
- * Protobuf type {@code hbase.pb.MergeTableRegionsStateData}
- */
- public static final class MergeTableRegionsStateData extends
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
- // @@protoc_insertion_point(message_implements:hbase.pb.MergeTableRegionsStateData)
- MergeTableRegionsStateDataOrBuilder {
- // Use MergeTableRegionsStateData.newBuilder() to construct.
- private MergeTableRegionsStateData(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
- super(builder);
- }
- private MergeTableRegionsStateData() {
- regionInfo_ = java.util.Collections.emptyList();
- forcible_ = false;
- }
-
- @java.lang.Override
- public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
- return this.unknownFields;
- }
- private MergeTableRegionsStateData(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- this();
- int mutable_bitField0_ = 0;
- org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
- try {
- boolean done = false;
- while (!done) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
- done = true;
- break;
- default: {
- if (!parseUnknownField(input, unknownFields,
- extensionRegistry, tag)) {
- done = true;
- }
- break;
- }
- case 10: {
- org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.Builder subBuilder = null;
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
- subBuilder = userInfo_.toBuilder();
- }
- userInfo_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.PARSER, extensionRegistry);
- if (subBuilder != null) {
- subBuilder.mergeFrom(userInfo_);
- userInfo_ = subBuilder.buildPartial();
- }
- bitField0_ |= 0x00000001;
- break;
- }
- case 18: {
- if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
- regionInfo_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo>();
- mutable_bitField0_ |= 0x00000002;
- }
- regionInfo_.add(
- input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.PARSER, extensionRegistry));
- break;
- }
- case 26: {
- org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder subBuilder = null;
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
- subBuilder = mergedRegionInfo_.toBuilder();
- }
- mergedRegionInfo_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.PARSER, extensionRegistry);
- if (subBuilder != null) {
- subBuilder.mergeFrom(mergedRegionInfo_);
- mergedRegionInfo_ = subBuilder.buildPartial();
- }
- bitField0_ |= 0x00000002;
- break;
- }
- case 32: {
- bitField0_ |= 0x00000004;
- forcible_ = input.readBool();
- break;
- }
- }
- }
- } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
- throw e.setUnfinishedMessage(this);
- } catch (java.io.IOException e) {
- throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
- e).setUnfinishedMessage(this);
- } finally {
- if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
- regionInfo_ = java.util.Collections.unmodifiableList(regionInfo_);
- }
- this.unknownFields = unknownFields.build();
- makeExtensionsImmutable();
- }
- }
- public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
- getDescriptor() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.internal_static_hbase_pb_MergeTableRegionsStateData_descriptor;
- }
-
- protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
- internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.internal_static_hbase_pb_MergeTableRegionsStateData_fieldAccessorTable
- .ensureFieldAccessorsInitialized(
- org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.MergeTableRegionsStateData.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.MergeTableRegionsStateData.Builder.class);
- }
-
- private int bitField0_;
- public static final int USER_INFO_FIELD_NUMBER = 1;
- private org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation userInfo_;
- /**
- * <code>required .hbase.pb.UserInformation user_info = 1;</code>
- */
- public boolean hasUserInfo() {
- return ((bitField0_ & 0x00000001) == 0x00000001);
- }
- /**
- * <code>required .hbase.pb.UserInformation user_info = 1;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation getUserInfo() {
- return userInfo_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance() : userInfo_;
- }
- /**
- * <code>required .hbase.pb.UserInformation user_info = 1;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationOrBuilder getUserInfoOrBuilder() {
- return userInfo_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance() : userInfo_;
- }
-
- public static final int REGION_INFO_FIELD_NUMBER = 2;
- private java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo> regionInfo_;
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 2;</code>
- */
- public java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo> getRegionInfoList() {
- return regionInfo_;
- }
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 2;</code>
- */
- public java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder>
- getRegionInfoOrBuilderList() {
- return regionInfo_;
- }
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 2;</code>
- */
- public int getRegionInfoCount() {
- return regionInfo_.size();
- }
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 2;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo(int index) {
- return regionInfo_.get(index);
- }
- /**
- * <code>repeated .hbase.pb.RegionInfo region_info = 2;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionInfoOrBuilder(
- int index) {
- return regionInfo_.get(index);
- }
-
- public static final int MERGED_REGION_INFO_FIELD_NUMBER = 3;
- private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo mergedRegionInfo_;
- /**
- * <code>required .hbase.pb.RegionInfo merged_region_info = 3;</code>
- */
- public boolean hasMergedRegionInfo() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
- }
- /**
- * <code>required .hbase.pb.RegionInfo merged_region_info = 3;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo getMergedRegionInfo() {
- return mergedRegionInfo_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance() : mergedRegionInfo_;
- }
- /**
- * <code>required .hbase.pb.RegionInfo merged_region_info = 3;</code>
- */
- public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getMergedRegionInfoOrBuilder() {
- return mergedRegionInfo_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance() : mergedRegionInfo_;
- }
-
- public static final int FORCIBLE_FIELD_NUMBER = 4;
- private boolean forcible_;
- /**
- * <code>optional bool forcible = 4 [default = false];</code>
- */
- public boolean hasForcible() {
- return ((bitField0_ & 0x00000004) == 0x00000004);
- }
- /**
- * <code>optional bool forcible = 4 [default = false];</code>
- */
- public boolean getForcible() {
- return forcible_;
- }
-
- private byte memoizedIsInitialized = -1;
- public final boolean isInitialized() {
- byte isInitialized = memoizedIsInitialized;
- if (isInitialized == 1) return true;
- if (isInitialized == 0) return false;
-
- if (!hasUserInfo()) {
- memoizedIsInitialized = 0;
- return false;
- }
- if (!hasMergedRegionInfo()) {
- memoizedIsInitialized = 0;
- return false;
- }
- if (!getUserInfo().isInitialized()) {
- memoizedIsInitialized = 0;
- return false;
- }
- for (int i = 0; i < getRegionInfoCount(); i++) {
- if (!getRegionInfo(i).isInitialized()) {
- memoizedIsInitialized = 0;
- return false;
- }
- }
- if (!getMergedRegionInfo().isInitialized()) {
- memoizedIsInitialized = 0;
- return false;
- }
- memoizedIsInitialized = 1;
- return true;
- }
-
- public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
- throws java.io.IOException {
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
- output.writeMessage(1, getUserInfo());
- }
- for (int i = 0; i < regionInfo_.size(); i++) {
- output.writeMessage(2, regionInfo_.get(i));
- }
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
- output.writeMessage(3, getMergedRegionInfo());
- }
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
- output.writeBool(4, forcible_);
- }
- unknownFields.writeTo(output);
- }
-
- public int getSerializedSize() {
- int size = memoizedSize;
- if (size != -1) return size;
-
- size = 0;
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
- .computeMessageSize(1, getUserInfo());
- }
- for (int i = 0; i < regionInfo_.size(); i++) {
- size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
- .computeMessageSize(2, regionInfo_.get(i));
- }
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
- .computeMessageSize(3, getMergedRegionInfo());
- }
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
- .computeBoolSize(4, forcible_);
- }
- size += unknownFields.getSerializedSize();
- memoizedSize = size;
- return size;
- }
-
- private static final long serialVersionUID = 0L;
- @java.lang.Override
- public boolean equals(final java.lang.Object obj) {
- if (obj == this) {
- return true;
- }
- if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.MergeTableRegionsStateData)) {
- return super.equals(obj);
- }
- org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.MergeTableRegionsStateData other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.MergeTableRegionsStateData) obj;
+ org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.MergeTableRegionsStateData other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.MergeTableRegionsStateData) obj;
boolean result = true;
result = result && (hasUserInfo() == other.hasUserInfo());
@@ -25422,11 +24034,6 @@ public final class MasterProcedureProtos {
org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hbase_pb_RestoreSnapshotStateData_fieldAccessorTable;
private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
- internal_static_hbase_pb_DispatchMergingRegionsStateData_descriptor;
- private static final
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
- internal_static_hbase_pb_DispatchMergingRegionsStateData_fieldAccessorTable;
- private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
internal_static_hbase_pb_MergeTableRegionsStateData_descriptor;
private static final
org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
@@ -25524,146 +24131,135 @@ public final class MasterProcedureProtos {
"o\0221\n\023region_info_for_add\030\006 \003(\0132\024.hbase.p" +
"b.RegionInfo\022T\n!parent_to_child_regions_" +
"pair_list\030\007 \003(\0132).hbase.pb.RestoreParent" +
- "ToChildRegionsPair\"\265\001\n\037DispatchMergingRe" +
- "gionsStateData\022,\n\tuser_info\030\001 \002(\0132\031.hbas" +
- "e.pb.UserInformation\022\'\n\ntable_name\030\002 \002(\013" +
- "2\023.hbase.pb.TableName\022)\n\013region_info\030\003 \003" +
- "(\0132\024.hbase.pb.RegionInfo\022\020\n\010forcible\030\004 \001" +
- "(\010\"\300\001\n\032MergeTableRegionsStateData\022,\n\tuse",
- "r_info\030\001 \002(\0132\031.hbase.pb.UserInformation\022" +
- ")\n\013region_info\030\002 \003(\0132\024.hbase.pb.RegionIn" +
- "fo\0220\n\022merged_region_info\030\003 \002(\0132\024.hbase.p" +
- "b.RegionInfo\022\027\n\010forcible\030\004 \001(\010:\005false\"\254\001" +
- "\n\031SplitTableRegionStateData\022,\n\tuser_info" +
- "\030\001 \002(\0132\031.hbase.pb.UserInformation\0220\n\022par" +
- "ent_region_info\030\002 \002(\0132\024.hbase.pb.RegionI" +
- "nfo\022/\n\021child_region_info\030\003 \003(\0132\024.hbase.p" +
- "b.RegionInfo\"\201\002\n\024ServerCrashStateData\022)\n" +
- "\013server_name\030\001 \002(\0132\024.hbase.pb.ServerName",
- "\022\036\n\026distributed_log_replay\030\002 \001(\010\0227\n\031regi" +
- "ons_on_crashed_server\030\003 \003(\0132\024.hbase.pb.R" +
- "egionInfo\022.\n\020regions_assigned\030\004 \003(\0132\024.hb" +
- "ase.pb.RegionInfo\022\025\n\rcarrying_meta\030\005 \001(\010" +
- "\022\036\n\020should_split_wal\030\006 \001(\010:\004true*\330\001\n\020Cre" +
- "ateTableState\022\036\n\032CREATE_TABLE_PRE_OPERAT" +
- "ION\020\001\022 \n\034CREATE_TABLE_WRITE_FS_LAYOUT\020\002\022" +
- "\034\n\030CREATE_TABLE_ADD_TO_META\020\003\022\037\n\033CREATE_" +
- "TABLE_ASSIGN_REGIONS\020\004\022\"\n\036CREATE_TABLE_U" +
- "PDATE_DESC_CACHE\020\005\022\037\n\033CREATE_TABLE_POST_",
- "OPERATION\020\006*\207\002\n\020ModifyTableState\022\030\n\024MODI" +
- "FY_TABLE_PREPARE\020\001\022\036\n\032MODIFY_TABLE_PRE_O" +
- "PERATION\020\002\022(\n$MODIFY_TABLE_UPDATE_TABLE_" +
- "DESCRIPTOR\020\003\022&\n\"MODIFY_TABLE_REMOVE_REPL" +
- "ICA_COLUMN\020\004\022!\n\035MODIFY_TABLE_DELETE_FS_L" +
- "AYOUT\020\005\022\037\n\033MODIFY_TABLE_POST_OPERATION\020\006" +
- "\022#\n\037MODIFY_TABLE_REOPEN_ALL_REGIONS\020\007*\212\002" +
- "\n\022TruncateTableState\022 \n\034TRUNCATE_TABLE_P" +
- "RE_OPERATION\020\001\022#\n\037TRUNCATE_TABLE_REMOVE_" +
- "FROM_META\020\002\022\"\n\036TRUNCATE_TABLE_CLEAR_FS_L",
- "AYOUT\020\003\022#\n\037TRUNCATE_TABLE_CREATE_FS_LAYO" +
- "UT\020\004\022\036\n\032TRUNCATE_TABLE_ADD_TO_META\020\005\022!\n\035" +
- "TRUNCATE_TABLE_ASSIGN_REGIONS\020\006\022!\n\035TRUNC" +
- "ATE_TABLE_POST_OPERATION\020\007*\337\001\n\020DeleteTab" +
- "leState\022\036\n\032DELETE_TABLE_PRE_OPERATION\020\001\022" +
- "!\n\035DELETE_TABLE_REMOVE_FROM_META\020\002\022 \n\034DE" +
- "LETE_TABLE_CLEAR_FS_LAYOUT\020\003\022\"\n\036DELETE_T" +
- "ABLE_UPDATE_DESC_CACHE\020\004\022!\n\035DELETE_TABLE" +
- "_UNASSIGN_REGIONS\020\005\022\037\n\033DELETE_TABLE_POST" +
- "_OPERATION\020\006*\320\001\n\024CreateNamespaceState\022\034\n",
- "\030CREATE_NAMESPACE_PREPARE\020\001\022%\n!CREATE_NA" +
- "MESPACE_CREATE_DIRECTORY\020\002\022)\n%CREATE_NAM" +
- "ESPACE_INSERT_INTO_NS_TABLE\020\003\022\036\n\032CREATE_" +
- "NAMESPACE_UPDATE_ZK\020\004\022(\n$CREATE_NAMESPAC" +
- "E_SET_NAMESPACE_QUOTA\020\005*z\n\024ModifyNamespa" +
- "ceState\022\034\n\030MODIFY_NAMESPACE_PREPARE\020\001\022$\n" +
- " MODIFY_NAMESPACE_UPDATE_NS_TABLE\020\002\022\036\n\032M" +
- "ODIFY_NAMESPACE_UPDATE_ZK\020\003*\332\001\n\024DeleteNa" +
- "mespaceState\022\034\n\030DELETE_NAMESPACE_PREPARE" +
- "\020\001\022)\n%DELETE_NAMESPACE_DELETE_FROM_NS_TA",
- "BLE\020\002\022#\n\037DELETE_NAMESPACE_REMOVE_FROM_ZK" +
- "\020\003\022\'\n#DELETE_NAMESPACE_DELETE_DIRECTORIE" +
- "S\020\004\022+\n\'DELETE_NAMESPACE_REMOVE_NAMESPACE" +
- "_QUOTA\020\005*\331\001\n\024AddColumnFamilyState\022\035\n\031ADD" +
- "_COLUMN_FAMILY_PREPARE\020\001\022#\n\037ADD_COLUMN_F" +
- "AMILY_PRE_OPERATION\020\002\022-\n)ADD_COLUMN_FAMI" +
- "LY_UPDATE_TABLE_DESCRIPTOR\020\003\022$\n ADD_COLU" +
- "MN_FAMILY_POST_OPERATION\020\004\022(\n$ADD_COLUMN" +
- "_FAMILY_REOPEN_ALL_REGIONS\020\005*\353\001\n\027ModifyC" +
- "olumnFamilyState\022 \n\034MODIFY_COLUMN_FAMILY",
- "_PREPARE\020\001\022&\n\"MODIFY_COLUMN_FAMILY_PRE_O" +
- "PERATION\020\002\0220\n,MODIFY_COLUMN_FAMILY_UPDAT" +
- "E_TABLE_DESCRIPTOR\020\003\022\'\n#MODIFY_COLUMN_FA" +
- "MILY_POST_OPERATION\020\004\022+\n\'MODIFY_COLUMN_F" +
- "AMILY_REOPEN_ALL_REGIONS\020\005*\226\002\n\027DeleteCol" +
- "umnFamilyState\022 \n\034DELETE_COLUMN_FAMILY_P" +
- "REPARE\020\001\022&\n\"DELETE_COLUMN_FAMILY_PRE_OPE" +
- "RATION\020\002\0220\n,DELETE_COLUMN_FAMILY_UPDATE_" +
- "TABLE_DESCRIPTOR\020\003\022)\n%DELETE_COLUMN_FAMI" +
- "LY_DELETE_FS_LAYOUT\020\004\022\'\n#DELETE_COLUMN_F",
- "AMILY_POST_OPERATION\020\005\022+\n\'DELETE_COLUMN_" +
- "FAMILY_REOPEN_ALL_REGIONS\020\006*\350\001\n\020EnableTa" +
- "bleState\022\030\n\024ENABLE_TABLE_PREPARE\020\001\022\036\n\032EN" +
- "ABLE_TABLE_PRE_OPERATION\020\002\022)\n%ENABLE_TAB" +
- "LE_SET_ENABLING_TABLE_STATE\020\003\022$\n ENABLE_" +
- "TABLE_MARK_REGIONS_ONLINE\020\004\022(\n$ENABLE_TA" +
- "BLE_SET_ENABLED_TABLE_STATE\020\005\022\037\n\033ENABLE_" +
- "TABLE_POST_OPERATION\020\006*\362\001\n\021DisableTableS" +
- "tate\022\031\n\025DISABLE_TABLE_PREPARE\020\001\022\037\n\033DISAB" +
- "LE_TABLE_PRE_OPERATION\020\002\022+\n\'DISABLE_TABL",
- "E_SET_DISABLING_TABLE_STATE\020\003\022&\n\"DISABLE" +
- "_TABLE_MARK_REGIONS_OFFLINE\020\004\022*\n&DISABLE" +
- "_TABLE_SET_DISABLED_TABLE_STATE\020\005\022 \n\034DIS" +
- "ABLE_TABLE_POST_OPERATION\020\006*\346\001\n\022CloneSna" +
- "pshotState\022 \n\034CLONE_SNAPSHOT_PRE_OPERATI" +
- "ON\020\001\022\"\n\036CLONE_SNAPSHOT_WRITE_FS_LAYOUT\020\002" +
- "\022\036\n\032CLONE_SNAPSHOT_ADD_TO_META\020\003\022!\n\035CLON" +
- "E_SNAPSHOT_ASSIGN_REGIONS\020\004\022$\n CLONE_SNA" +
- "PSHOT_UPDATE_DESC_CACHE\020\005\022!\n\035CLONE_SNAPS" +
- "HOT_POST_OPERATION\020\006*\260\001\n\024RestoreSnapshot",
- "State\022\"\n\036RESTORE_SNAPSHOT_PRE_OPERATION\020" +
- "\001\022,\n(RESTORE_SNAPSHOT_UPDATE_TABLE_DESCR" +
- "IPTOR\020\002\022$\n RESTORE_SNAPSHOT_WRITE_FS_LAY" +
- "OUT\020\003\022 \n\034RESTORE_SNAPSHOT_UPDATE_META\020\004*" +
- "\376\001\n\033DispatchMergingRegionsState\022$\n DISPA" +
- "TCH_MERGING_REGIONS_PREPARE\020\001\022*\n&DISPATC" +
- "H_MERGING_REGIONS_PRE_OPERATION\020\002\0223\n/DIS" +
- "PATCH_MERGING_REGIONS_MOVE_REGION_TO_SAM" +
- "E_RS\020\003\022+\n\'DISPATCH_MERGING_REGIONS_DO_ME" +
- "RGE_IN_RS\020\004\022+\n\'DISPATCH_MERGING_REGIONS_",
- "POST_OPERATION\020\005*\376\003\n\026MergeTableRegionsSt" +
- "ate\022\037\n\033MERGE_TABLE_REGIONS_PREPARE\020\001\022.\n*" +
- "MERGE_TABLE_REGIONS_MOVE_REGION_TO_SAME_" +
- "RS\020\002\022+\n\'MERGE_TABLE_
<TRUNCATED>
[3/7] hbase git commit: HBASE-17470 Remove merge region code from
region server (Stephen Yuan Jiang)
Posted by sy...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-protocol-shaded/src/main/protobuf/Admin.proto
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/protobuf/Admin.proto b/hbase-protocol-shaded/src/main/protobuf/Admin.proto
index e8cf10c..338c80b 100644
--- a/hbase-protocol-shaded/src/main/protobuf/Admin.proto
+++ b/hbase-protocol-shaded/src/main/protobuf/Admin.proto
@@ -189,22 +189,6 @@ message UpdateFavoredNodesResponse {
optional uint32 response = 1;
}
-/**
- * Merges the specified regions.
- * <p>
- * This method currently closes the regions and then merges them
- */
-message MergeRegionsRequest {
- required RegionSpecifier region_a = 1;
- required RegionSpecifier region_b = 2;
- optional bool forcible = 3 [default = false];
- // wall clock time from master
- optional uint64 master_system_time = 4;
-}
-
-message MergeRegionsResponse {
-}
-
// Protocol buffer version of WAL for replication
message WALEntry {
required WALKey key = 1;
@@ -307,9 +291,6 @@ service AdminService {
rpc CompactRegion(CompactRegionRequest)
returns(CompactRegionResponse);
- rpc MergeRegions(MergeRegionsRequest)
- returns(MergeRegionsResponse);
-
rpc ReplicateWALEntry(ReplicateWALEntryRequest)
returns(ReplicateWALEntryResponse);
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-protocol-shaded/src/main/protobuf/Master.proto
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/protobuf/Master.proto b/hbase-protocol-shaded/src/main/protobuf/Master.proto
index 7cd9921..e62f52c 100644
--- a/hbase-protocol-shaded/src/main/protobuf/Master.proto
+++ b/hbase-protocol-shaded/src/main/protobuf/Master.proto
@@ -80,21 +80,6 @@ message MoveRegionResponse {
}
/**
- * Dispatch merging the specified regions.
- */
-message DispatchMergingRegionsRequest {
- required RegionSpecifier region_a = 1;
- required RegionSpecifier region_b = 2;
- optional bool forcible = 3 [default = false];
- optional uint64 nonce_group = 4 [default = 0];
- optional uint64 nonce = 5 [default = 0];
-}
-
-message DispatchMergingRegionsResponse {
- optional uint64 proc_id = 1;
-}
-
-/**
* Merging the specified regions in a table.
*/
message MergeTableRegionsRequest {
@@ -625,10 +610,6 @@ service MasterService {
rpc MoveRegion(MoveRegionRequest)
returns(MoveRegionResponse);
- /** Master dispatch merging the regions */
- rpc DispatchMergingRegions(DispatchMergingRegionsRequest)
- returns(DispatchMergingRegionsResponse);
-
/** Master merge the regions */
rpc MergeTableRegions(MergeTableRegionsRequest)
returns(MergeTableRegionsResponse);
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-protocol-shaded/src/main/protobuf/MasterProcedure.proto
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/protobuf/MasterProcedure.proto b/hbase-protocol-shaded/src/main/protobuf/MasterProcedure.proto
index 23d914e..ef3f973 100644
--- a/hbase-protocol-shaded/src/main/protobuf/MasterProcedure.proto
+++ b/hbase-protocol-shaded/src/main/protobuf/MasterProcedure.proto
@@ -262,21 +262,6 @@ message RestoreSnapshotStateData {
repeated RestoreParentToChildRegionsPair parent_to_child_regions_pair_list = 7;
}
-enum DispatchMergingRegionsState {
- DISPATCH_MERGING_REGIONS_PREPARE = 1;
- DISPATCH_MERGING_REGIONS_PRE_OPERATION = 2;
- DISPATCH_MERGING_REGIONS_MOVE_REGION_TO_SAME_RS = 3;
- DISPATCH_MERGING_REGIONS_DO_MERGE_IN_RS = 4;
- DISPATCH_MERGING_REGIONS_POST_OPERATION = 5;
-}
-
-message DispatchMergingRegionsStateData {
- required UserInformation user_info = 1;
- required TableName table_name = 2;
- repeated RegionInfo region_info = 3;
- optional bool forcible = 4;
-}
-
enum MergeTableRegionsState {
MERGE_TABLE_REGIONS_PREPARE = 1;
MERGE_TABLE_REGIONS_MOVE_REGION_TO_SAME_RS = 2;
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseMasterAndRegionObserver.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseMasterAndRegionObserver.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseMasterAndRegionObserver.java
index 00376cb..a0a1d49 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseMasterAndRegionObserver.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseMasterAndRegionObserver.java
@@ -59,11 +59,23 @@ public class BaseMasterAndRegionObserver extends BaseRegionObserver
HTableDescriptor desc, HRegionInfo[] regions) throws IOException {
}
+ /**
+ * @deprecated As of release 2.0.0, this will be removed in HBase 3.0.0
+ * (<a href="https://issues.apache.org/jira/browse/HBASE-">HBASE-</a>).
+ * Use {@link #preMergeRegions(ObserverContext, HRegionInfo[])}
+ */
+ @Deprecated
@Override
public void preDispatchMerge(final ObserverContext<MasterCoprocessorEnvironment> ctx,
HRegionInfo regionA, HRegionInfo regionB) throws IOException {
}
+ /**
+ * @deprecated As of release 2.0.0, this will be removed in HBase 3.0.0
+ * (<a href="https://issues.apache.org/jira/browse/HBASE-">HBASE-</a>).
+ * Use {@link #postMergeRegions(ObserverContext, HRegionInfo[])}
+ */
+ @Deprecated
@Override
public void postDispatchMerge(final ObserverContext<MasterCoprocessorEnvironment> ctx,
HRegionInfo regionA, HRegionInfo regionB) throws IOException {
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseMasterObserver.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseMasterObserver.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseMasterObserver.java
index 461148b..98f21b2 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseMasterObserver.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseMasterObserver.java
@@ -111,11 +111,23 @@ public class BaseMasterObserver implements MasterObserver {
final HRegionInfo[] regions) throws IOException {
}
+ /**
+ * @deprecated As of release 2.0.0, this will be removed in HBase 3.0.0
+ * (<a href="https://issues.apache.org/jira/browse/HBASE-">HBASE-</a>).
+ * Use {@link #preMergeRegions(ObserverContext, HRegionInfo[])}
+ */
+ @Deprecated
@Override
public void preDispatchMerge(final ObserverContext<MasterCoprocessorEnvironment> ctx,
HRegionInfo regionA, HRegionInfo regionB) throws IOException {
}
+ /**
+ * @deprecated As of release 2.0.0, this will be removed in HBase 3.0.0
+ * (<a href="https://issues.apache.org/jira/browse/HBASE-">HBASE-</a>).
+ * Use {@link #postMergeRegions(ObserverContext, HRegionInfo[])}
+ */
+ @Deprecated
@Override
public void postDispatchMerge(final ObserverContext<MasterCoprocessorEnvironment> ctx,
HRegionInfo regionA, HRegionInfo regionB) throws IOException {
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterObserver.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterObserver.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterObserver.java
index 82b3cfa..bb0e732 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterObserver.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterObserver.java
@@ -1699,7 +1699,11 @@ public interface MasterObserver extends Coprocessor {
* @param regionA first region to be merged
* @param regionB second region to be merged
* @throws IOException if an error occurred on the coprocessor
+ * @deprecated As of release 2.0.0, this will be removed in HBase 3.0.0
+ * (<a href="https://issues.apache.org/jira/browse/HBASE-">HBASE-</a>).
+ * Use {@link #preMergeRegions(ObserverContext, HRegionInfo[])}
*/
+ @Deprecated
void preDispatchMerge(final ObserverContext<MasterCoprocessorEnvironment> ctx,
HRegionInfo regionA, HRegionInfo regionB) throws IOException;
@@ -1709,7 +1713,11 @@ public interface MasterObserver extends Coprocessor {
* @param regionA first region to be merged
* @param regionB second region to be merged
* @throws IOException if an error occurred on the coprocessor
+ * @deprecated As of release 2.0.0, this will be removed in HBase 3.0.0
+ * (<a href="https://issues.apache.org/jira/browse/HBASE-">HBASE-</a>).
+ * Use {@link #postMergeRegions(ObserverContext, HRegionInfo[])}
*/
+ @Deprecated
void postDispatchMerge(final ObserverContext<MasterCoprocessorEnvironment> c,
final HRegionInfo regionA, final HRegionInfo regionB) throws IOException;
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
index a41960b..ab7a25e 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
@@ -109,7 +109,6 @@ import org.apache.hadoop.hbase.master.procedure.CreateTableProcedure;
import org.apache.hadoop.hbase.master.procedure.DeleteColumnFamilyProcedure;
import org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure;
import org.apache.hadoop.hbase.master.procedure.DisableTableProcedure;
-import org.apache.hadoop.hbase.master.procedure.DispatchMergingRegionsProcedure;
import org.apache.hadoop.hbase.master.procedure.EnableTableProcedure;
import org.apache.hadoop.hbase.master.procedure.MasterProcedureConstants;
import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
@@ -1419,55 +1418,6 @@ public class HMaster extends HRegionServer implements MasterServices {
}
@Override
- public long dispatchMergingRegions(
- final HRegionInfo regionInfoA,
- final HRegionInfo regionInfoB,
- final boolean forcible,
- final long nonceGroup,
- final long nonce) throws IOException {
- checkInitialized();
-
- TableName tableName = regionInfoA.getTable();
- if (tableName == null || regionInfoB.getTable() == null) {
- throw new UnknownRegionException ("Can't merge regions without table associated");
- }
-
- if (!tableName.equals(regionInfoB.getTable())) {
- throw new IOException ("Cannot merge regions from two different tables");
- }
-
- if (regionInfoA.compareTo(regionInfoB) == 0) {
- throw new MergeRegionException(
- "Cannot merge a region to itself " + regionInfoA + ", " + regionInfoB);
- }
-
- HRegionInfo [] regionsToMerge = new HRegionInfo[2];
- regionsToMerge [0] = regionInfoA;
- regionsToMerge [1] = regionInfoB;
-
- return MasterProcedureUtil.submitProcedure(
- new MasterProcedureUtil.NonceProcedureRunnable(this, nonceGroup, nonce) {
- @Override
- protected void run() throws IOException {
- getMaster().getMasterCoprocessorHost().preDispatchMerge(regionInfoA, regionInfoB);
-
- LOG.info(getClientIdAuditPrefix() + " Merge regions "
- + regionInfoA.getEncodedName() + " and " + regionInfoB.getEncodedName());
-
- submitProcedure(new DispatchMergingRegionsProcedure(procedureExecutor.getEnvironment(),
- tableName, regionsToMerge, forcible));
-
- getMaster().getMasterCoprocessorHost().postDispatchMerge(regionInfoA, regionInfoB);
- }
-
- @Override
- protected String getDescription() {
- return "DisableTableProcedure";
- }
- });
- }
-
- @Override
public long mergeRegions(
final HRegionInfo[] regionsToMerge,
final boolean forcible,
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterCoprocessorHost.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterCoprocessorHost.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterCoprocessorHost.java
index 2f17a5f..3dec2e8 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterCoprocessorHost.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterCoprocessorHost.java
@@ -773,28 +773,6 @@ public class MasterCoprocessorHost
});
}
- public void preDispatchMerge(final HRegionInfo regionInfoA, final HRegionInfo regionInfoB)
- throws IOException {
- execOperation(coprocessors.isEmpty() ? null : new CoprocessorOperation() {
- @Override
- public void call(MasterObserver oserver, ObserverContext<MasterCoprocessorEnvironment> ctx)
- throws IOException {
- oserver.preDispatchMerge(ctx, regionInfoA, regionInfoB);
- }
- });
- }
-
- public void postDispatchMerge(final HRegionInfo regionInfoA, final HRegionInfo regionInfoB)
- throws IOException {
- execOperation(coprocessors.isEmpty() ? null : new CoprocessorOperation() {
- @Override
- public void call(MasterObserver oserver, ObserverContext<MasterCoprocessorEnvironment> ctx)
- throws IOException {
- oserver.postDispatchMerge(ctx, regionInfoA, regionInfoB);
- }
- });
- }
-
public void preMergeRegions(final HRegionInfo[] regionsToMerge)
throws IOException {
execOperation(coprocessors.isEmpty() ? null : new CoprocessorOperation() {
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java
index 5873986..1151c92 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java
@@ -501,52 +501,6 @@ public class MasterRpcServices extends RSRpcServices
}
@Override
- public DispatchMergingRegionsResponse dispatchMergingRegions(RpcController c,
- DispatchMergingRegionsRequest request) throws ServiceException {
- try {
- master.checkInitialized();
- } catch (IOException ioe) {
- throw new ServiceException(ioe);
- }
-
- final byte[] encodedNameOfRegionA = request.getRegionA().getValue()
- .toByteArray();
- final byte[] encodedNameOfRegionB = request.getRegionB().getValue()
- .toByteArray();
- if (request.getRegionA().getType() != RegionSpecifierType.ENCODED_REGION_NAME
- || request.getRegionB().getType() != RegionSpecifierType.ENCODED_REGION_NAME) {
- LOG.warn("mergeRegions specifier type: expected: "
- + RegionSpecifierType.ENCODED_REGION_NAME + " actual: region_a="
- + request.getRegionA().getType() + ", region_b="
- + request.getRegionB().getType());
- }
-
- RegionStates regionStates = master.getAssignmentManager().getRegionStates();
- RegionState regionStateA = regionStates.getRegionState(Bytes.toString(encodedNameOfRegionA));
- RegionState regionStateB = regionStates.getRegionState(Bytes.toString(encodedNameOfRegionB));
- if (regionStateA == null || regionStateB == null) {
- throw new ServiceException(new UnknownRegionException(
- Bytes.toStringBinary(regionStateA == null ? encodedNameOfRegionA
- : encodedNameOfRegionB)));
- }
-
- final HRegionInfo regionInfoA = regionStateA.getRegion();
- final HRegionInfo regionInfoB = regionStateB.getRegion();
-
- try {
- long procId = master.dispatchMergingRegions(
- regionInfoA,
- regionInfoB,
- request.getForcible(),
- request.getNonceGroup(),
- request.getNonce());
- return DispatchMergingRegionsResponse.newBuilder().setProcId(procId).build();
- } catch (IOException ioe) {
- throw new ServiceException(ioe);
- }
- }
-
- @Override
public EnableCatalogJanitorResponse enableCatalogJanitor(RpcController c,
EnableCatalogJanitorRequest req) throws ServiceException {
try {
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java
index 5019eda..79ebca5 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java
@@ -326,23 +326,6 @@ public interface MasterServices extends Server {
boolean registerService(Service instance);
/**
- * Merge two regions. The real implementation is on the regionserver, master
- * just move the regions together and send MERGE RPC to regionserver
- * @param region_a region to merge
- * @param region_b region to merge
- * @param forcible true if do a compulsory merge, otherwise we will only merge
- * two adjacent regions
- * @return procedure Id
- * @throws IOException
- */
- long dispatchMergingRegions(
- final HRegionInfo region_a,
- final HRegionInfo region_b,
- final boolean forcible,
- final long nonceGroup,
- final long nonce) throws IOException;
-
- /**
* @return true if master is the active one
*/
boolean isActiveMaster();
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
index f3b21ac..38493cd 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
@@ -896,37 +896,6 @@ public class ServerManager {
}
/**
- * Sends an MERGE REGIONS RPC to the specified server to merge the specified
- * regions.
- * <p>
- * A region server could reject the close request because it either does not
- * have the specified region.
- * @param server server to merge regions
- * @param region_a region to merge
- * @param region_b region to merge
- * @param forcible true if do a compulsory merge, otherwise we will only merge
- * two adjacent regions
- * @throws IOException
- */
- public void sendRegionsMerge(ServerName server, HRegionInfo region_a,
- HRegionInfo region_b, boolean forcible, final User user) throws IOException {
- if (server == null)
- throw new NullPointerException("Passed server is null");
- if (region_a == null || region_b == null)
- throw new NullPointerException("Passed region is null");
- AdminService.BlockingInterface admin = getRsAdmin(server);
- if (admin == null) {
- throw new IOException("Attempting to send MERGE REGIONS RPC to server "
- + server.toString() + " for region "
- + region_a.getRegionNameAsString() + ","
- + region_b.getRegionNameAsString()
- + " failed because no RPC connection found to this server");
- }
- HBaseRpcController controller = newRpcController();
- ProtobufUtil.mergeRegions(controller, admin, region_a, region_b, forcible, user);
- }
-
- /**
* Check if a region server is reachable and has the expected start code
*/
public boolean isServerReachable(ServerName server) {
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/DispatchMergingRegionsProcedure.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/DispatchMergingRegionsProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/DispatchMergingRegionsProcedure.java
deleted file mode 100644
index ee92932..0000000
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/DispatchMergingRegionsProcedure.java
+++ /dev/null
@@ -1,579 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-package org.apache.hadoop.hbase.master.procedure;
-
-import java.io.IOException;
-import java.io.InputStream;
-import java.io.InterruptedIOException;
-import java.io.OutputStream;
-import java.util.Map;
-import java.util.concurrent.atomic.AtomicBoolean;
-
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
-import org.apache.hadoop.hbase.HRegionInfo;
-import org.apache.hadoop.hbase.RegionLoad;
-import org.apache.hadoop.hbase.ServerLoad;
-import org.apache.hadoop.hbase.ServerName;
-import org.apache.hadoop.hbase.TableName;
-import org.apache.hadoop.hbase.UnknownRegionException;
-import org.apache.hadoop.hbase.classification.InterfaceAudience;
-import org.apache.hadoop.hbase.exceptions.MergeRegionException;
-import org.apache.hadoop.hbase.exceptions.RegionOpeningException;
-import org.apache.hadoop.hbase.master.AssignmentManager;
-import org.apache.hadoop.hbase.master.CatalogJanitor;
-import org.apache.hadoop.hbase.master.RegionPlan;
-import org.apache.hadoop.hbase.master.RegionState;
-import org.apache.hadoop.hbase.master.RegionStates;
-import org.apache.hadoop.hbase.master.ServerManager;
-import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsState;
-import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-
-/**
- * The procedure to Merge a region in a table.
- */
-@InterfaceAudience.Private
-public class DispatchMergingRegionsProcedure
- extends AbstractStateMachineTableProcedure<DispatchMergingRegionsState> {
- private static final Log LOG = LogFactory.getLog(DispatchMergingRegionsProcedure.class);
-
- private final AtomicBoolean aborted = new AtomicBoolean(false);
- private Boolean traceEnabled;
- private AssignmentManager assignmentManager;
- private int timeout;
- private ServerName regionLocation;
- private String regionsToMergeListFullName;
- private String regionsToMergeListEncodedName;
-
- private TableName tableName;
- private HRegionInfo [] regionsToMerge;
- private boolean forcible;
-
- public DispatchMergingRegionsProcedure() {
- this.traceEnabled = isTraceEnabled();
- this.assignmentManager = null;
- this.timeout = -1;
- this.regionLocation = null;
- this.regionsToMergeListFullName = null;
- this.regionsToMergeListEncodedName = null;
- }
-
- public DispatchMergingRegionsProcedure(
- final MasterProcedureEnv env,
- final TableName tableName,
- final HRegionInfo [] regionsToMerge,
- final boolean forcible) {
- super(env);
- this.traceEnabled = isTraceEnabled();
- this.assignmentManager = getAssignmentManager(env);
- this.tableName = tableName;
- // For now, we only merge 2 regions. It could be extended to more than 2 regions in
- // the future.
- assert(regionsToMerge.length == 2);
- this.regionsToMerge = regionsToMerge;
- this.forcible = forcible;
-
- this.timeout = -1;
- this.regionsToMergeListFullName = getRegionsToMergeListFullNameString();
- this.regionsToMergeListEncodedName = getRegionsToMergeListEncodedNameString();
- }
-
- @Override
- protected Flow executeFromState(
- final MasterProcedureEnv env,
- final DispatchMergingRegionsState state) throws InterruptedException {
- if (isTraceEnabled()) {
- LOG.trace(this + " execute state=" + state);
- }
-
- try {
- switch (state) {
- case DISPATCH_MERGING_REGIONS_PREPARE:
- prepareMergeRegion(env);
- setNextState(DispatchMergingRegionsState.DISPATCH_MERGING_REGIONS_PRE_OPERATION);
- break;
- case DISPATCH_MERGING_REGIONS_PRE_OPERATION:
- //Unused for now - reserve to add preMerge coprocessor in the future
- setNextState(DispatchMergingRegionsState.DISPATCH_MERGING_REGIONS_MOVE_REGION_TO_SAME_RS);
- break;
- case DISPATCH_MERGING_REGIONS_MOVE_REGION_TO_SAME_RS:
- if (MoveRegionsToSameRS(env)) {
- setNextState(DispatchMergingRegionsState.DISPATCH_MERGING_REGIONS_DO_MERGE_IN_RS);
- } else {
- LOG.info("Cancel merging regions " + getRegionsToMergeListFullNameString()
- + ", because can't move them to the same RS");
- setNextState(DispatchMergingRegionsState.DISPATCH_MERGING_REGIONS_POST_OPERATION);
- }
- break;
- case DISPATCH_MERGING_REGIONS_DO_MERGE_IN_RS:
- doMergeInRS(env);
- setNextState(DispatchMergingRegionsState.DISPATCH_MERGING_REGIONS_POST_OPERATION);
- break;
- case DISPATCH_MERGING_REGIONS_POST_OPERATION:
- //Unused for now - reserve to add postCompletedMerge coprocessor in the future
- return Flow.NO_MORE_STATE;
- default:
- throw new UnsupportedOperationException(this + " unhandled state=" + state);
- }
- } catch (IOException e) {
- LOG.warn("Error trying to merge regions " + getRegionsToMergeListFullNameString() +
- " in the table " + tableName + " (in state=" + state + ")", e);
-
- setFailure("master-merge-regions", e);
- }
- return Flow.HAS_MORE_STATE;
- }
-
- @Override
- protected void rollbackState(
- final MasterProcedureEnv env,
- final DispatchMergingRegionsState state) throws IOException, InterruptedException {
- if (isTraceEnabled()) {
- LOG.trace(this + " rollback state=" + state);
- }
-
- try {
- switch (state) {
- case DISPATCH_MERGING_REGIONS_POST_OPERATION:
- case DISPATCH_MERGING_REGIONS_DO_MERGE_IN_RS:
- String msg = this + " We are in the " + state + " state."
- + " It is complicated to rollback the merge operation that region server is working on."
- + " Rollback is not supported and we should let the merge operation to complete";
- LOG.warn(msg);
- // PONR
- throw new UnsupportedOperationException(this + " unhandled state=" + state);
- case DISPATCH_MERGING_REGIONS_MOVE_REGION_TO_SAME_RS:
- break; // nothing to rollback
- case DISPATCH_MERGING_REGIONS_PRE_OPERATION:
- break; // nothing to rollback
- case DISPATCH_MERGING_REGIONS_PREPARE:
- break; // nothing to rollback
- default:
- throw new UnsupportedOperationException(this + " unhandled state=" + state);
- }
- } catch (Exception e) {
- // This will be retried. Unless there is a bug in the code,
- // this should be just a "temporary error" (e.g. network down)
- LOG.warn("Failed rollback attempt step " + state + " for merging the regions "
- + getRegionsToMergeListFullNameString() + " in table " + tableName, e);
- throw e;
- }
- }
-
- @Override
- protected DispatchMergingRegionsState getState(final int stateId) {
- return DispatchMergingRegionsState.valueOf(stateId);
- }
-
- @Override
- protected int getStateId(final DispatchMergingRegionsState state) {
- return state.getNumber();
- }
-
- @Override
- protected DispatchMergingRegionsState getInitialState() {
- return DispatchMergingRegionsState.DISPATCH_MERGING_REGIONS_PREPARE;
- }
-
- /*
- * Check whether we are in the state that can be rollback
- */
- @Override
- protected boolean isRollbackSupported(final DispatchMergingRegionsState state) {
- switch (state) {
- case DISPATCH_MERGING_REGIONS_POST_OPERATION:
- case DISPATCH_MERGING_REGIONS_DO_MERGE_IN_RS:
- // It is not safe to rollback if we reach to these states.
- return false;
- default:
- break;
- }
- return true;
- }
-
- @Override
- public void serializeStateData(final OutputStream stream) throws IOException {
- super.serializeStateData(stream);
-
- MasterProcedureProtos.DispatchMergingRegionsStateData.Builder dispatchMergingRegionsMsg =
- MasterProcedureProtos.DispatchMergingRegionsStateData.newBuilder()
- .setUserInfo(MasterProcedureUtil.toProtoUserInfo(getUser()))
- .setTableName(ProtobufUtil.toProtoTableName(tableName))
- .setForcible(forcible);
- for (HRegionInfo hri: regionsToMerge) {
- dispatchMergingRegionsMsg.addRegionInfo(HRegionInfo.convert(hri));
- }
- dispatchMergingRegionsMsg.build().writeDelimitedTo(stream);
- }
-
- @Override
- public void deserializeStateData(final InputStream stream) throws IOException {
- super.deserializeStateData(stream);
-
- MasterProcedureProtos.DispatchMergingRegionsStateData dispatchMergingRegionsMsg =
- MasterProcedureProtos.DispatchMergingRegionsStateData.parseDelimitedFrom(stream);
- setUser(MasterProcedureUtil.toUserInfo(dispatchMergingRegionsMsg.getUserInfo()));
- tableName = ProtobufUtil.toTableName(dispatchMergingRegionsMsg.getTableName());
-
- assert(dispatchMergingRegionsMsg.getRegionInfoCount() == 2);
- regionsToMerge = new HRegionInfo[dispatchMergingRegionsMsg.getRegionInfoCount()];
- for (int i = 0; i < regionsToMerge.length; i++) {
- regionsToMerge[i] = HRegionInfo.convert(dispatchMergingRegionsMsg.getRegionInfo(i));
- }
- }
-
- @Override
- public void toStringClassDetails(StringBuilder sb) {
- sb.append(getClass().getSimpleName());
- sb.append(" (table=");
- sb.append(tableName);
- sb.append(" regions=");
- sb.append(getRegionsToMergeListFullNameString());
- sb.append(" forcible=");
- sb.append(forcible);
- sb.append(")");
- }
-
- @Override
- protected boolean acquireLock(final MasterProcedureEnv env) {
- return !env.getProcedureQueue().waitRegions(
- this, getTableName(), regionsToMerge[0], regionsToMerge[1]);
- }
-
- @Override
- protected void releaseLock(final MasterProcedureEnv env) {
- env.getProcedureQueue().wakeRegions(this, getTableName(), regionsToMerge[0], regionsToMerge[1]);
- }
-
- @Override
- public TableName getTableName() {
- return tableName;
- }
-
- @Override
- public TableOperationType getTableOperationType() {
- return TableOperationType.MERGE;
- }
-
- /**
- * Prepare merge and do some check
- * @param env MasterProcedureEnv
- * @throws IOException
- */
- private void prepareMergeRegion(final MasterProcedureEnv env) throws IOException {
- // Note: the following logic assumes that we only have 2 regions to merge. In the future,
- // if we want to extend to more than 2 regions, the code needs to modify a little bit.
- //
- CatalogJanitor catalogJanitor = env.getMasterServices().getCatalogJanitor();
- boolean regionAHasMergeQualifier = !catalogJanitor.cleanMergeQualifier(regionsToMerge[0]);
- if (regionAHasMergeQualifier
- || !catalogJanitor.cleanMergeQualifier(regionsToMerge[1])) {
- String msg = "Skip merging regions " + regionsToMerge[0].getRegionNameAsString()
- + ", " + regionsToMerge[1].getRegionNameAsString() + ", because region "
- + (regionAHasMergeQualifier ? regionsToMerge[0].getEncodedName() : regionsToMerge[1]
- .getEncodedName()) + " has merge qualifier";
- LOG.info(msg);
- throw new MergeRegionException(msg);
- }
-
- RegionStates regionStates = getAssignmentManager(env).getRegionStates();
- RegionState regionStateA = regionStates.getRegionState(regionsToMerge[0].getEncodedName());
- RegionState regionStateB = regionStates.getRegionState(regionsToMerge[1].getEncodedName());
- if (regionStateA == null || regionStateB == null) {
- throw new UnknownRegionException(
- regionStateA == null ?
- regionsToMerge[0].getEncodedName() : regionsToMerge[1].getEncodedName());
- }
-
- if (!regionStateA.isOpened() || !regionStateB.isOpened()) {
- throw new MergeRegionException(
- "Unable to merge regions not online " + regionStateA + ", " + regionStateB);
- }
-
- if (regionsToMerge[0].getReplicaId() != HRegionInfo.DEFAULT_REPLICA_ID ||
- regionsToMerge[1].getReplicaId() != HRegionInfo.DEFAULT_REPLICA_ID) {
- throw new MergeRegionException("Can't merge non-default replicas");
- }
-
- if (!forcible && !HRegionInfo.areAdjacent(regionsToMerge[0], regionsToMerge[1])) {
- throw new MergeRegionException(
- "Unable to merge not adjacent regions "
- + regionsToMerge[0].getRegionNameAsString() + ", "
- + regionsToMerge[1].getRegionNameAsString()
- + " where forcible = " + forcible);
- }
- }
-
- /**
- * Move all regions to the same region server
- * @param env MasterProcedureEnv
- * @return whether target regions hosted by the same RS
- * @throws IOException
- */
- private boolean MoveRegionsToSameRS(final MasterProcedureEnv env) throws IOException {
- // Make sure regions are on the same regionserver before send merge
- // regions request to region server.
- //
- boolean onSameRS = isRegionsOnTheSameServer(env);
- if (!onSameRS) {
- // Note: the following logic assumes that we only have 2 regions to merge. In the future,
- // if we want to extend to more than 2 regions, the code needs to modify a little bit.
- //
- RegionStates regionStates = getAssignmentManager(env).getRegionStates();
- ServerName regionLocation2 = regionStates.getRegionServerOfRegion(regionsToMerge[1]);
-
- RegionLoad loadOfRegionA = getRegionLoad(env, regionLocation, regionsToMerge[0]);
- RegionLoad loadOfRegionB = getRegionLoad(env, regionLocation2, regionsToMerge[1]);
- if (loadOfRegionA != null && loadOfRegionB != null
- && loadOfRegionA.getRequestsCount() < loadOfRegionB.getRequestsCount()) {
- // switch regionsToMerge[0] and regionsToMerge[1]
- HRegionInfo tmpRegion = this.regionsToMerge[0];
- this.regionsToMerge[0] = this.regionsToMerge[1];
- this.regionsToMerge[1] = tmpRegion;
- ServerName tmpLocation = regionLocation;
- regionLocation = regionLocation2;
- regionLocation2 = tmpLocation;
- }
-
- long startTime = EnvironmentEdgeManager.currentTime();
-
- RegionPlan regionPlan = new RegionPlan(regionsToMerge[1], regionLocation2, regionLocation);
- LOG.info("Moving regions to same server for merge: " + regionPlan.toString());
- getAssignmentManager(env).balance(regionPlan);
- do {
- try {
- Thread.sleep(20);
- // Make sure check RIT first, then get region location, otherwise
- // we would make a wrong result if region is online between getting
- // region location and checking RIT
- boolean isRIT = regionStates.isRegionInTransition(regionsToMerge[1]);
- regionLocation2 = regionStates.getRegionServerOfRegion(regionsToMerge[1]);
- onSameRS = regionLocation.equals(regionLocation2);
- if (onSameRS || !isRIT) {
- // Regions are on the same RS, or regionsToMerge[1] is not in
- // RegionInTransition any more
- break;
- }
- } catch (InterruptedException e) {
- InterruptedIOException iioe = new InterruptedIOException();
- iioe.initCause(e);
- throw iioe;
- }
- } while ((EnvironmentEdgeManager.currentTime() - startTime) <= getTimeout(env));
- }
- return onSameRS;
- }
-
- /**
- * Do the real merge operation in the region server that hosts regions
- * @param env MasterProcedureEnv
- * @throws IOException
- */
- private void doMergeInRS(final MasterProcedureEnv env) throws IOException {
- long duration = 0;
- long startTime = EnvironmentEdgeManager.currentTime();
- do {
- try {
- if (getServerName(env) == null) {
- // The merge probably already happen. Check
- RegionState regionState = getAssignmentManager(env).getRegionStates().getRegionState(
- regionsToMerge[0].getEncodedName());
- if (regionState.isMerging() || regionState.isMerged()) {
- LOG.info("Merge regions " + getRegionsToMergeListEncodedNameString() +
- " is in progress or completed. No need to send a new request.");
- } else {
- LOG.warn("Cannot sending merge to hosting server of the regions " +
- getRegionsToMergeListEncodedNameString() + " as the server is unknown");
- }
- return;
- }
- // TODO: the following RPC call is not idempotent. Multiple calls (eg. after master
- // failover, re-execute this step) could result in some exception thrown that does not
- // paint the correct picture. This behavior is on-par with old releases. Improvement
- // could happen in the future.
- env.getMasterServices().getServerManager().sendRegionsMerge(
- getServerName(env),
- regionsToMerge[0],
- regionsToMerge[1],
- forcible,
- getUser());
- LOG.info("Sent merge to server " + getServerName(env) + " for region " +
- getRegionsToMergeListEncodedNameString() + ", forcible=" + forcible);
- return;
- } catch (RegionOpeningException roe) {
- // Do a retry since region should be online on RS immediately
- LOG.warn("Failed mergering regions in " + getServerName(env) + ", retrying...", roe);
- } catch (Exception ie) {
- LOG.warn("Failed sending merge to " + getServerName(env) + " for regions " +
- getRegionsToMergeListEncodedNameString() + ", forcible=" + forcible, ie);
- return;
- }
- } while ((duration = EnvironmentEdgeManager.currentTime() - startTime) <= getTimeout(env));
-
- // If we reaches here, it means that we get timed out.
- String msg = "Failed sending merge to " + getServerName(env) + " after " + duration + "ms";
- LOG.warn(msg);
- throw new IOException(msg);
- }
-
- private RegionLoad getRegionLoad(
- final MasterProcedureEnv env,
- final ServerName sn,
- final HRegionInfo hri) {
- ServerManager serverManager = env.getMasterServices().getServerManager();
- ServerLoad load = serverManager.getLoad(sn);
- if (load != null) {
- Map<byte[], RegionLoad> regionsLoad = load.getRegionsLoad();
- if (regionsLoad != null) {
- return regionsLoad.get(hri.getRegionName());
- }
- }
- return null;
- }
-
- /**
- * The procedure could be restarted from a different machine. If the variable is null, we need to
- * retrieve it.
- * @param env MasterProcedureEnv
- * @return whether target regions hosted by the same RS
- */
- private boolean isRegionsOnTheSameServer(final MasterProcedureEnv env) throws IOException{
- Boolean onSameRS = true;
- int i = 0;
- RegionStates regionStates = getAssignmentManager(env).getRegionStates();
- regionLocation = regionStates.getRegionServerOfRegion(regionsToMerge[i]);
- if (regionLocation != null) {
- for(i = 1; i < regionsToMerge.length; i++) {
- ServerName regionLocation2 = regionStates.getRegionServerOfRegion(regionsToMerge[i]);
- if (regionLocation2 != null) {
- if (onSameRS) {
- onSameRS = regionLocation.equals(regionLocation2);
- }
- } else {
- // At least one region is not online, merge will fail, no need to continue.
- break;
- }
- }
- if (i == regionsToMerge.length) {
- // Finish checking all regions, return the result;
- return onSameRS;
- }
- }
-
- // If reaching here, at least one region is not online.
- String msg = "Skip merging regions " + getRegionsToMergeListFullNameString() +
- ", because region " + regionsToMerge[i].getEncodedName() + " is not online now.";
- LOG.warn(msg);
- throw new IOException(msg);
- }
-
- /**
- * The procedure could be restarted from a different machine. If the variable is null, we need to
- * retrieve it.
- * @param env MasterProcedureEnv
- * @return assignmentManager
- */
- private AssignmentManager getAssignmentManager(final MasterProcedureEnv env) {
- if (assignmentManager == null) {
- assignmentManager = env.getMasterServices().getAssignmentManager();
- }
- return assignmentManager;
- }
-
- /**
- * The procedure could be restarted from a different machine. If the variable is null, we need to
- * retrieve it.
- * @param env MasterProcedureEnv
- * @return timeout value
- */
- private int getTimeout(final MasterProcedureEnv env) {
- if (timeout == -1) {
- timeout = env.getMasterConfiguration().getInt(
- "hbase.master.regionmerge.timeout", regionsToMerge.length * 60 * 1000);
- }
- return timeout;
- }
-
- /**
- * The procedure could be restarted from a different machine. If the variable is null, we need to
- * retrieve it.
- * @param env MasterProcedureEnv
- * @return serverName
- */
- private ServerName getServerName(final MasterProcedureEnv env) {
- if (regionLocation == null) {
- regionLocation =
- getAssignmentManager(env).getRegionStates().getRegionServerOfRegion(regionsToMerge[0]);
- }
- return regionLocation;
- }
-
- /**
- * The procedure could be restarted from a different machine. If the variable is null, we need to
- * retrieve it.
- * @param fullName whether return only encoded name
- * @return region names in a list
- */
- private String getRegionsToMergeListFullNameString() {
- if (regionsToMergeListFullName == null) {
- StringBuilder sb = new StringBuilder("[");
- int i = 0;
- while(i < regionsToMerge.length - 1) {
- sb.append(regionsToMerge[i].getRegionNameAsString() + ", ");
- i++;
- }
- sb.append(regionsToMerge[i].getRegionNameAsString() + " ]");
- regionsToMergeListFullName = sb.toString();
- }
- return regionsToMergeListFullName;
- }
-
- /**
- * The procedure could be restarted from a different machine. If the variable is null, we need to
- * retrieve it.
- * @return encoded region names
- */
- private String getRegionsToMergeListEncodedNameString() {
- if (regionsToMergeListEncodedName == null) {
- StringBuilder sb = new StringBuilder("[");
- int i = 0;
- while(i < regionsToMerge.length - 1) {
- sb.append(regionsToMerge[i].getEncodedName() + ", ");
- i++;
- }
- sb.append(regionsToMerge[i].getEncodedName() + " ]");
- regionsToMergeListEncodedName = sb.toString();
- }
- return regionsToMergeListEncodedName;
- }
-
- /**
- * The procedure could be restarted from a different machine. If the variable is null, we need to
- * retrieve it.
- * @return traceEnabled
- */
- private Boolean isTraceEnabled() {
- if (traceEnabled == null) {
- traceEnabled = LOG.isTraceEnabled();
- }
- return traceEnabled;
- }
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplitThread.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplitThread.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplitThread.java
index 1331b86..63929a8 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplitThread.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplitThread.java
@@ -217,20 +217,6 @@ public class CompactSplitThread implements CompactionRequestor, PropagatingConfi
return queueLists.toString();
}
- public synchronized void requestRegionsMerge(final Region a,
- final Region b, final boolean forcible, long masterSystemTime, User user) {
- try {
- mergePool.execute(new RegionMergeRequest(a, b, this.server, forcible, masterSystemTime,user));
- if (LOG.isDebugEnabled()) {
- LOG.debug("Region merge requested for " + a + "," + b + ", forcible="
- + forcible + ". " + this);
- }
- } catch (RejectedExecutionException ree) {
- LOG.warn("Could not execute merge for " + a + "," + b + ", forcible="
- + forcible, ree);
- }
- }
-
public synchronized boolean requestSplit(final Region r) {
// don't split regions that are blocking
if (shouldSplitRegion() && ((HRegion)r).getCompactPriority() >= Store.PRIORITY_USER) {
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
index a5172bb..b574c50 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
@@ -6881,112 +6881,6 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
(Bytes.compareTo(info.getEndKey(), 0, info.getEndKey().length, row, offset, length) > 0));
}
- /**
- * Merge two HRegions. The regions must be adjacent and must not overlap.
- *
- * @return new merged HRegion
- * @throws IOException
- */
- public static HRegion mergeAdjacent(final HRegion srcA, final HRegion srcB)
- throws IOException {
- HRegion a = srcA;
- HRegion b = srcB;
-
- // Make sure that srcA comes first; important for key-ordering during
- // write of the merged file.
- if (srcA.getRegionInfo().getStartKey() == null) {
- if (srcB.getRegionInfo().getStartKey() == null) {
- throw new IOException("Cannot merge two regions with null start key");
- }
- // A's start key is null but B's isn't. Assume A comes before B
- } else if ((srcB.getRegionInfo().getStartKey() == null) ||
- (Bytes.compareTo(srcA.getRegionInfo().getStartKey(),
- srcB.getRegionInfo().getStartKey()) > 0)) {
- a = srcB;
- b = srcA;
- }
-
- if (!(Bytes.compareTo(a.getRegionInfo().getEndKey(),
- b.getRegionInfo().getStartKey()) == 0)) {
- throw new IOException("Cannot merge non-adjacent regions");
- }
- return merge(a, b);
- }
-
- /**
- * Merge two regions whether they are adjacent or not.
- *
- * @param a region a
- * @param b region b
- * @return new merged region
- * @throws IOException
- */
- public static HRegion merge(final HRegion a, final HRegion b) throws IOException {
- if (!a.getRegionInfo().getTable().equals(b.getRegionInfo().getTable())) {
- throw new IOException("Regions do not belong to the same table");
- }
-
- FileSystem fs = a.getRegionFileSystem().getFileSystem();
- // Make sure each region's cache is empty
- a.flush(true);
- b.flush(true);
-
- // Compact each region so we only have one store file per family
- a.compact(true);
- if (LOG.isDebugEnabled()) {
- LOG.debug("Files for region: " + a);
- a.getRegionFileSystem().logFileSystemState(LOG);
- }
- b.compact(true);
- if (LOG.isDebugEnabled()) {
- LOG.debug("Files for region: " + b);
- b.getRegionFileSystem().logFileSystemState(LOG);
- }
-
- RegionMergeTransactionImpl rmt = new RegionMergeTransactionImpl(a, b, true);
- if (!rmt.prepare(null)) {
- throw new IOException("Unable to merge regions " + a + " and " + b);
- }
- HRegionInfo mergedRegionInfo = rmt.getMergedRegionInfo();
- LOG.info("starting merge of regions: " + a + " and " + b
- + " into new region " + mergedRegionInfo.getRegionNameAsString()
- + " with start key <"
- + Bytes.toStringBinary(mergedRegionInfo.getStartKey())
- + "> and end key <"
- + Bytes.toStringBinary(mergedRegionInfo.getEndKey()) + ">");
- HRegion dstRegion;
- try {
- dstRegion = (HRegion)rmt.execute(null, null);
- } catch (IOException ioe) {
- rmt.rollback(null, null);
- throw new IOException("Failed merging region " + a + " and " + b
- + ", and successfully rolled back");
- }
- dstRegion.compact(true);
-
- if (LOG.isDebugEnabled()) {
- LOG.debug("Files for new region");
- dstRegion.getRegionFileSystem().logFileSystemState(LOG);
- }
-
- // clear the compacted files if any
- for (Store s : dstRegion.getStores()) {
- s.closeAndArchiveCompactedFiles();
- }
- if (dstRegion.getRegionFileSystem().hasReferences(dstRegion.getTableDesc())) {
- throw new IOException("Merged region " + dstRegion
- + " still has references after the compaction, is compaction canceled?");
- }
-
- // Archiving the 'A' region
- HFileArchiver.archiveRegion(a.getBaseConf(), fs, a.getRegionInfo());
- // Archiving the 'B' region
- HFileArchiver.archiveRegion(b.getBaseConf(), fs, b.getRegionInfo());
-
- LOG.info("merge completed. New region is " + dstRegion);
- return dstRegion;
- }
-
@Override
public Result get(final Get get) throws IOException {
prepareGet(get);
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java
index 7b5f799..1e9f16b 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java
@@ -115,8 +115,6 @@ import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetServerIn
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetServerInfoResponse;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileResponse;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.OpenRegionRequest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.OpenRegionResponse;
@@ -1642,46 +1640,6 @@ public class RSRpcServices implements HBaseRPCErrorHandler,
}
/**
- * Merge regions on the region server.
- *
- * @param controller the RPC controller
- * @param request the request
- * @return merge regions response
- * @throws ServiceException
- */
- @Override
- @QosPriority(priority = HConstants.ADMIN_QOS)
- public MergeRegionsResponse mergeRegions(final RpcController controller,
- final MergeRegionsRequest request) throws ServiceException {
- try {
- checkOpen();
- requestCount.increment();
- Region regionA = getRegion(request.getRegionA());
- Region regionB = getRegion(request.getRegionB());
- boolean forcible = request.getForcible();
- long masterSystemTime = request.hasMasterSystemTime() ? request.getMasterSystemTime() : -1;
- regionA.startRegionOperation(Operation.MERGE_REGION);
- regionB.startRegionOperation(Operation.MERGE_REGION);
- if (regionA.getRegionInfo().getReplicaId() != HRegionInfo.DEFAULT_REPLICA_ID ||
- regionB.getRegionInfo().getReplicaId() != HRegionInfo.DEFAULT_REPLICA_ID) {
- throw new ServiceException(new MergeRegionException("Can't merge non-default replicas"));
- }
- LOG.info("Receiving merging request for " + regionA + ", " + regionB
- + ",forcible=" + forcible);
- regionA.flush(true);
- regionB.flush(true);
- regionServer.compactSplitThread.requestRegionsMerge(regionA, regionB, forcible,
- masterSystemTime, RpcServer.getRequestUser());
- return MergeRegionsResponse.newBuilder().build();
- } catch (DroppedSnapshotException ex) {
- regionServer.abort("Replay of WAL required. Forcing server shutdown", ex);
- throw new ServiceException(ex);
- } catch (IOException ie) {
- throw new ServiceException(ie);
- }
- }
-
- /**
* Open asynchronously a region or a set of regions on the region server.
*
* The opening is coordinated by ZooKeeper, and this method requires the znode to be created
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionMergeRequest.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionMergeRequest.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionMergeRequest.java
deleted file mode 100644
index ce69ad3..0000000
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionMergeRequest.java
+++ /dev/null
@@ -1,153 +0,0 @@
-/**
- * Copyright The Apache Software Foundation
- *
- * Licensed to the Apache Software Foundation (ASF) under one or more
- * contributor license agreements. See the NOTICE file distributed with this
- * work for additional information regarding copyright ownership. The ASF
- * licenses this file to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance with the License.
- * You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
- * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
- * License for the specific language governing permissions and limitations
- * under the License.
- */
-package org.apache.hadoop.hbase.regionserver;
-
-import java.io.IOException;
-
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
-import org.apache.hadoop.hbase.classification.InterfaceAudience;
-import org.apache.hadoop.hbase.DroppedSnapshotException;
-import org.apache.hadoop.hbase.master.TableLockManager.TableLock;
-import org.apache.hadoop.hbase.security.User;
-import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-import org.apache.hadoop.ipc.RemoteException;
-import org.apache.hadoop.util.StringUtils;
-
-import com.google.common.base.Preconditions;
-
-/**
- * Handles processing region merges. Put in a queue, owned by HRegionServer.
- */
-@InterfaceAudience.Private
-class RegionMergeRequest implements Runnable {
- private static final Log LOG = LogFactory.getLog(RegionMergeRequest.class);
- private final HRegion region_a;
- private final HRegion region_b;
- private final HRegionServer server;
- private final boolean forcible;
- private TableLock tableLock;
- private final long masterSystemTime;
- private final User user;
-
- RegionMergeRequest(Region a, Region b, HRegionServer hrs, boolean forcible,
- long masterSystemTime, User user) {
- Preconditions.checkNotNull(hrs);
- this.region_a = (HRegion)a;
- this.region_b = (HRegion)b;
- this.server = hrs;
- this.forcible = forcible;
- this.masterSystemTime = masterSystemTime;
- this.user = user;
- }
-
- @Override
- public String toString() {
- return "MergeRequest,regions:" + region_a + ", " + region_b + ", forcible="
- + forcible;
- }
-
- @Override
- public void run() {
- if (this.server.isStopping() || this.server.isStopped()) {
- LOG.debug("Skipping merge because server is stopping="
- + this.server.isStopping() + " or stopped=" + this.server.isStopped());
- return;
- }
- try {
- final long startTime = EnvironmentEdgeManager.currentTime();
- RegionMergeTransactionImpl mt = new RegionMergeTransactionImpl(region_a,
- region_b, forcible, masterSystemTime);
-
- //acquire a shared read lock on the table, so that table schema modifications
- //do not happen concurrently
- tableLock = server.getTableLockManager().readLock(region_a.getTableDesc().getTableName()
- , "MERGE_REGIONS:" + region_a.getRegionInfo().getRegionNameAsString() + ", " +
- region_b.getRegionInfo().getRegionNameAsString());
- try {
- tableLock.acquire();
- } catch (IOException ex) {
- tableLock = null;
- throw ex;
- }
-
- // If prepare does not return true, for some reason -- logged inside in
- // the prepare call -- we are not ready to merge just now. Just return.
- if (!mt.prepare(this.server)) return;
- try {
- mt.execute(this.server, this.server, this.user);
- } catch (Exception e) {
- if (this.server.isStopping() || this.server.isStopped()) {
- LOG.info(
- "Skip rollback/cleanup of failed merge of " + region_a + " and "
- + region_b + " because server is"
- + (this.server.isStopping() ? " stopping" : " stopped"), e);
- return;
- }
- if (e instanceof DroppedSnapshotException) {
- server.abort("Replay of WAL required. Forcing server shutdown", e);
- return;
- }
- try {
- LOG.warn("Running rollback/cleanup of failed merge of "
- + region_a +" and "+ region_b + "; " + e.getMessage(), e);
- if (mt.rollback(this.server, this.server)) {
- LOG.info("Successful rollback of failed merge of "
- + region_a +" and "+ region_b);
- } else {
- this.server.abort("Abort; we got an error after point-of-no-return"
- + "when merging " + region_a + " and " + region_b);
- }
- } catch (RuntimeException ee) {
- String msg = "Failed rollback of failed merge of "
- + region_a +" and "+ region_b + " -- aborting server";
- // If failed rollback, kill this server to avoid having a hole in
- // table.
- LOG.info(msg, ee);
- this.server.abort(msg);
- }
- return;
- }
- LOG.info("Regions merged, hbase:meta updated, and report to master. region_a="
- + region_a + ", region_b=" + region_b + ",merged region="
- + mt.getMergedRegionInfo().getRegionNameAsString()
- + ". Region merge took "
- + StringUtils.formatTimeDiff(EnvironmentEdgeManager.currentTime(), startTime));
- } catch (IOException ex) {
- ex = ex instanceof RemoteException ? ((RemoteException) ex).unwrapRemoteException() : ex;
- LOG.error("Merge failed " + this, ex);
- server.checkFileSystem();
- } finally {
- releaseTableLock();
- }
- }
-
- protected void releaseTableLock() {
- if (this.tableLock != null) {
- try {
- this.tableLock.release();
- } catch (IOException ex) {
- LOG.error("Could not release the table lock (something is really wrong). "
- + "Aborting this server to avoid holding the lock forever.");
- this.server.abort("Abort; we got an error when releasing the table lock "
- + "on " + region_a.getRegionInfo().getRegionNameAsString());
- }
- }
- }
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/805d39fc/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionMergeTransactionFactory.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionMergeTransactionFactory.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionMergeTransactionFactory.java
deleted file mode 100644
index c844d54..0000000
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionMergeTransactionFactory.java
+++ /dev/null
@@ -1,76 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.hadoop.hbase.regionserver;
-
-import org.apache.hadoop.conf.Configurable;
-import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.hbase.HBaseInterfaceAudience;
-import org.apache.hadoop.hbase.classification.InterfaceAudience;
-import org.apache.hadoop.hbase.classification.InterfaceStability;
-import org.apache.hadoop.hbase.util.ReflectionUtils;
-
-/**
- * A factory for creating RegionMergeTransactions, which execute region split as a "transaction".
- * See {@link RegionMergeTransactionImpl}
- */
-@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.COPROC)
-@InterfaceStability.Evolving
-public class RegionMergeTransactionFactory implements Configurable {
-
- public static final String MERGE_TRANSACTION_IMPL_KEY =
- "hbase.regionserver.merge.transaction.impl";
-
- private Configuration conf;
-
- public RegionMergeTransactionFactory(Configuration conf) {
- this.conf = conf;
- }
-
- @Override
- public Configuration getConf() {
- return conf;
- }
-
- @Override
- public void setConf(Configuration conf) {
- this.conf = conf;
- }
-
- /**
- * Create a merge transaction
- * @param a region a to merge
- * @param b region b to merge
- * @param forcible if false, we will only merge adjacent regions
- * @return transaction instance
- */
- public RegionMergeTransactionImpl create(final Region a, final Region b,
- final boolean forcible) {
- // The implementation class must extend RegionMergeTransactionImpl, not only
- // implement the RegionMergeTransaction interface like you might expect,
- // because various places such as AssignmentManager use static methods
- // from RegionMergeTransactionImpl. Whatever we use for implementation must
- // be compatible, so it's safest to require ? extends RegionMergeTransactionImpl.
- // If not compatible we will throw a runtime exception from here.
- return ReflectionUtils.instantiateWithCustomCtor(
- conf.getClass(MERGE_TRANSACTION_IMPL_KEY, RegionMergeTransactionImpl.class,
- RegionMergeTransactionImpl.class).getName(),
- new Class[] { Region.class, Region.class, boolean.class },
- new Object[] { a, b, forcible });
- }
-
-}