You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2016/09/30 21:56:50 UTC
[01/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Repository: hbase
Updated Branches:
refs/heads/HBASE-16264 35cf55048 -> 32be831ce
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/util/ByteStringer.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/util/ByteStringer.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/util/ByteStringer.java
deleted file mode 100644
index 38cc0bd..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/util/ByteStringer.java
+++ /dev/null
@@ -1,48 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.hadoop.hbase.shaded.util;
-
-import org.apache.hadoop.hbase.classification.InterfaceAudience;
-
-import com.google.protobuf.ByteString;
-import com.google.protobuf.UnsafeByteOperations;
-
-/**
- * Hack to workaround HBASE-10304 issue that keeps bubbling up when a mapreduce context.
- */
-// Depends on protobuf-3.1.0 feature.
-@InterfaceAudience.Private
-public class ByteStringer {
- private ByteStringer() {
- super();
- }
-
- /**
- * Wraps a byte array in a {@link ByteString} without copying it.
- */
- public static ByteString wrap(final byte[] array) {
- return UnsafeByteOperations.unsafeWrap(array);
- }
-
- /**
- * Wraps a subset of a byte array in a {@link ByteString} without copying it.
- */
- public static ByteString wrap(final byte[] array, int offset, int length) {
- return UnsafeByteOperations.unsafeWrap(array, offset, length);
- }
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/protobuf/RSGroup.proto
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/protobuf/RSGroup.proto b/hbase-protocol-shaded/src/main/protobuf/RSGroup.proto
deleted file mode 100644
index d65cad4..0000000
--- a/hbase-protocol-shaded/src/main/protobuf/RSGroup.proto
+++ /dev/null
@@ -1,34 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-package hbase.pb;
-
-option java_package = "org.apache.hadoop.hbase.shaded.protobuf.generated";
-option java_outer_classname = "RSGroupProtos";
-option java_generic_services = true;
-option java_generate_equals_and_hash = true;
-option optimize_for = SPEED;
-
-import "HBase.proto";
-
-message RSGroupInfo {
- required string name = 1;
- repeated ServerName servers = 4;
- repeated TableName tables = 3;
-}
-
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/protobuf/RSGroupAdmin.proto
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/protobuf/RSGroupAdmin.proto b/hbase-protocol-shaded/src/main/protobuf/RSGroupAdmin.proto
deleted file mode 100644
index 550cdcb..0000000
--- a/hbase-protocol-shaded/src/main/protobuf/RSGroupAdmin.proto
+++ /dev/null
@@ -1,136 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-package hbase.pb;
-
-option java_package = "org.apache.hadoop.hbase.shaded.protobuf.generated";
-option java_outer_classname = "RSGroupAdminProtos";
-option java_generic_services = true;
-option java_generate_equals_and_hash = true;
-option optimize_for = SPEED;
-
-import "HBase.proto";
-import "RSGroup.proto";
-
-/** Group level protobufs */
-
-message ListTablesOfRSGroupRequest {
- required string r_s_group_name = 1;
-}
-
-message ListTablesOfRSGroupResponse {
- repeated TableName table_name = 1;
-}
-
-message GetRSGroupInfoRequest {
- required string r_s_group_name = 1;
-}
-
-message GetRSGroupInfoResponse {
- optional RSGroupInfo r_s_group_info = 1;
-}
-
-message GetRSGroupInfoOfTableRequest {
- required TableName table_name = 1;
-}
-
-message GetRSGroupInfoOfTableResponse {
- optional RSGroupInfo r_s_group_info = 1;
-}
-
-message MoveServersRequest {
- required string target_group = 1;
- repeated ServerName servers = 3;
-}
-
-message MoveServersResponse {
-}
-
-message MoveTablesRequest {
- required string target_group = 1;
- repeated TableName table_name = 2;
-}
-
-message MoveTablesResponse {
-}
-
-message AddRSGroupRequest {
- required string r_s_group_name = 1;
-}
-
-message AddRSGroupResponse {
-}
-
-message RemoveRSGroupRequest {
- required string r_s_group_name = 1;
-}
-
-message RemoveRSGroupResponse {
-}
-
-message BalanceRSGroupRequest {
- required string r_s_group_name = 1;
-}
-
-message BalanceRSGroupResponse {
- required bool balanceRan = 1;
-}
-
-message ListRSGroupInfosRequest {
-}
-
-message ListRSGroupInfosResponse {
- repeated RSGroupInfo r_s_group_info = 1;
-}
-
-message GetRSGroupInfoOfServerRequest {
- required ServerName server = 2;
-}
-
-message GetRSGroupInfoOfServerResponse {
- optional RSGroupInfo r_s_group_info = 1;
-}
-
-service RSGroupAdminService {
- rpc GetRSGroupInfo(GetRSGroupInfoRequest)
- returns (GetRSGroupInfoResponse);
-
- rpc GetRSGroupInfoOfTable(GetRSGroupInfoOfTableRequest)
- returns (GetRSGroupInfoOfTableResponse);
-
- rpc GetRSGroupInfoOfServer(GetRSGroupInfoOfServerRequest)
- returns (GetRSGroupInfoOfServerResponse);
-
- rpc MoveServers(MoveServersRequest)
- returns (MoveServersResponse);
-
- rpc MoveTables(MoveTablesRequest)
- returns (MoveTablesResponse);
-
- rpc AddRSGroup(AddRSGroupRequest)
- returns (AddRSGroupResponse);
-
- rpc RemoveRSGroup(RemoveRSGroupRequest)
- returns (RemoveRSGroupResponse);
-
- rpc BalanceRSGroup(BalanceRSGroupRequest)
- returns (BalanceRSGroupResponse);
-
- rpc ListRSGroupInfos(ListRSGroupInfosRequest)
- returns (ListRSGroupInfosResponse);
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-server/src/main/java/org/apache/hadoop/hbase/codec/MessageCodec.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/codec/MessageCodec.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/codec/MessageCodec.java
index 9ce3a33..ea162fc 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/codec/MessageCodec.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/codec/MessageCodec.java
@@ -22,12 +22,12 @@ import java.io.InputStream;
import java.io.OutputStream;
import java.nio.ByteBuffer;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.classification.InterfaceAudience;
import org.apache.hadoop.hbase.io.ByteBufferInputStream;
import org.apache.hadoop.hbase.Cell;
import org.apache.hadoop.hbase.CellUtil;
import org.apache.hadoop.hbase.HBaseInterfaceAudience;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
import org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos;
/**
@@ -47,15 +47,16 @@ public class MessageCodec implements Codec {
CellProtos.Cell.Builder builder = CellProtos.Cell.newBuilder();
// This copies bytes from Cell to ByteString. I don't see anyway around the copy.
// ByteString is final.
- builder.setRow(ByteStringer.wrap(cell.getRowArray(), cell.getRowOffset(),
+ builder.setRow(UnsafeByteOperations.unsafeWrap(cell.getRowArray(), cell.getRowOffset(),
cell.getRowLength()));
- builder.setFamily(ByteStringer.wrap(cell.getFamilyArray(), cell.getFamilyOffset(),
+ builder.setFamily(UnsafeByteOperations.unsafeWrap(cell.getFamilyArray(),
+ cell.getFamilyOffset(),
cell.getFamilyLength()));
- builder.setQualifier(ByteStringer.wrap(cell.getQualifierArray(),
+ builder.setQualifier(UnsafeByteOperations.unsafeWrap(cell.getQualifierArray(),
cell.getQualifierOffset(), cell.getQualifierLength()));
builder.setTimestamp(cell.getTimestamp());
builder.setCellType(CellProtos.CellType.valueOf(cell.getTypeByte()));
- builder.setValue(ByteStringer.wrap(cell.getValueArray(), cell.getValueOffset(),
+ builder.setValue(UnsafeByteOperations.unsafeWrap(cell.getValueArray(), cell.getValueOffset(),
cell.getValueLength()));
CellProtos.Cell pbcell = builder.build();
pbcell.writeDelimitedTo(this.out);
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-server/src/main/java/org/apache/hadoop/hbase/io/Reference.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/Reference.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/Reference.java
index 4d36fc4..62fd3fb 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/Reference.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/Reference.java
@@ -25,12 +25,12 @@ import java.io.IOException;
import java.io.InputStream;
import java.util.Arrays;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.classification.InterfaceAudience;
import org.apache.hadoop.fs.FSDataOutputStream;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hbase.KeyValueUtil;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
import org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos;
import org.apache.hadoop.hbase.util.Bytes;
@@ -196,7 +196,7 @@ public class Reference {
FSProtos.Reference.Builder builder = FSProtos.Reference.newBuilder();
builder.setRange(isTopFileRegion(getFileRegion())?
FSProtos.Reference.Range.TOP: FSProtos.Reference.Range.BOTTOM);
- builder.setSplitkey(ByteStringer.wrap(getSplitKey()));
+ builder.setSplitkey(UnsafeByteOperations.unsafeWrap(getSplitKey()));
return builder.build();
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/FixedFileTrailer.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/FixedFileTrailer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/FixedFileTrailer.java
index df39ea5..7eac9c6 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/FixedFileTrailer.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/FixedFileTrailer.java
@@ -26,13 +26,13 @@ import java.io.DataOutputStream;
import java.io.IOException;
import java.nio.ByteBuffer;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
-import org.apache.hadoop.hbase.CellComparator.MetaCellComparator;
-import org.apache.hadoop.hbase.classification.InterfaceAudience;
import org.apache.hadoop.fs.FSDataInputStream;
import org.apache.hadoop.hbase.CellComparator;
+import org.apache.hadoop.hbase.CellComparator.MetaCellComparator;
import org.apache.hadoop.hbase.KeyValue;
+import org.apache.hadoop.hbase.classification.InterfaceAudience;
import org.apache.hadoop.hbase.io.compress.Compression;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
import org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos;
import org.apache.hadoop.hbase.util.Bytes;
@@ -207,7 +207,7 @@ public class FixedFileTrailer {
.setComparatorClassName(comparatorClassName)
.setCompressionCodec(compressionCodec.ordinal());
if (encryptionKey != null) {
- builder.setEncryptionKey(ByteStringer.wrap(encryptionKey));
+ builder.setEncryptionKey(UnsafeByteOperations.unsafeWrap(encryptionKey));
}
// We need this extra copy unfortunately to determine the final size of the
// delimited output, see use of baos.size() below.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java
index 5fd2de8..d3669f4 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java
@@ -18,9 +18,6 @@
*/
package org.apache.hadoop.hbase.io.hfile;
-import com.google.common.annotations.VisibleForTesting;
-import com.google.common.base.Preconditions;
-
import java.io.ByteArrayInputStream;
import java.io.Closeable;
import java.io.DataInput;
@@ -39,7 +36,6 @@ import java.util.SortedMap;
import java.util.TreeMap;
import java.util.concurrent.atomic.LongAdder;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.conf.Configuration;
@@ -58,6 +54,7 @@ import org.apache.hadoop.hbase.io.FSDataInputStreamWrapper;
import org.apache.hadoop.hbase.io.compress.Compression;
import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding;
import org.apache.hadoop.hbase.protobuf.ProtobufMagic;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair;
@@ -67,6 +64,9 @@ import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.hbase.util.FSUtils;
import org.apache.hadoop.io.Writable;
+import com.google.common.annotations.VisibleForTesting;
+import com.google.common.base.Preconditions;
+
/**
* File format for hbase.
* A file of sorted key/value pairs. Both keys and values are byte arrays.
@@ -746,8 +746,8 @@ public class HFile {
HFileProtos.FileInfoProto.Builder builder = HFileProtos.FileInfoProto.newBuilder();
for (Map.Entry<byte [], byte[]> e: this.map.entrySet()) {
HBaseProtos.BytesBytesPair.Builder bbpBuilder = HBaseProtos.BytesBytesPair.newBuilder();
- bbpBuilder.setFirst(ByteStringer.wrap(e.getKey()));
- bbpBuilder.setSecond(ByteStringer.wrap(e.getValue()));
+ bbpBuilder.setFirst(UnsafeByteOperations.unsafeWrap(e.getKey()));
+ bbpBuilder.setSecond(UnsafeByteOperations.unsafeWrap(e.getValue()));
builder.addMapEntry(bbpBuilder.build());
}
out.write(ProtobufMagic.PB_MAGIC);
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java
index 3854fa6..11788fa 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java
@@ -90,7 +90,7 @@ import org.apache.hadoop.hbase.security.visibility.VisibilityController;
import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString;
import org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController;
import org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
import org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils;
import org.apache.hadoop.hbase.snapshot.SnapshotDescriptionUtils;
import org.apache.hadoop.hbase.util.Bytes;
@@ -919,7 +919,7 @@ public class MasterRpcServices extends RSRpcServices
builder.setException(ForeignExceptionUtil.toProtoForeignException(result.getException()));
}
if (result.hasResultData()) {
- builder.setResult(ByteStringer.wrap(result.getResult()));
+ builder.setResult(UnsafeByteOperations.unsafeWrap(result.getResult()));
}
master.getMasterProcedureExecutor().removeResult(request.getProcId());
} else {
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/DeleteColumnFamilyProcedure.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/DeleteColumnFamilyProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/DeleteColumnFamilyProcedure.java
index aa618db..238ecf2 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/DeleteColumnFamilyProcedure.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/DeleteColumnFamilyProcedure.java
@@ -32,10 +32,10 @@ import org.apache.hadoop.hbase.TableName;
import org.apache.hadoop.hbase.classification.InterfaceAudience;
import org.apache.hadoop.hbase.client.TableState;
import org.apache.hadoop.hbase.master.MasterCoprocessorHost;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos;
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DeleteColumnFamilyState;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.util.Bytes;
/**
@@ -176,7 +176,7 @@ public class DeleteColumnFamilyProcedure
MasterProcedureProtos.DeleteColumnFamilyStateData.newBuilder()
.setUserInfo(MasterProcedureUtil.toProtoUserInfo(getUser()))
.setTableName(ProtobufUtil.toProtoTableName(tableName))
- .setColumnfamilyName(ByteStringer.wrap(familyName));
+ .setColumnfamilyName(UnsafeByteOperations.unsafeWrap(familyName));
if (unmodifiedHTableDescriptor != null) {
deleteCFMsg
.setUnmodifiedTableSchema(ProtobufUtil.convertToTableSchema(unmodifiedHTableDescriptor));
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-server/src/main/java/org/apache/hadoop/hbase/protobuf/ReplicationProtbufUtil.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/protobuf/ReplicationProtbufUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/protobuf/ReplicationProtbufUtil.java
index ba93a17..c301759 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/protobuf/ReplicationProtbufUtil.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/protobuf/ReplicationProtbufUtil.java
@@ -37,12 +37,12 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience;
import org.apache.hadoop.hbase.io.SizedCellScanner;
import org.apache.hadoop.hbase.ipc.HBaseRpcController;
import org.apache.hadoop.hbase.ipc.HBaseRpcControllerImpl;
+import org.apache.hadoop.hbase.regionserver.wal.WALEdit;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService;
import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos;
-import org.apache.hadoop.hbase.regionserver.wal.WALEdit;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.util.Pair;
import org.apache.hadoop.hbase.wal.WAL.Entry;
import org.apache.hadoop.hbase.wal.WALKey;
@@ -112,10 +112,10 @@ public class ReplicationProtbufUtil {
WALProtos.WALKey.Builder keyBuilder = entryBuilder.getKeyBuilder();
WALKey key = entry.getKey();
keyBuilder.setEncodedRegionName(
- ByteStringer.wrap(encodedRegionName == null
+ UnsafeByteOperations.unsafeWrap(encodedRegionName == null
? key.getEncodedRegionName()
: encodedRegionName));
- keyBuilder.setTableName(ByteStringer.wrap(key.getTablename().getName()));
+ keyBuilder.setTableName(UnsafeByteOperations.unsafeWrap(key.getTablename().getName()));
keyBuilder.setLogSequenceNumber(key.getLogSeqNum());
keyBuilder.setWriteTime(key.getWriteTime());
if (key.getNonce() != HConstants.NO_NONCE) {
@@ -136,7 +136,7 @@ public class ReplicationProtbufUtil {
NavigableMap<byte[], Integer> scopes = key.getReplicationScopes();
if (scopes != null && !scopes.isEmpty()) {
for (Map.Entry<byte[], Integer> scope: scopes.entrySet()) {
- scopeBuilder.setFamily(ByteStringer.wrap(scope.getKey()));
+ scopeBuilder.setFamily(UnsafeByteOperations.unsafeWrap(scope.getKey()));
WALProtos.ScopeType scopeType =
WALProtos.ScopeType.valueOf(scope.getValue().intValue());
scopeBuilder.setScopeType(scopeType);
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
index 565c1c0..07d1fa5 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
@@ -129,21 +129,6 @@ import org.apache.hadoop.hbase.ipc.RpcServer;
import org.apache.hadoop.hbase.mob.MobUtils;
import org.apache.hadoop.hbase.monitoring.MonitoredTask;
import org.apache.hadoop.hbase.monitoring.TaskMonitor;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState;
-import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceCall;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionLoad;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.SnapshotDescription;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.CompactionDescriptor;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FlushDescriptor;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FlushDescriptor.FlushAction;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FlushDescriptor.StoreFlushDescriptor;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.RegionEventDescriptor;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.RegionEventDescriptor.EventType;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.StoreDescriptor;
import org.apache.hadoop.hbase.regionserver.MultiVersionConcurrencyControl.WriteEntry;
import org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope;
import org.apache.hadoop.hbase.regionserver.ScannerContext.NextState;
@@ -158,9 +143,24 @@ import org.apache.hadoop.hbase.regionserver.wal.WALUtil;
import org.apache.hadoop.hbase.security.User;
import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString;
import org.apache.hadoop.hbase.shaded.com.google.protobuf.TextFormat;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
+import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceCall;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionLoad;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.SnapshotDescription;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.CompactionDescriptor;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FlushDescriptor;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FlushDescriptor.FlushAction;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FlushDescriptor.StoreFlushDescriptor;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.RegionEventDescriptor;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.RegionEventDescriptor.EventType;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.StoreDescriptor;
import org.apache.hadoop.hbase.snapshot.SnapshotDescriptionUtils;
import org.apache.hadoop.hbase.snapshot.SnapshotManifest;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.hbase.util.CancelableProgressable;
import org.apache.hadoop.hbase.util.ClassSize;
@@ -5553,7 +5553,8 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
try {
WALProtos.BulkLoadDescriptor loadDescriptor =
ProtobufUtil.toBulkLoadDescriptor(this.getRegionInfo().getTable(),
- ByteStringer.wrap(this.getRegionInfo().getEncodedNameAsBytes()), storeFiles,
+ UnsafeByteOperations.unsafeWrap(this.getRegionInfo().getEncodedNameAsBytes()),
+ storeFiles,
storeFilesSizes, seqId);
WALUtil.writeBulkLoadMarkerAndSync(this.wal, this.getReplicationScope(), getRegionInfo(),
loadDescriptor, mvcc);
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java
index a3d0a4b..b07a40c 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java
@@ -127,11 +127,12 @@ import org.apache.hadoop.hbase.security.UserProvider;
import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingRpcChannel;
import org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController;
import org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
+import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
+import org.apache.hadoop.hbase.shaded.protobuf.RequestConverter;
import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceCall;
import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceResponse;
-import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-import org.apache.hadoop.hbase.shaded.protobuf.RequestConverter;
import org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos;
import org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionLoad;
import org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStoreSequenceIds;
@@ -152,7 +153,6 @@ import org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProto
import org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.ReportRSFatalErrorRequest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.ReportRegionStateTransitionRequest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.ReportRegionStateTransitionResponse;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.trace.SpanReceiverHost;
import org.apache.hadoop.hbase.util.Addressing;
import org.apache.hadoop.hbase.util.Bytes;
@@ -1537,7 +1537,7 @@ public class HRegionServer extends HasThread implements
regionSpecifier = RegionSpecifier.newBuilder();
}
regionSpecifier.setType(RegionSpecifierType.REGION_NAME);
- regionSpecifier.setValue(ByteStringer.wrap(name));
+ regionSpecifier.setValue(UnsafeByteOperations.unsafeWrap(name));
regionLoadBldr.setRegionSpecifier(regionSpecifier.build())
.setStores(stores)
.setStorefiles(storefiles)
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/AbstractProtobufLogWriter.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/AbstractProtobufLogWriter.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/AbstractProtobufLogWriter.java
index 569a16a..55c4771 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/AbstractProtobufLogWriter.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/AbstractProtobufLogWriter.java
@@ -41,11 +41,11 @@ import org.apache.hadoop.hbase.io.crypto.Cipher;
import org.apache.hadoop.hbase.io.crypto.Encryption;
import org.apache.hadoop.hbase.io.crypto.Encryptor;
import org.apache.hadoop.hbase.io.util.LRUDictionary;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALHeader;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALTrailer;
import org.apache.hadoop.hbase.security.EncryptionUtil;
import org.apache.hadoop.hbase.security.User;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALHeader;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALTrailer;
import org.apache.hadoop.hbase.util.EncryptionTest;
import org.apache.hadoop.hbase.util.FSUtils;
@@ -112,7 +112,7 @@ public abstract class AbstractProtobufLogWriter {
byte[] keyBytes = new byte[cipher.getKeyLength()];
rng.nextBytes(keyBytes);
Key key = new SecretKeySpec(keyBytes, cipher.getName());
- builder.setEncryptionKey(ByteStringer.wrap(EncryptionUtil.wrapKey(conf,
+ builder.setEncryptionKey(UnsafeByteOperations.unsafeWrap(EncryptionUtil.wrapKey(conf,
conf.get(HConstants.CRYPTO_WAL_KEY_NAME_CONF_KEY,
conf.get(HConstants.CRYPTO_MASTERKEY_NAME_CONF_KEY,
User.getCurrent().getShortName())),
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV1.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV1.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV1.java
index 93c8cfa..cceeebc 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV1.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV1.java
@@ -24,23 +24,23 @@ import java.util.ArrayList;
import java.util.Collection;
import java.util.List;
import java.util.concurrent.Callable;
-import java.util.concurrent.Executor;
import java.util.concurrent.ExecutionException;
+import java.util.concurrent.Executor;
import java.util.concurrent.ExecutorCompletionService;
import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
-import org.apache.hadoop.hbase.classification.InterfaceAudience;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hbase.HRegionInfo;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.SnapshotDescription;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
+import org.apache.hadoop.hbase.classification.InterfaceAudience;
import org.apache.hadoop.hbase.regionserver.HRegionFileSystem;
import org.apache.hadoop.hbase.regionserver.StoreFileInfo;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.SnapshotDescription;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.hbase.util.FSUtils;
@@ -184,7 +184,7 @@ public final class SnapshotManifestV1 {
// 2.1. build the snapshot reference for the store
SnapshotRegionManifest.FamilyFiles.Builder family =
SnapshotRegionManifest.FamilyFiles.newBuilder();
- family.setFamilyName(ByteStringer.wrap(Bytes.toBytes(familyName)));
+ family.setFamilyName(UnsafeByteOperations.unsafeWrap(Bytes.toBytes(familyName)));
if (LOG.isDebugEnabled()) {
LOG.debug("Adding snapshot references for " + storeFiles + " hfiles");
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV2.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV2.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV2.java
index 36e5121..18c2a13 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV2.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV2.java
@@ -19,6 +19,8 @@
package org.apache.hadoop.hbase.snapshot;
import org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
+
import java.io.IOException;
import java.io.InterruptedIOException;
import java.util.ArrayList;
@@ -42,7 +44,6 @@ import org.apache.hadoop.hbase.HRegionInfo;
import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.SnapshotDescription;
import org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
import org.apache.hadoop.hbase.regionserver.StoreFileInfo;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.util.FSUtils;
/**
@@ -101,7 +102,7 @@ public final class SnapshotManifestV2 {
final SnapshotRegionManifest.Builder region, final byte[] familyName) {
SnapshotRegionManifest.FamilyFiles.Builder family =
SnapshotRegionManifest.FamilyFiles.newBuilder();
- family.setFamilyName(ByteStringer.wrap(familyName));
+ family.setFamilyName(UnsafeByteOperations.unsafeWrap(familyName));
return family;
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionInfo.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionInfo.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionInfo.java
index fea65f5..cd76b17 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionInfo.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionInfo.java
@@ -35,9 +35,9 @@ import org.apache.hadoop.hbase.HTableDescriptor;
import org.apache.hadoop.hbase.TableName;
import org.apache.hadoop.hbase.exceptions.DeserializationException;
import org.apache.hadoop.hbase.master.RegionState;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.testclassification.RegionServerTests;
import org.apache.hadoop.hbase.testclassification.SmallTests;
import org.apache.hadoop.hbase.util.Bytes;
@@ -245,11 +245,11 @@ public class TestHRegionInfo {
// test convert RegionInfo without replicaId
RegionInfo info = RegionInfo.newBuilder()
.setTableName(HBaseProtos.TableName.newBuilder()
- .setQualifier(ByteStringer.wrap(tableName.getQualifier()))
- .setNamespace(ByteStringer.wrap(tableName.getNamespace()))
+ .setQualifier(UnsafeByteOperations.unsafeWrap(tableName.getQualifier()))
+ .setNamespace(UnsafeByteOperations.unsafeWrap(tableName.getNamespace()))
.build())
- .setStartKey(ByteStringer.wrap(startKey))
- .setEndKey(ByteStringer.wrap(endKey))
+ .setStartKey(UnsafeByteOperations.unsafeWrap(startKey))
+ .setEndKey(UnsafeByteOperations.unsafeWrap(endKey))
.setSplit(split)
.setRegionId(regionId)
.build();
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionReplayEvents.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionReplayEvents.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionReplayEvents.java
index e1a2486..bcdbff6 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionReplayEvents.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionReplayEvents.java
@@ -64,6 +64,7 @@ import org.apache.hadoop.hbase.client.Put;
import org.apache.hadoop.hbase.executor.ExecutorService;
import org.apache.hadoop.hbase.io.hfile.HFile;
import org.apache.hadoop.hbase.io.hfile.HFileContext;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.MutationProto.MutationType;
import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.BulkLoadDescriptor;
@@ -78,7 +79,6 @@ import org.apache.hadoop.hbase.regionserver.HRegion.FlushResultImpl;
import org.apache.hadoop.hbase.regionserver.HRegion.PrepareFlushResult;
import org.apache.hadoop.hbase.regionserver.throttle.NoLimitThroughputController;
import org.apache.hadoop.hbase.regionserver.wal.WALEdit;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.testclassification.MediumTests;
import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
@@ -1133,11 +1133,13 @@ public class TestHRegionReplayEvents {
putDataByReplay(secondaryRegion, 0, 10, cq, families);
secondaryRegion.replayWALFlushStartMarker(FlushDescriptor.newBuilder().
setFlushSequenceNumber(10)
- .setTableName(ByteStringer.wrap(primaryRegion.getTableDesc().getTableName().getName()))
+ .setTableName(UnsafeByteOperations.unsafeWrap(
+ primaryRegion.getTableDesc().getTableName().getName()))
.setAction(FlushAction.START_FLUSH)
.setEncodedRegionName(
- ByteStringer.wrap(primaryRegion.getRegionInfo().getEncodedNameAsBytes()))
- .setRegionName(ByteStringer.wrap(primaryRegion.getRegionInfo().getRegionName()))
+ UnsafeByteOperations.unsafeWrap(primaryRegion.getRegionInfo().getEncodedNameAsBytes()))
+ .setRegionName(UnsafeByteOperations.unsafeWrap(
+ primaryRegion.getRegionInfo().getRegionName()))
.build());
verify(walSecondary, times(0)).append((HRegionInfo)any(),
@@ -1541,13 +1543,14 @@ public class TestHRegionReplayEvents {
// from primary and also deleted from the archive directory
secondaryRegion.replayWALFlushCommitMarker(FlushDescriptor.newBuilder().
setFlushSequenceNumber(Long.MAX_VALUE)
- .setTableName(ByteStringer.wrap(primaryRegion.getTableDesc().getTableName().getName()))
+ .setTableName(UnsafeByteOperations.unsafeWrap(primaryRegion.getTableDesc().getTableName().getName()))
.setAction(FlushAction.COMMIT_FLUSH)
.setEncodedRegionName(
- ByteStringer.wrap(primaryRegion.getRegionInfo().getEncodedNameAsBytes()))
- .setRegionName(ByteStringer.wrap(primaryRegion.getRegionInfo().getRegionName()))
+ UnsafeByteOperations.unsafeWrap(primaryRegion.getRegionInfo().getEncodedNameAsBytes()))
+ .setRegionName(UnsafeByteOperations.unsafeWrap(
+ primaryRegion.getRegionInfo().getRegionName()))
.addStoreFlushes(StoreFlushDescriptor.newBuilder()
- .setFamilyName(ByteStringer.wrap(families[0]))
+ .setFamilyName(UnsafeByteOperations.unsafeWrap(families[0]))
.setStoreHomeDir("/store_home_dir")
.addFlushOutput("/foo/baz/bar")
.build())
@@ -1559,14 +1562,15 @@ public class TestHRegionReplayEvents {
// tests replaying compaction marker, but the compaction output file has already been compacted
// from primary and also deleted from the archive directory
secondaryRegion.replayWALCompactionMarker(CompactionDescriptor.newBuilder()
- .setTableName(ByteStringer.wrap(primaryRegion.getTableDesc().getTableName().getName()))
+ .setTableName(UnsafeByteOperations.unsafeWrap(
+ primaryRegion.getTableDesc().getTableName().getName()))
.setEncodedRegionName(
- ByteStringer.wrap(primaryRegion.getRegionInfo().getEncodedNameAsBytes()))
- .setFamilyName(ByteStringer.wrap(families[0]))
+ UnsafeByteOperations.unsafeWrap(primaryRegion.getRegionInfo().getEncodedNameAsBytes()))
+ .setFamilyName(UnsafeByteOperations.unsafeWrap(families[0]))
.addCompactionInput("/foo")
.addCompactionOutput("/bar")
.setStoreHomeDir("/store_home_dir")
- .setRegionName(ByteStringer.wrap(primaryRegion.getRegionInfo().getRegionName()))
+ .setRegionName(UnsafeByteOperations.unsafeWrap(primaryRegion.getRegionInfo().getRegionName()))
.build()
, true, true, Long.MAX_VALUE);
}
@@ -1576,15 +1580,16 @@ public class TestHRegionReplayEvents {
// tests replaying region open event marker, but the region files have already been compacted
// from primary and also deleted from the archive directory
secondaryRegion.replayWALRegionEventMarker(RegionEventDescriptor.newBuilder()
- .setTableName(ByteStringer.wrap(primaryRegion.getTableDesc().getTableName().getName()))
+ .setTableName(UnsafeByteOperations.unsafeWrap(
+ primaryRegion.getTableDesc().getTableName().getName()))
.setEncodedRegionName(
- ByteStringer.wrap(primaryRegion.getRegionInfo().getEncodedNameAsBytes()))
- .setRegionName(ByteStringer.wrap(primaryRegion.getRegionInfo().getRegionName()))
+ UnsafeByteOperations.unsafeWrap(primaryRegion.getRegionInfo().getEncodedNameAsBytes()))
+ .setRegionName(UnsafeByteOperations.unsafeWrap(primaryRegion.getRegionInfo().getRegionName()))
.setEventType(EventType.REGION_OPEN)
.setServer(ProtobufUtil.toServerName(ServerName.valueOf("foo", 1, 1)))
.setLogSequenceNumber(Long.MAX_VALUE)
.addStores(StoreDescriptor.newBuilder()
- .setFamilyName(ByteStringer.wrap(families[0]))
+ .setFamilyName(UnsafeByteOperations.unsafeWrap(families[0]))
.setStoreHomeDir("/store_home_dir")
.addStoreFile("/foo")
.build())
@@ -1598,10 +1603,10 @@ public class TestHRegionReplayEvents {
secondaryRegion.replayWALBulkLoadEventMarker(BulkLoadDescriptor.newBuilder()
.setTableName(ProtobufUtil.toProtoTableName(primaryRegion.getTableDesc().getTableName()))
.setEncodedRegionName(
- ByteStringer.wrap(primaryRegion.getRegionInfo().getEncodedNameAsBytes()))
+ UnsafeByteOperations.unsafeWrap(primaryRegion.getRegionInfo().getEncodedNameAsBytes()))
.setBulkloadSeqNum(Long.MAX_VALUE)
.addStores(StoreDescriptor.newBuilder()
- .setFamilyName(ByteStringer.wrap(families[0]))
+ .setFamilyName(UnsafeByteOperations.unsafeWrap(families[0]))
.setStoreHomeDir("/store_home_dir")
.addStoreFile("/foo")
.build())
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestPriorityRpc.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestPriorityRpc.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestPriorityRpc.java
index ed77e11..4c4bae4 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestPriorityRpc.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestPriorityRpc.java
@@ -27,7 +27,6 @@ import java.io.IOException;
import org.apache.hadoop.hbase.security.User;
import org.apache.hadoop.hbase.testclassification.MediumTests;
import org.apache.hadoop.hbase.testclassification.RegionServerTests;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hbase.CoordinatedStateManagerFactory;
import org.apache.hadoop.hbase.HBaseConfiguration;
@@ -48,6 +47,7 @@ import org.junit.experimental.categories.Category;
import org.mockito.Mockito;
import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
/**
* Tests that verify certain RPCs get a higher QoS.
@@ -80,12 +80,13 @@ public class TestPriorityRpc {
GetRequest.Builder getRequestBuilder = GetRequest.newBuilder();
RegionSpecifier.Builder regionSpecifierBuilder = RegionSpecifier.newBuilder();
regionSpecifierBuilder.setType(RegionSpecifierType.REGION_NAME);
- ByteString name = ByteStringer.wrap(HRegionInfo.FIRST_META_REGIONINFO.getRegionName());
+ ByteString name = UnsafeByteOperations.unsafeWrap(
+ HRegionInfo.FIRST_META_REGIONINFO.getRegionName());
regionSpecifierBuilder.setValue(name);
RegionSpecifier regionSpecifier = regionSpecifierBuilder.build();
getRequestBuilder.setRegion(regionSpecifier);
Get.Builder getBuilder = Get.newBuilder();
- getBuilder.setRow(ByteStringer.wrap("somerow".getBytes()));
+ getBuilder.setRow(UnsafeByteOperations.unsafeWrap("somerow".getBytes()));
getRequestBuilder.setGet(getBuilder.build());
GetRequest getRequest = getRequestBuilder.build();
RequestHeader header = headerBuilder.build();
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestReplicationSink.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestReplicationSink.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestReplicationSink.java
index 90e2d80..62a94f1 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestReplicationSink.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestReplicationSink.java
@@ -57,13 +57,13 @@ import org.apache.hadoop.hbase.client.Scan;
import org.apache.hadoop.hbase.client.Table;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.WALEntry;
import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.UUID;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos;
import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALKey;
import org.apache.hadoop.hbase.regionserver.wal.WALEdit;
import org.apache.hadoop.hbase.testclassification.MediumTests;
import org.apache.hadoop.hbase.testclassification.ReplicationTests;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.hbase.util.FSUtils;
import org.apache.hadoop.hbase.util.HFileTestUtil;
@@ -309,7 +309,8 @@ public class TestReplicationSink {
HRegionInfo regionInfo = l.getAllRegionLocations().get(0).getRegionInfo();
loadDescriptor =
ProtobufUtil.toBulkLoadDescriptor(TABLE_NAME1,
- ByteStringer.wrap(regionInfo.getEncodedNameAsBytes()), storeFiles, storeFilesSize, 1);
+ UnsafeByteOperations.unsafeWrap(regionInfo.getEncodedNameAsBytes()),
+ storeFiles, storeFilesSize, 1);
edit = WALEdit.createBulkLoadEvent(regionInfo, loadDescriptor);
}
List<WALEntry> entries = new ArrayList<WALEntry>(1);
@@ -386,9 +387,9 @@ public class TestReplicationSink {
uuidBuilder.setLeastSigBits(HConstants.DEFAULT_CLUSTER_ID.getLeastSignificantBits());
uuidBuilder.setMostSigBits(HConstants.DEFAULT_CLUSTER_ID.getMostSignificantBits());
keyBuilder.setClusterId(uuidBuilder.build());
- keyBuilder.setTableName(ByteStringer.wrap(table.getName()));
+ keyBuilder.setTableName(UnsafeByteOperations.unsafeWrap(table.getName()));
keyBuilder.setWriteTime(System.currentTimeMillis());
- keyBuilder.setEncodedRegionName(ByteStringer.wrap(HConstants.EMPTY_BYTE_ARRAY));
+ keyBuilder.setEncodedRegionName(UnsafeByteOperations.unsafeWrap(HConstants.EMPTY_BYTE_ARRAY));
keyBuilder.setLogSequenceNumber(-1);
builder.setKey(keyBuilder.build());
return builder;
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestReplicationSourceManager.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestReplicationSourceManager.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestReplicationSourceManager.java
index 44821e9..7174d5f 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestReplicationSourceManager.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestReplicationSourceManager.java
@@ -58,6 +58,7 @@ import org.apache.hadoop.hbase.Server;
import org.apache.hadoop.hbase.ServerName;
import org.apache.hadoop.hbase.TableName;
import org.apache.hadoop.hbase.client.ClusterConnection;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.BulkLoadDescriptor;
import org.apache.hadoop.hbase.regionserver.MultiVersionConcurrencyControl;
@@ -72,7 +73,6 @@ import org.apache.hadoop.hbase.replication.ReplicationStateZKBase;
import org.apache.hadoop.hbase.replication.regionserver.ReplicationSourceManager.NodeFailoverWorker;
import org.apache.hadoop.hbase.testclassification.MediumTests;
import org.apache.hadoop.hbase.testclassification.ReplicationTests;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.hbase.util.FSUtils;
import org.apache.hadoop.hbase.util.Pair;
@@ -492,7 +492,7 @@ public abstract class TestReplicationSourceManager {
// 2. Create bulk load descriptor
BulkLoadDescriptor desc =
ProtobufUtil.toBulkLoadDescriptor(hri.getTable(),
- ByteStringer.wrap(hri.getEncodedNameAsBytes()), storeFiles, storeFilesSize, 1);
+ UnsafeByteOperations.unsafeWrap(hri.getEncodedNameAsBytes()), storeFiles, storeFilesSize, 1);
// 3. create bulk load wal edit event
WALEdit logEdit = WALEdit.createBulkLoadEvent(hri, desc);
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestSnapshotManifest.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestSnapshotManifest.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestSnapshotManifest.java
index 9eeb08c..2a610bf 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestSnapshotManifest.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestSnapshotManifest.java
@@ -31,11 +31,11 @@ import org.apache.hadoop.hbase.HBaseTestingUtility;
import org.apache.hadoop.hbase.HColumnDescriptor;
import org.apache.hadoop.hbase.HRegionInfo;
import org.apache.hadoop.hbase.TableName;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.SnapshotDescription;
import org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDataManifest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.testclassification.MasterTests;
import org.apache.hadoop.hbase.testclassification.SmallTests;
import org.apache.hadoop.hbase.util.Bytes;
@@ -87,7 +87,7 @@ public class TestSnapshotManifest {
for (HColumnDescriptor hcd: builder.getTableDescriptor().getFamilies()) {
SnapshotRegionManifest.FamilyFiles.Builder family =
SnapshotRegionManifest.FamilyFiles.newBuilder();
- family.setFamilyName(ByteStringer.wrap(hcd.getName()));
+ family.setFamilyName(UnsafeByteOperations.unsafeWrap(hcd.getName()));
for (int j = 0; j < 100; ++j) {
SnapshotRegionManifest.StoreFile.Builder sfManifest =
SnapshotRegionManifest.StoreFile.newBuilder();
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/pom.xml
----------------------------------------------------------------------
diff --git a/pom.xml b/pom.xml
index 08f5cb3..320486d 100644
--- a/pom.xml
+++ b/pom.xml
@@ -655,6 +655,7 @@
</execution>
</executions>
<configuration>
+ <skipIfEmpty>true</skipIfEmpty>
<excludes>
<exclude>hbase-site.xml</exclude>
<exclude>hdfs-site.xml</exclude>
@@ -2839,7 +2840,7 @@
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-compiler-plugin</artifactId>
- <version>3.1</version>
+ <version>3.5.1</version>
<configuration>
<compilerId>javac-with-errorprone</compilerId>
<forceJavacCompilerUse>true</forceJavacCompilerUse>
[06/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/RegionServerStatusProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/RegionServerStatusProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/RegionServerStatusProtos.java
index 540df5a..b8712b7 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/RegionServerStatusProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/RegionServerStatusProtos.java
@@ -6,17 +6,17 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class RegionServerStatusProtos {
private RegionServerStatusProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
public interface RegionServerStartupRequestOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.RegionServerStartupRequest)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <pre>
@@ -92,18 +92,18 @@ public final class RegionServerStatusProtos {
*
* <code>optional string use_this_hostname_instead = 4;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getUseThisHostnameInsteadBytes();
}
/**
* Protobuf type {@code hbase.pb.RegionServerStartupRequest}
*/
public static final class RegionServerStartupRequest extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.RegionServerStartupRequest)
RegionServerStartupRequestOrBuilder {
// Use RegionServerStartupRequest.newBuilder() to construct.
- private RegionServerStartupRequest(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private RegionServerStartupRequest(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private RegionServerStartupRequest() {
@@ -114,18 +114,18 @@ public final class RegionServerStatusProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private RegionServerStartupRequest(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -157,29 +157,29 @@ public final class RegionServerStatusProtos {
break;
}
case 34: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000008;
useThisHostnameInstead_ = bs;
break;
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.internal_static_hbase_pb_RegionServerStartupRequest_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.internal_static_hbase_pb_RegionServerStartupRequest_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -280,8 +280,8 @@ public final class RegionServerStatusProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
useThisHostnameInstead_ = s;
@@ -296,17 +296,17 @@ public final class RegionServerStatusProtos {
*
* <code>optional string use_this_hostname_instead = 4;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getUseThisHostnameInsteadBytes() {
java.lang.Object ref = useThisHostnameInstead_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
useThisHostnameInstead_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -332,7 +332,7 @@ public final class RegionServerStatusProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeUInt32(1, port_);
@@ -344,7 +344,7 @@ public final class RegionServerStatusProtos {
output.writeUInt64(3, serverCurrentTime_);
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 4, useThisHostnameInstead_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 4, useThisHostnameInstead_);
}
unknownFields.writeTo(output);
}
@@ -355,19 +355,19 @@ public final class RegionServerStatusProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt32Size(1, port_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(2, serverStartCode_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(3, serverCurrentTime_);
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(4, useThisHostnameInstead_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(4, useThisHostnameInstead_);
}
size += unknownFields.getSerializedSize();
memoizedSize = size;
@@ -423,12 +423,12 @@ public final class RegionServerStatusProtos {
}
if (hasServerStartCode()) {
hash = (37 * hash) + SERVER_START_CODE_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getServerStartCode());
}
if (hasServerCurrentTime()) {
hash = (37 * hash) + SERVER_CURRENT_TIME_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getServerCurrentTime());
}
if (hasUseThisHostnameInstead()) {
@@ -441,61 +441,61 @@ public final class RegionServerStatusProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupRequest parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupRequest parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupRequest parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupRequest parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupRequest parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupRequest parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupRequest parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupRequest parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupRequest parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupRequest parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -513,7 +513,7 @@ public final class RegionServerStatusProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -521,15 +521,15 @@ public final class RegionServerStatusProtos {
* Protobuf type {@code hbase.pb.RegionServerStartupRequest}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.RegionServerStartupRequest)
org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupRequestOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.internal_static_hbase_pb_RegionServerStartupRequest_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.internal_static_hbase_pb_RegionServerStartupRequest_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -542,12 +542,12 @@ public final class RegionServerStatusProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -564,7 +564,7 @@ public final class RegionServerStatusProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.internal_static_hbase_pb_RegionServerStartupRequest_descriptor;
}
@@ -610,29 +610,29 @@ public final class RegionServerStatusProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupRequest) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupRequest)other);
} else {
@@ -676,13 +676,13 @@ public final class RegionServerStatusProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupRequest parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupRequest) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -859,8 +859,8 @@ public final class RegionServerStatusProtos {
public java.lang.String getUseThisHostnameInstead() {
java.lang.Object ref = useThisHostnameInstead_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
useThisHostnameInstead_ = s;
@@ -877,17 +877,17 @@ public final class RegionServerStatusProtos {
*
* <code>optional string use_this_hostname_instead = 4;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getUseThisHostnameInsteadBytes() {
java.lang.Object ref = useThisHostnameInstead_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
useThisHostnameInstead_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -928,7 +928,7 @@ public final class RegionServerStatusProtos {
* <code>optional string use_this_hostname_instead = 4;</code>
*/
public Builder setUseThisHostnameInsteadBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -938,12 +938,12 @@ public final class RegionServerStatusProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -961,22 +961,22 @@ public final class RegionServerStatusProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<RegionServerStartupRequest>
- PARSER = new com.google.protobuf.AbstractParser<RegionServerStartupRequest>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RegionServerStartupRequest>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<RegionServerStartupRequest>() {
public RegionServerStartupRequest parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new RegionServerStartupRequest(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<RegionServerStartupRequest> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RegionServerStartupRequest> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<RegionServerStartupRequest> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RegionServerStartupRequest> getParserForType() {
return PARSER;
}
@@ -988,7 +988,7 @@ public final class RegionServerStatusProtos {
public interface RegionServerStartupResponseOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.RegionServerStartupResponse)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <pre>
@@ -1053,11 +1053,11 @@ public final class RegionServerStatusProtos {
* Protobuf type {@code hbase.pb.RegionServerStartupResponse}
*/
public static final class RegionServerStartupResponse extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.RegionServerStartupResponse)
RegionServerStartupResponseOrBuilder {
// Use RegionServerStartupResponse.newBuilder() to construct.
- private RegionServerStartupResponse(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private RegionServerStartupResponse(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private RegionServerStartupResponse() {
@@ -1065,18 +1065,18 @@ public final class RegionServerStatusProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private RegionServerStartupResponse(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -1103,10 +1103,10 @@ public final class RegionServerStatusProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
@@ -1116,12 +1116,12 @@ public final class RegionServerStatusProtos {
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.internal_static_hbase_pb_RegionServerStartupResponse_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.internal_static_hbase_pb_RegionServerStartupResponse_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1214,7 +1214,7 @@ public final class RegionServerStatusProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
for (int i = 0; i < mapEntries_.size(); i++) {
output.writeMessage(1, mapEntries_.get(i));
@@ -1228,7 +1228,7 @@ public final class RegionServerStatusProtos {
size = 0;
for (int i = 0; i < mapEntries_.size(); i++) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(1, mapEntries_.get(i));
}
size += unknownFields.getSerializedSize();
@@ -1271,61 +1271,61 @@ public final class RegionServerStatusProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupResponse parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupResponse parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupResponse parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupResponse parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupResponse parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupResponse parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupResponse parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupResponse parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupResponse parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupResponse parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -1343,7 +1343,7 @@ public final class RegionServerStatusProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1351,15 +1351,15 @@ public final class RegionServerStatusProtos {
* Protobuf type {@code hbase.pb.RegionServerStartupResponse}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.RegionServerStartupResponse)
org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupResponseOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.internal_static_hbase_pb_RegionServerStartupResponse_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.internal_static_hbase_pb_RegionServerStartupResponse_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1372,12 +1372,12 @@ public final class RegionServerStatusProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getMapEntriesFieldBuilder();
}
@@ -1393,7 +1393,7 @@ public final class RegionServerStatusProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.internal_static_hbase_pb_RegionServerStartupResponse_descriptor;
}
@@ -1430,29 +1430,29 @@ public final class RegionServerStatusProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupResponse) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupResponse)other);
} else {
@@ -1482,7 +1482,7 @@ public final class RegionServerStatusProtos {
mapEntries_ = other.mapEntries_;
bitField0_ = (bitField0_ & ~0x00000001);
mapEntriesBuilder_ =
- com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getMapEntriesFieldBuilder() : null;
} else {
mapEntriesBuilder_.addAllMessages(other.mapEntries_);
@@ -1504,13 +1504,13 @@ public final class RegionServerStatusProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupResponse parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupResponse) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -1531,7 +1531,7 @@ public final class RegionServerStatusProtos {
}
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPairOrBuilder> mapEntriesBuilder_;
/**
@@ -1733,7 +1733,7 @@ public final class RegionServerStatusProtos {
java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair> values) {
if (mapEntriesBuilder_ == null) {
ensureMapEntriesIsMutable();
- com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
values, mapEntries_);
onChanged();
} else {
@@ -1873,11 +1873,11 @@ public final class RegionServerStatusProtos {
getMapEntriesBuilderList() {
return getMapEntriesFieldBuilder().getBuilderList();
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPairOrBuilder>
getMapEntriesFieldBuilder() {
if (mapEntriesBuilder_ == null) {
- mapEntriesBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3<
+ mapEntriesBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPairOrBuilder>(
mapEntries_,
((bitField0_ & 0x00000001) == 0x00000001),
@@ -1888,12 +1888,12 @@ public final class RegionServerStatusProtos {
return mapEntriesBuilder_;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -1911,22 +1911,22 @@ public final class RegionServerStatusProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<RegionServerStartupResponse>
- PARSER = new com.google.protobuf.AbstractParser<RegionServerStartupResponse>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RegionServerStartupResponse>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<RegionServerStartupResponse>() {
public RegionServerStartupResponse parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new RegionServerStartupResponse(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<RegionServerStartupResponse> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RegionServerStartupResponse> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<RegionServerStartupResponse> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RegionServerStartupResponse> getParserForType() {
return PARSER;
}
@@ -1938,7 +1938,7 @@ public final class RegionServerStatusProtos {
public interface RegionServerReportRequestOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.RegionServerReportRequest)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required .hbase.pb.ServerName server = 1;</code>
@@ -1982,29 +1982,29 @@ public final class RegionServerStatusProtos {
* Protobuf type {@code hbase.pb.RegionServerReportRequest}
*/
public static final class RegionServerReportRequest extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.RegionServerReportRequest)
RegionServerReportRequestOrBuilder {
// Use RegionServerReportRequest.newBuilder() to construct.
- private RegionServerReportRequest(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private RegionServerReportRequest(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private RegionServerReportRequest() {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private RegionServerReportRequest(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -2048,22 +2048,22 @@ public final class RegionServerStatusProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.internal_static_hbase_pb_RegionServerReportRequest_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.internal_static_hbase_pb_RegionServerReportRequest_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2149,7 +2149,7 @@ public final class RegionServerStatusProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeMessage(1, getServer());
@@ -2166,11 +2166,11 @@ public final class RegionServerStatusProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(1, getServer());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(2, getLoad());
}
size += unknownFields.getSerializedSize();
@@ -2225,61 +2225,61 @@ public final class RegionServerStatusProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportRequest parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportRequest parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportRequest parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportRequest parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportRequest parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportRequest parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportRequest parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportRequest parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportRequest parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportRequest parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -2297,7 +2297,7 @@ public final class RegionServerStatusProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -2305,15 +2305,15 @@ public final class RegionServerStatusProtos {
* Protobuf type {@code hbase.pb.RegionServerReportRequest}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.RegionServerReportRequest)
org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportRequestOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.internal_static_hbase_pb_RegionServerReportRequest_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.internal_static_hbase_pb_RegionServerReportRequest_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2326,12 +2326,12 @@ public final class RegionServerStatusProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getServerFieldBuilder();
getLoadFieldBuilder();
@@ -2354,7 +2354,7 @@ public final class RegionServerStatusProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.internal_static_hbase_pb_RegionServerReportRequest_descriptor;
}
@@ -2400,29 +2400,29 @@ public final class RegionServerStatusProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportRequest) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportRequest)other);
} else {
@@ -2460,13 +2460,13 @@ public final class RegionServerStatusProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportRequest parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportRequest) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -2479,7 +2479,7 @@ public final class RegionServerStatusProtos {
private int bitField0_;
private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName server_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverBuilder_;
/**
* <code>required .hbase.pb.ServerName server = 1;</code>
@@ -2582,11 +2582,11 @@ public final class RegionServerStatusProtos {
/**
* <code>required .hbase.pb.ServerName server = 1;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder>
getServerFieldBuilder() {
if (serverBuilder_ == null) {
- serverBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ serverBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder>(
getServer(),
getParentForChildren(),
@@ -2597,7 +2597,7 @@ public final class RegionServerStatusProtos {
}
private org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ServerLoad load_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ServerLoad, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ServerLoad.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ServerLoadOrBuilder> loadBuilder_;
/**
* <pre>
@@ -2736,11 +2736,11 @@ public final class RegionServerStatusProtos {
*
* <code>optional .hbase.pb.ServerLoad load = 2;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ServerLoad, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ServerLoad.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ServerLoadOrBuilder>
getLoadFieldBuilder() {
if (loadBuilder_ == null) {
- loadBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ loadBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ServerLoad, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ServerLoad.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ServerLoadOrBuilder>(
getLoad(),
getParentForChildren(),
@@ -2750,12 +2750,12 @@ public final class RegionServerStatusProtos {
return loadBuilder_;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -2773,22 +2773,22 @@ public final class RegionServerStatusProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<RegionServerReportRequest>
- PARSER = new com.google.protobuf.AbstractParser<RegionServerReportRequest>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RegionServerReportRequest>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<RegionServerReportRequest>() {
public RegionServerReportRequest parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new RegionServerReportRequest(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<RegionServerReportRequest> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RegionServerReportRequest> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<RegionServerReportRequest> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RegionServerReportRequest> getParserForType() {
return PARSER;
}
@@ -2800,34 +2800,34 @@ public final class RegionServerStatusProtos {
public interface RegionServerReportResponseOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.RegionServerReportResponse)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
}
/**
* Protobuf type {@code hbase.pb.RegionServerReportResponse}
*/
public static final class RegionServerReportResponse extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.RegionServerReportResponse)
RegionServerReportResponseOrBuilder {
// Use RegionServerReportResponse.newBuilder() to construct.
- private RegionServerReportResponse(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private RegionServerReportResponse(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private RegionServerReportResponse() {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private RegionServerReportResponse(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -2845,22 +2845,22 @@ public final class RegionServerStatusProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.internal_static_hbase_pb_RegionServerReportResponse_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.internal_static_hbase_pb_RegionServerReportResponse_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2877,7 +2877,7 @@ public final class RegionServerStatusProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
unknownFields.writeTo(output);
}
@@ -2921,61 +2921,61 @@ public final class RegionServerStatusProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportResponse parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportResponse parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportResponse parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportResponse parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportResponse parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportResponse parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportResponse parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportResponse parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportResponse parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportResponse parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -2993,7 +2993,7 @@ public final class RegionServerStatusProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -3001,15 +3001,15 @@ public final class RegionServerStatusProtos {
* Protobuf type {@code hbase.pb.RegionServerReportResponse}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.RegionServerReportResponse)
org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportResponseOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.internal_static_hbase_pb_RegionServerReportResponse_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.internal_static_hbase_pb_RegionServerReportResponse_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -3022,12 +3022,12 @@ public final class RegionServerStatusProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -3036,7 +3036,7 @@ public final class RegionServerStatusProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.internal_static_hbase_pb_RegionServerReportResponse_descriptor;
}
@@ -3063,29 +3063,29 @@ public final class RegionServerStatusProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportResponse) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportResponse)other);
} else {
@@ -3106,13 +3106,13 @@ public final class RegionServerStatusProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportResponse parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerReportResponse) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -3123,12 +3123,12 @@ public final class RegionServerStatusProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -3146,22 +3146,22 @@ public final class RegionServerStatusProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<RegionServerReportResponse>
- PARSER = new com.google.protobuf.AbstractParser<RegionServerReportResponse>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RegionServerReportResponse>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<RegionServerReportResponse>() {
public RegionServerReportResponse parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new RegionServerReportResponse(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<RegionServerReportResponse> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RegionServerReportResponse> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<RegionServerReportResponse> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RegionServerReportResponse> getParserForType() {
return PARSER;
}
@@ -3173,7 +3173,7 @@ public final class RegionServerStatusProtos {
public interface ReportRSFatalErrorRequestOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.ReportRSFatalErrorRequest)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <pre>
@@ -3223,18 +3223,18 @@ public final class RegionServerStatusProtos {
*
* <code>required string error_message = 2;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getErrorMessageBytes();
}
/**
* Protobuf type {@code hbase.pb.ReportRSFatalErrorRequest}
*/
public static final class ReportRSFatalErrorRequest extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.ReportRSFatalErrorRequest)
ReportRSFatalErrorRequestOrBuilder {
// Use ReportRSFatalErrorRequest.newBuilder() to construct.
- private ReportRSFatalErrorRequest(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private ReportRSFatalErrorRequest(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private ReportRSFatalErrorRequest() {
@@ -3242,18 +3242,18 @@ public final class RegionServerStatusProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private ReportRSFatalErrorRequest(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws
<TRUNCATED>
[28/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/pom.xml
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/pom.xml b/hbase-protocol-shaded/pom.xml
index 1bd0aa6..e0e9eec 100644
--- a/hbase-protocol-shaded/pom.xml
+++ b/hbase-protocol-shaded/pom.xml
@@ -1,6 +1,6 @@
<?xml version="1.0"?>
<project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd">
- <!--
+ <!--
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
@@ -19,112 +19,44 @@
* limitations under the License.
*/
-->
- <modelVersion>4.0.0</modelVersion>
- <parent>
- <artifactId>hbase</artifactId>
- <groupId>org.apache.hbase</groupId>
- <version>2.0.0-SNAPSHOT</version>
- <relativePath>..</relativePath>
- </parent>
-
- <artifactId>hbase-protocol-shaded</artifactId>
- <name>Apache HBase - Shaded Protocol</name>
- <description>Shaded protobuf protocol classes used by HBase internally.</description>
-
- <properties>
- <maven.javadoc.skip>true</maven.javadoc.skip>
- <!--Version of protobuf that hbase uses internally (we shade our pb)
+ <modelVersion>4.0.0</modelVersion>
+ <parent>
+ <artifactId>hbase</artifactId>
+ <groupId>org.apache.hbase</groupId>
+ <version>2.0.0-SNAPSHOT</version>
+ <relativePath>..</relativePath>
+ </parent>
+ <artifactId>hbase-protocol-shaded</artifactId>
+ <name>Apache HBase - Shaded Protocol</name>
+ <description>Shaded protobuf protocol classes used by HBase internally.</description>
+ <properties>
+ <maven.javadoc.skip>true</maven.javadoc.skip>
+ <!--Version of protobuf that hbase uses internally (we shade our pb)
-->
- <internal.protobuf.version>3.1.0</internal.protobuf.version>
- </properties>
-
- <build>
- <plugins>
- <plugin>
- <groupId>org.apache.maven.plugins</groupId>
- <artifactId>maven-shade-plugin</artifactId>
- <version>2.4.3</version>
- <executions>
- <execution>
- <phase>package</phase>
- <goals>
- <goal>shade</goal>
- </goals>
- <configuration>
- <relocations>
- <relocation>
- <pattern>com.google.protobuf</pattern>
- <shadedPattern>org.apache.hadoop.hbase.shaded.com.google.protobuf</shadedPattern>
- </relocation>
- </relocations>
- <!-- What I got when I did a mvn dependency:list for this
- module. Exclude all but the protobuf
- [INFO] commons-logging:commons-logging:jar:1.2:compile
- [INFO] com.github.stephenc.findbugs:findbugs-annotations:jar:1.3.9-1:compile
- [INFO] log4j:log4j:jar:1.2.17:compile
- [INFO] com.google.protobuf:protobuf-java:jar:2.5.0:compile
- [INFO] org.hamcrest:hamcrest-core:jar:1.3:test
- [INFO] org.mockito:mockito-all:jar:1.10.8:test
- [INFO] junit:junit:jar:4.12:compile
- [INFO] org.apache.hbase:hbase-annotations:jar:2.0.0-SNAPSHOT:compile
-
- The list below must exlude all of the above except protobuf.
- -->
- <artifactSet>
- <excludes>
- <exclude>commons-logging:commons-logging</exclude>
- <exclude>com.github.stephenc.findbugs:findbugs-annotations</exclude>
- <exclude>log4j:log4j</exclude>
- <exclude>org.hamcrest:hamcrest-core</exclude>
- <exclude>org.mockito:mockito-all</exclude>
- <exclude>junit:junit</exclude>
- <exclude>org.apache.hbase:hbase-annotations</exclude>
- </excludes>
- </artifactSet>
- </configuration>
- </execution>
- </executions>
- </plugin>
- <plugin>
- <!--Now unpack the shaded jar made above so the shaded classes
- are available to subsequent modules-->
- <groupId>org.apache.maven.plugins</groupId>
- <artifactId>maven-dependency-plugin</artifactId>
- <version>2.10</version>
- <executions>
- <execution>
- <id>unpack</id>
- <phase>package</phase>
- <goals>
- <goal>unpack</goal>
- </goals>
- <configuration>
- <artifactItems>
- <artifactItem>
- <groupId>org.apache.hbase</groupId>
- <artifactId>hbase-protocol-shaded</artifactId>
- <version>${project.version}</version>
- <type>jar</type>
- <overWrite>true</overWrite>
- <outputDirectory>${project.build.directory}/classes</outputDirectory>
- </artifactItem>
- </artifactItems>
- </configuration>
- </execution>
- </executions>
- </plugin>
- <plugin>
- <groupId>org.apache.maven.plugins</groupId>
- <artifactId>maven-site-plugin</artifactId>
- <configuration>
- <skip>true</skip>
- </configuration>
- </plugin>
- <!-- Make a jar and put the sources in the jar -->
- <plugin>
- <groupId>org.apache.maven.plugins</groupId>
- <artifactId>maven-source-plugin</artifactId>
- </plugin>
+ <internal.protobuf.version>3.1.0</internal.protobuf.version>
+ <!--The Default target dir-->
+ <classes.dir>${basedir}/target</classes.dir>
+ <!--The Default location for sources-->
+ <sources.dir>src/main/java</sources.dir>
+ </properties>
+ <build>
+ <!--I want to override these in profile so define them
+ with variables up here-->
+ <sourceDirectory>${sources.dir}</sourceDirectory>
+ <outputDirectory>${classes.dir}</outputDirectory>
+ <plugins>
+ <plugin>
+ <groupId>org.apache.maven.plugins</groupId>
+ <artifactId>maven-site-plugin</artifactId>
+ <configuration>
+ <skip>true</skip>
+ </configuration>
+ </plugin>
+ <!-- Make a jar and put the sources in the jar -->
+ <plugin>
+ <groupId>org.apache.maven.plugins</groupId>
+ <artifactId>maven-source-plugin</artifactId>
+ </plugin>
<plugin>
<!--Make it so assembly:single does nothing in here-->
<artifactId>maven-assembly-plugin</artifactId>
@@ -133,163 +65,283 @@
<skipAssembly>true</skipAssembly>
</configuration>
</plugin>
+ <plugin>
+ <artifactId>maven-surefire-plugin</artifactId>
+ <!-- Always skip the second part executions, since we only run simple unit tests in this module -->
+ <executions>
+ <execution>
+ <id>secondPartTestsExecution</id>
+ <phase>test</phase>
+ <goals>
+ <goal>test</goal>
+ </goals>
+ <configuration>
+ <skip>true</skip>
+ </configuration>
+ </execution>
+ </executions>
+ </plugin>
+ </plugins>
+ <pluginManagement>
+ <plugins>
+ <!--This plugin's configuration is used to store Eclipse m2e settings only. It has no influence on the Maven build itself.-->
<plugin>
- <artifactId>maven-surefire-plugin</artifactId>
- <!-- Always skip the second part executions, since we only run simple unit tests in this module -->
- <executions>
- <execution>
- <id>secondPartTestsExecution</id>
- <phase>test</phase>
+ <groupId>org.eclipse.m2e</groupId>
+ <artifactId>lifecycle-mapping</artifactId>
+ <version>1.0.0</version>
+ <configuration>
+ <lifecycleMappingMetadata>
+ <pluginExecutions>
+ <pluginExecution>
+ <pluginExecutionFilter>
+ <groupId>org.apache.hadoop</groupId>
+ <artifactId>hadoop-maven-plugins</artifactId>
+ <versionRange>[2.0.5-alpha,)</versionRange>
<goals>
- <goal>test</goal>
+ <goal>protoc</goal>
</goals>
- <configuration>
- <skip>true</skip>
- </configuration>
- </execution>
- </executions>
+ </pluginExecutionFilter>
+ <action>
+ <ignore/>
+ </action>
+ </pluginExecution>
+ </pluginExecutions>
+ </lifecycleMappingMetadata>
+ </configuration>
</plugin>
- </plugins>
- <pluginManagement>
+ </plugins>
+ </pluginManagement>
+ </build>
+ <dependencies>
+ <!--BE CAREFUL! Any dependency added here needs to be
+ excluded above in the shade plugin else the dependency
+ will get bundled-->
+ <!-- Intra-project dependencies -->
+ <dependency>
+ <groupId>org.apache.hbase</groupId>
+ <artifactId>hbase-annotations</artifactId>
+ <exclusions>
+ <exclusion>
+ <groupId>jdk.tools</groupId>
+ <artifactId>jdk.tools</artifactId>
+ </exclusion>
+ </exclusions>
+ </dependency>
+ <!-- General dependencies -->
+ <dependency>
+ <groupId>com.google.protobuf</groupId>
+ <artifactId>protobuf-java</artifactId>
+ <version>${internal.protobuf.version}</version>
+ </dependency>
+ <dependency>
+ <groupId>commons-logging</groupId>
+ <artifactId>commons-logging</artifactId>
+ </dependency>
+ </dependencies>
+ <profiles>
+ <!-- Skip the tests in this module -->
+ <profile>
+ <id>skip-protocol-shaded-tests</id>
+ <activation>
+ <property>
+ <name>skip-protocol-shaded-tests</name>
+ </property>
+ </activation>
+ <properties>
+ <surefire.skipFirstPart>true</surefire.skipFirstPart>
+ </properties>
+ </profile>
+ <!--
+ Generate shaded classes using proto files and
+ the protobuf lib we depend on. Drops generated
+ files under src/main/java. Check in the generated
+ files so available at build time. Run this
+ profile/step everytime you change proto
+ files or update the protobuf version. If you add a
+ proto, be sure to add it to the list below in the
+ hadoop-maven-plugin else we won't 'see' it.
+
+ The below first generates java files from protos.
+ We then compile the generated files and make a jar
+ file. The jar file is then passed to the shade plugin
+ which makes a new fat jar that includes the protobuf
+ lib but with all relocated given the
+ org.apache.hadoop.hbase.shaded prefix. The shading
+ step as by-product produces a jar with relocated
+ java source files in it. This jar we then unpack over
+ the src/main/java directory and we're done.
+
+ User is expected to check in the changes if they look
+ good.
+
+ TODO: Patch the protobuf lib using maven-patch-plugin
+ with changes we need.
+ -->
+ <profile>
+ <id>generate-shaded-classes</id>
+ <activation>
+ <property>
+ <name>generate-shaded-classes</name>
+ </property>
+ </activation>
+ <properties>
+ <profile.id>generate-shaded-classes</profile.id>
+ <sources.dir>${project.build.directory}/protoc-generated-sources</sources.dir>
+ <classes.dir>${project.build.directory}/protoc-generated-classes</classes.dir>
+ <!--When the compile for this profile runs, make sure it makes jars that
+ can be related back to this shading profile. Give them a shading prefix.
+ -->
+ <jar.finalName>${profile.id}.${artifactId}-${project.version}</jar.finalName>
+ </properties>
+ <build>
<plugins>
- <!--This plugin's configuration is used to store Eclipse m2e settings only. It has no influence on the Maven build itself.-->
<plugin>
- <groupId>org.eclipse.m2e</groupId>
- <artifactId>lifecycle-mapping</artifactId>
- <version>1.0.0</version>
+ <groupId>org.apache.hadoop</groupId>
+ <artifactId>hadoop-maven-plugins</artifactId>
<configuration>
- <lifecycleMappingMetadata>
- <pluginExecutions>
- <pluginExecution>
- <pluginExecutionFilter>
- <groupId>org.apache.hadoop</groupId>
- <artifactId>hadoop-maven-plugins</artifactId>
- <versionRange>[2.0.5-alpha,)</versionRange>
- <goals>
- <goal>protoc</goal>
- </goals>
- </pluginExecutionFilter>
- <action>
- <ignore></ignore>
- </action>
- </pluginExecution>
- </pluginExecutions>
- </lifecycleMappingMetadata>
+ <protocVersion>${internal.protobuf.version}</protocVersion>
</configuration>
+ <executions>
+ <execution>
+ <id>compile-protoc</id>
+ <phase>generate-sources</phase>
+ <goals>
+ <goal>protoc</goal>
+ </goals>
+ <configuration>
+ <imports>
+ <param>${basedir}/src/main/protobuf</param>
+ </imports>
+ <source>
+ <directory>${basedir}/src/main/protobuf</directory>
+ <!-- Unfortunately, Hadoop plugin does not support *.proto.
+ We have to individually list every proto file here -->
+ <includes>
+ <include>Admin.proto</include>
+ <include>Cell.proto</include>
+ <include>Client.proto</include>
+ <include>ClusterId.proto</include>
+ <include>ClusterStatus.proto</include>
+ <include>Comparator.proto</include>
+ <include>Encryption.proto</include>
+ <include>ErrorHandling.proto</include>
+ <include>FS.proto</include>
+ <include>Filter.proto</include>
+ <include>HBase.proto</include>
+ <include>HFile.proto</include>
+ <include>LoadBalancer.proto</include>
+ <include>MapReduce.proto</include>
+ <include>Master.proto</include>
+ <include>MasterProcedure.proto</include>
+ <include>Procedure.proto</include>
+ <include>Quota.proto</include>
+ <include>RPC.proto</include>
+ <include>RegionNormalizer.proto</include>
+ <include>RegionServerStatus.proto</include>
+ <include>Snapshot.proto</include>
+ <include>Tracing.proto</include>
+ <include>WAL.proto</include>
+ <include>ZooKeeper.proto</include>
+ <include>TestProcedure.proto</include>
+ <include>test.proto</include>
+ <include>test_rpc_service.proto</include>
+ </includes>
+ </source>
+ <output>${sources.dir}</output>
+ </configuration>
+ </execution>
+ </executions>
</plugin>
- </plugins>
- </pluginManagement>
- </build>
-
- <dependencies>
- <!--BE CAREFUL! Any dependency added here needs to be
- excluded above in the shade plugin else the dependency
- will get bundled here-->
- <!-- Intra-project dependencies -->
- <dependency>
- <groupId>org.apache.hbase</groupId>
- <artifactId>hbase-annotations</artifactId>
- <exclusions>
- <exclusion>
- <groupId>jdk.tools</groupId>
- <artifactId>jdk.tools</artifactId>
- </exclusion>
- </exclusions>
- </dependency>
- <!-- General dependencies -->
- <dependency>
- <groupId>com.google.protobuf</groupId>
- <artifactId>protobuf-java</artifactId>
- <version>${internal.protobuf.version}</version>
- </dependency>
- <dependency>
- <groupId>commons-logging</groupId>
- <artifactId>commons-logging</artifactId>
- </dependency>
- </dependencies>
+ <plugin>
+ <groupId>org.apache.maven.plugins</groupId>
+ <artifactId>maven-jar-plugin</artifactId>
+ <version>3.0.2</version>
+ <configuration>
+ <finalName>${jar.finalName}</finalName>
+ </configuration>
+ </plugin>
+ <plugin>
+ <groupId>org.apache.maven.plugins</groupId>
+ <artifactId>maven-shade-plugin</artifactId>
+ <version>2.4.3</version>
+ <executions>
+ <execution>
+ <phase>package</phase>
+ <goals>
+ <goal>shade</goal>
+ </goals>
+ <configuration>
+ <shadeSourcesContent>true</shadeSourcesContent>
+ <createSourcesJar>true</createSourcesJar>
+ <relocations>
+ <relocation>
+ <pattern>com.google.protobuf</pattern>
+ <shadedPattern>org.apache.hadoop.hbase.shaded.com.google.protobuf</shadedPattern>
+ </relocation>
+ </relocations>
+ <!-- What I got when I did a mvn dependency:list for this
+ module. Exclude all but the protobuf
+ [INFO] commons-logging:commons-logging:jar:1.2:compile
+ [INFO] com.github.stephenc.findbugs:findbugs-annotations:jar:1.3.9-1:compile
+ [INFO] log4j:log4j:jar:1.2.17:compile
+ [INFO] com.google.protobuf:protobuf-java:jar:2.5.0:compile
+ [INFO] org.hamcrest:hamcrest-core:jar:1.3:test
+ [INFO] org.mockito:mockito-all:jar:1.10.8:test
+ [INFO] junit:junit:jar:4.12:compile
+ [INFO] org.apache.hbase:hbase-annotations:jar:2.0.0-SNAPSHOT:compile
- <profiles>
- <!-- Skip the tests in this module -->
- <profile>
- <id>skip-protocol-shaded-tests</id>
- <activation>
- <property>
- <name>skip-protocol-shaded-tests</name>
- </property>
- </activation>
- <properties>
- <surefire.skipFirstPart>true</surefire.skipFirstPart>
- </properties>
- </profile>
- <profile>
- <id>compile-protobuf</id>
- <activation>
- <property>
- <name>compile-protobuf</name>
- </property>
- </activation>
- <build>
- <plugins>
- <plugin>
- <groupId>org.apache.hadoop</groupId>
- <artifactId>hadoop-maven-plugins</artifactId>
+ The list below must exlude all of the above except protobuf.
+ -->
+ <artifactSet>
+ <excludes>
+ <exclude>commons-logging:commons-logging</exclude>
+ <exclude>com.github.stephenc.findbugs:findbugs-annotations</exclude>
+ <exclude>log4j:log4j</exclude>
+ <exclude>org.hamcrest:hamcrest-core</exclude>
+ <exclude>org.mockito:mockito-all</exclude>
+ <exclude>junit:junit</exclude>
+ <exclude>org.apache.hbase:hbase-annotations</exclude>
+ </excludes>
+ </artifactSet>
+ </configuration>
+ </execution>
+ </executions>
+ </plugin>
+ <!--Now unpack the shaded jar made above so the shaded classes
+ are available to subsequent modules-->
+ <plugin>
+ <groupId>org.apache.maven.plugins</groupId>
+ <artifactId>maven-dependency-plugin</artifactId>
+ <version>2.10</version>
+ <executions>
+ <execution>
+ <id>unpack</id>
+ <phase>package</phase>
+ <goals>
+ <goal>unpack</goal>
+ </goals>
<configuration>
- <protocVersion>${internal.protobuf.version}</protocVersion>
+ <artifactItems>
+ <artifactItem>
+ <groupId>${project.groupId}</groupId>
+ <artifactId>${project.artifactId}</artifactId>
+ <version>${project.version}</version>
+ <classifier>sources</classifier>
+ <type>jar</type>
+ <overWrite>true</overWrite>
+ <outputDirectory>${default.sources.dir}</outputDirectory>
+ <includes>**/*.java</includes>
+ </artifactItem>
+ </artifactItems>
</configuration>
- <executions>
- <execution>
- <id>compile-protoc</id>
- <phase>generate-sources</phase>
- <goals>
- <goal>protoc</goal>
- </goals>
- <configuration>
- <imports>
- <param>${basedir}/src/main/protobuf</param>
- </imports>
- <source>
- <directory>${basedir}/src/main/protobuf</directory>
- <!-- Unfortunately, Hadoop plugin does not support *.proto.
- We have to individually list every proto file here -->
- <includes>
- <include>Admin.proto</include>
- <include>Cell.proto</include>
- <include>Client.proto</include>
- <include>ClusterId.proto</include>
- <include>ClusterStatus.proto</include>
- <include>Comparator.proto</include>
- <include>Encryption.proto</include>
- <include>ErrorHandling.proto</include>
- <include>FS.proto</include>
- <include>Filter.proto</include>
- <include>HBase.proto</include>
- <include>HFile.proto</include>
- <include>LoadBalancer.proto</include>
- <include>MapReduce.proto</include>
- <include>Master.proto</include>
- <include>MasterProcedure.proto</include>
- <include>Procedure.proto</include>
- <include>Quota.proto</include>
- <include>RPC.proto</include>
- <include>RegionNormalizer.proto</include>
- <include>RegionServerStatus.proto</include>
- <include>Snapshot.proto</include>
- <include>Tracing.proto</include>
- <include>WAL.proto</include>
- <include>ZooKeeper.proto</include>
- <include>TestProcedure.proto</include>
- <include>test.proto</include>
- <include>test_rpc_service.proto</include>
- </includes>
- </source>
- <!--<output>${project.build.directory}/generated-sources/java</output>-->
- <output>${basedir}/src/main/java/</output>
- </configuration>
- </execution>
- </executions>
- </plugin>
- </plugins>
- </build>
- </profile>
- </profiles>
+ </execution>
+ </executions>
+ </plugin>
+ <!--Patch the files here!!!
+ Use maven-patch-plugin
+ -->
+ </plugins>
+ </build>
+ </profile>
+ </profiles>
</project>
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestProcedureProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestProcedureProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestProcedureProtos.java
index 641bde3..2e78905 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestProcedureProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestProcedureProtos.java
@@ -6,17 +6,17 @@ package org.apache.hadoop.hbase.shaded.ipc.protobuf.generated;
public final class TestProcedureProtos {
private TestProcedureProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
public interface TestTableDDLStateDataOrBuilder extends
// @@protoc_insertion_point(interface_extends:TestTableDDLStateData)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required string table_name = 1;</code>
@@ -29,18 +29,18 @@ public final class TestProcedureProtos {
/**
* <code>required string table_name = 1;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getTableNameBytes();
}
/**
* Protobuf type {@code TestTableDDLStateData}
*/
public static final class TestTableDDLStateData extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:TestTableDDLStateData)
TestTableDDLStateDataOrBuilder {
// Use TestTableDDLStateData.newBuilder() to construct.
- private TestTableDDLStateData(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private TestTableDDLStateData(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private TestTableDDLStateData() {
@@ -48,18 +48,18 @@ public final class TestProcedureProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private TestTableDDLStateData(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -76,29 +76,29 @@ public final class TestProcedureProtos {
break;
}
case 10: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000001;
tableName_ = bs;
break;
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.internal_static_TestTableDDLStateData_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.internal_static_TestTableDDLStateData_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -122,8 +122,8 @@ public final class TestProcedureProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
tableName_ = s;
@@ -134,17 +134,17 @@ public final class TestProcedureProtos {
/**
* <code>required string table_name = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getTableNameBytes() {
java.lang.Object ref = tableName_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
tableName_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -162,10 +162,10 @@ public final class TestProcedureProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 1, tableName_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, tableName_);
}
unknownFields.writeTo(output);
}
@@ -176,7 +176,7 @@ public final class TestProcedureProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, tableName_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, tableName_);
}
size += unknownFields.getSerializedSize();
memoizedSize = size;
@@ -221,61 +221,61 @@ public final class TestProcedureProtos {
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -293,7 +293,7 @@ public final class TestProcedureProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -301,15 +301,15 @@ public final class TestProcedureProtos {
* Protobuf type {@code TestTableDDLStateData}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:TestTableDDLStateData)
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateDataOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.internal_static_TestTableDDLStateData_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.internal_static_TestTableDDLStateData_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -322,12 +322,12 @@ public final class TestProcedureProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -338,7 +338,7 @@ public final class TestProcedureProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.internal_static_TestTableDDLStateData_descriptor;
}
@@ -372,29 +372,29 @@ public final class TestProcedureProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData) {
return mergeFrom((org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData)other);
} else {
@@ -423,13 +423,13 @@ public final class TestProcedureProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -454,8 +454,8 @@ public final class TestProcedureProtos {
public java.lang.String getTableName() {
java.lang.Object ref = tableName_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
tableName_ = s;
@@ -468,17 +468,17 @@ public final class TestProcedureProtos {
/**
* <code>required string table_name = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getTableNameBytes() {
java.lang.Object ref = tableName_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
tableName_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -507,7 +507,7 @@ public final class TestProcedureProtos {
* <code>required string table_name = 1;</code>
*/
public Builder setTableNameBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -517,12 +517,12 @@ public final class TestProcedureProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -540,22 +540,22 @@ public final class TestProcedureProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<TestTableDDLStateData>
- PARSER = new com.google.protobuf.AbstractParser<TestTableDDLStateData>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<TestTableDDLStateData>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<TestTableDDLStateData>() {
public TestTableDDLStateData parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new TestTableDDLStateData(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<TestTableDDLStateData> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<TestTableDDLStateData> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<TestTableDDLStateData> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<TestTableDDLStateData> getParserForType() {
return PARSER;
}
@@ -565,17 +565,17 @@ public final class TestProcedureProtos {
}
- private static final com.google.protobuf.Descriptors.Descriptor
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
internal_static_TestTableDDLStateData_descriptor;
private static final
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_TestTableDDLStateData_fieldAccessorTable;
- public static com.google.protobuf.Descriptors.FileDescriptor
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
getDescriptor() {
return descriptor;
}
- private static com.google.protobuf.Descriptors.FileDescriptor
+ private static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
descriptor;
static {
java.lang.String[] descriptorData = {
@@ -584,22 +584,22 @@ public final class TestProcedureProtos {
".hadoop.hbase.shaded.ipc.protobuf.genera" +
"tedB\023TestProcedureProtos\210\001\001"
};
- com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
- new com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
- public com.google.protobuf.ExtensionRegistry assignDescriptors(
- com.google.protobuf.Descriptors.FileDescriptor root) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
descriptor = root;
return null;
}
};
- com.google.protobuf.Descriptors.FileDescriptor
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
.internalBuildGeneratedFileFrom(descriptorData,
- new com.google.protobuf.Descriptors.FileDescriptor[] {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
}, assigner);
internal_static_TestTableDDLStateData_descriptor =
getDescriptor().getMessageTypes().get(0);
internal_static_TestTableDDLStateData_fieldAccessorTable = new
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_TestTableDDLStateData_descriptor,
new java.lang.String[] { "TableName", });
}
[15/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/HBaseProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/HBaseProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/HBaseProtos.java
index 2ae9f9c..0617426 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/HBaseProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/HBaseProtos.java
@@ -6,13 +6,13 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class HBaseProtos {
private HBaseProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
/**
* <pre>
@@ -22,7 +22,7 @@ public final class HBaseProtos {
* Protobuf enum {@code hbase.pb.CompareType}
*/
public enum CompareType
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>LESS = 0;</code>
*/
@@ -108,27 +108,27 @@ public final class HBaseProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<CompareType>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<CompareType>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
CompareType> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<CompareType>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<CompareType>() {
public CompareType findValueByNumber(int number) {
return CompareType.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.getDescriptor().getEnumTypes().get(0);
}
@@ -136,7 +136,7 @@ public final class HBaseProtos {
private static final CompareType[] VALUES = values();
public static CompareType valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -157,7 +157,7 @@ public final class HBaseProtos {
* Protobuf enum {@code hbase.pb.TimeUnit}
*/
public enum TimeUnit
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>NANOSECONDS = 1;</code>
*/
@@ -243,27 +243,27 @@ public final class HBaseProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<TimeUnit>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<TimeUnit>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
TimeUnit> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<TimeUnit>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<TimeUnit>() {
public TimeUnit findValueByNumber(int number) {
return TimeUnit.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.getDescriptor().getEnumTypes().get(1);
}
@@ -271,7 +271,7 @@ public final class HBaseProtos {
private static final TimeUnit[] VALUES = values();
public static TimeUnit valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -290,7 +290,7 @@ public final class HBaseProtos {
public interface TableNameOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.TableName)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required bytes namespace = 1;</code>
@@ -299,7 +299,7 @@ public final class HBaseProtos {
/**
* <code>required bytes namespace = 1;</code>
*/
- com.google.protobuf.ByteString getNamespace();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getNamespace();
/**
* <code>required bytes qualifier = 2;</code>
@@ -308,7 +308,7 @@ public final class HBaseProtos {
/**
* <code>required bytes qualifier = 2;</code>
*/
- com.google.protobuf.ByteString getQualifier();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getQualifier();
}
/**
* <pre>
@@ -319,31 +319,31 @@ public final class HBaseProtos {
* Protobuf type {@code hbase.pb.TableName}
*/
public static final class TableName extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.TableName)
TableNameOrBuilder {
// Use TableName.newBuilder() to construct.
- private TableName(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private TableName(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private TableName() {
- namespace_ = com.google.protobuf.ByteString.EMPTY;
- qualifier_ = com.google.protobuf.ByteString.EMPTY;
+ namespace_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
+ qualifier_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private TableName(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -371,22 +371,22 @@ public final class HBaseProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableName_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableName_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -395,7 +395,7 @@ public final class HBaseProtos {
private int bitField0_;
public static final int NAMESPACE_FIELD_NUMBER = 1;
- private com.google.protobuf.ByteString namespace_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString namespace_;
/**
* <code>required bytes namespace = 1;</code>
*/
@@ -405,12 +405,12 @@ public final class HBaseProtos {
/**
* <code>required bytes namespace = 1;</code>
*/
- public com.google.protobuf.ByteString getNamespace() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getNamespace() {
return namespace_;
}
public static final int QUALIFIER_FIELD_NUMBER = 2;
- private com.google.protobuf.ByteString qualifier_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString qualifier_;
/**
* <code>required bytes qualifier = 2;</code>
*/
@@ -420,7 +420,7 @@ public final class HBaseProtos {
/**
* <code>required bytes qualifier = 2;</code>
*/
- public com.google.protobuf.ByteString getQualifier() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getQualifier() {
return qualifier_;
}
@@ -442,7 +442,7 @@ public final class HBaseProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, namespace_);
@@ -459,11 +459,11 @@ public final class HBaseProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(1, namespace_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(2, qualifier_);
}
size += unknownFields.getSerializedSize();
@@ -518,61 +518,61 @@ public final class HBaseProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -590,7 +590,7 @@ public final class HBaseProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -603,15 +603,15 @@ public final class HBaseProtos {
* Protobuf type {@code hbase.pb.TableName}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.TableName)
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableName_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableName_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -624,25 +624,25 @@ public final class HBaseProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
public Builder clear() {
super.clear();
- namespace_ = com.google.protobuf.ByteString.EMPTY;
+ namespace_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000001);
- qualifier_ = com.google.protobuf.ByteString.EMPTY;
+ qualifier_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000002);
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableName_descriptor;
}
@@ -680,29 +680,29 @@ public final class HBaseProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName)other);
} else {
@@ -735,13 +735,13 @@ public final class HBaseProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -753,7 +753,7 @@ public final class HBaseProtos {
}
private int bitField0_;
- private com.google.protobuf.ByteString namespace_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString namespace_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>required bytes namespace = 1;</code>
*/
@@ -763,13 +763,13 @@ public final class HBaseProtos {
/**
* <code>required bytes namespace = 1;</code>
*/
- public com.google.protobuf.ByteString getNamespace() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getNamespace() {
return namespace_;
}
/**
* <code>required bytes namespace = 1;</code>
*/
- public Builder setNamespace(com.google.protobuf.ByteString value) {
+ public Builder setNamespace(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -788,7 +788,7 @@ public final class HBaseProtos {
return this;
}
- private com.google.protobuf.ByteString qualifier_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString qualifier_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>required bytes qualifier = 2;</code>
*/
@@ -798,13 +798,13 @@ public final class HBaseProtos {
/**
* <code>required bytes qualifier = 2;</code>
*/
- public com.google.protobuf.ByteString getQualifier() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getQualifier() {
return qualifier_;
}
/**
* <code>required bytes qualifier = 2;</code>
*/
- public Builder setQualifier(com.google.protobuf.ByteString value) {
+ public Builder setQualifier(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -823,12 +823,12 @@ public final class HBaseProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -846,22 +846,22 @@ public final class HBaseProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<TableName>
- PARSER = new com.google.protobuf.AbstractParser<TableName>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<TableName>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<TableName>() {
public TableName parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new TableName(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<TableName> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<TableName> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<TableName> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<TableName> getParserForType() {
return PARSER;
}
@@ -873,7 +873,7 @@ public final class HBaseProtos {
public interface TableSchemaOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.TableSchema)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
@@ -970,11 +970,11 @@ public final class HBaseProtos {
* Protobuf type {@code hbase.pb.TableSchema}
*/
public static final class TableSchema extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.TableSchema)
TableSchemaOrBuilder {
// Use TableSchema.newBuilder() to construct.
- private TableSchema(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private TableSchema(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private TableSchema() {
@@ -984,18 +984,18 @@ public final class HBaseProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private TableSchema(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -1053,10 +1053,10 @@ public final class HBaseProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
@@ -1072,12 +1072,12 @@ public final class HBaseProtos {
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableSchema_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableSchema_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1245,7 +1245,7 @@ public final class HBaseProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeMessage(1, getTableName());
@@ -1268,19 +1268,19 @@ public final class HBaseProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(1, getTableName());
}
for (int i = 0; i < attributes_.size(); i++) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(2, attributes_.get(i));
}
for (int i = 0; i < columnFamilies_.size(); i++) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(3, columnFamilies_.get(i));
}
for (int i = 0; i < configuration_.size(); i++) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(4, configuration_.get(i));
}
size += unknownFields.getSerializedSize();
@@ -1344,61 +1344,61 @@ public final class HBaseProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -1416,7 +1416,7 @@ public final class HBaseProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1430,15 +1430,15 @@ public final class HBaseProtos {
* Protobuf type {@code hbase.pb.TableSchema}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.TableSchema)
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableSchema_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableSchema_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1451,12 +1451,12 @@ public final class HBaseProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getTableNameFieldBuilder();
getAttributesFieldBuilder();
@@ -1493,7 +1493,7 @@ public final class HBaseProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableSchema_descriptor;
}
@@ -1558,29 +1558,29 @@ public final class HBaseProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema)other);
} else {
@@ -1613,7 +1613,7 @@ public final class HBaseProtos {
attributes_ = other.attributes_;
bitField0_ = (bitField0_ & ~0x00000002);
attributesBuilder_ =
- com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getAttributesFieldBuilder() : null;
} else {
attributesBuilder_.addAllMessages(other.attributes_);
@@ -1639,7 +1639,7 @@ public final class HBaseProtos {
columnFamilies_ = other.columnFamilies_;
bitField0_ = (bitField0_ & ~0x00000004);
columnFamiliesBuilder_ =
- com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getColumnFamiliesFieldBuilder() : null;
} else {
columnFamiliesBuilder_.addAllMessages(other.columnFamilies_);
@@ -1665,7 +1665,7 @@ public final class HBaseProtos {
configuration_ = other.configuration_;
bitField0_ = (bitField0_ & ~0x00000008);
configurationBuilder_ =
- com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getConfigurationFieldBuilder() : null;
} else {
configurationBuilder_.addAllMessages(other.configuration_);
@@ -1702,13 +1702,13 @@ public final class HBaseProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -1721,7 +1721,7 @@ public final class HBaseProtos {
private int bitField0_;
private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
@@ -1824,11 +1824,11 @@ public final class HBaseProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>
getTableNameFieldBuilder() {
if (tableNameBuilder_ == null) {
- tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ tableNameBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
getTableName(),
getParentForChildren(),
@@ -1847,7 +1847,7 @@ public final class HBaseProtos {
}
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder> attributesBuilder_;
/**
@@ -1979,7 +1979,7 @@ public final class HBaseProtos {
java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair> values) {
if (attributesBuilder_ == null) {
ensureAttributesIsMutable();
- com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
values, attributes_);
onChanged();
} else {
@@ -2063,11 +2063,11 @@ public final class HBaseProtos {
getAttributesBuilderList() {
return getAttributesFieldBuilder().getBuilderList();
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder>
getAttributesFieldBuilder() {
if (attributesBuilder_ == null) {
- attributesBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3<
+ attributesBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder>(
attributes_,
((bitField0_ & 0x00000002) == 0x00000002),
@@ -2087,7 +2087,7 @@ public final class HBaseProtos {
}
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> columnFamiliesBuilder_;
/**
@@ -2219,7 +2219,7 @@ public final class HBaseProtos {
java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema> values) {
if (columnFamiliesBuilder_ == null) {
ensureColumnFamiliesIsMutable();
- com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
values, columnFamilies_);
onChanged();
} else {
@@ -2303,11 +2303,11 @@ public final class HBaseProtos {
getColumnFamiliesBuilderList() {
return getColumnFamiliesFieldBuilder().getBuilderList();
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder>
getColumnFamiliesFieldBuilder() {
if (columnFamiliesBuilder_ == null) {
- columnFamiliesBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3<
+ columnFamiliesBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder>(
columnFamilies_,
((bitField0_ & 0x00000004) == 0x00000004),
@@ -2327,7 +2327,7 @@ public final class HBaseProtos {
}
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPairOrBuilder> configurationBuilder_;
/**
@@ -2459,7 +2459,7 @@ public final class HBaseProtos {
java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair> values) {
if (configurationBuilder_ == null) {
ensureConfigurationIsMutable();
- com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
values, configuration_);
onChanged();
} else {
@@ -2543,11 +2543,11 @@ public final class HBaseProtos {
getConfigurationBuilderList() {
return getConfigurationFieldBuilder().getBuilderList();
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPairOrBuilder>
getConfigurationFieldBuilder() {
if (configurationBuilder_ == null) {
- configurationBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3<
+ configurationBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPairOrBuilder>(
configuration_,
((bitField0_ & 0x00000008) == 0x00000008),
@@ -2558,12 +2558,12 @@ public final class HBaseProtos {
return configurationBuilder_;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -2581,22 +2581,22 @@ public final class HBaseProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<TableSchema>
- PARSER = new com.google.protobuf.AbstractParser<TableSchema>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<TableSchema>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<TableSchema>() {
public TableSchema parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new TableSchema(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<TableSchema> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<TableSchema> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<TableSchema> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<TableSchema> getParserForType() {
return PARSER;
}
@@ -2608,7 +2608,7 @@ public final class HBaseProtos {
public interface TableStateOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.TableState)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <pre>
@@ -2635,11 +2635,11 @@ public final class HBaseProtos {
* Protobuf type {@code hbase.pb.TableState}
*/
public static final class TableState extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.TableState)
TableStateOrBuilder {
// Use TableState.newBuilder() to construct.
- private TableState(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private TableState(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private TableState() {
@@ -2647,18 +2647,18 @@ public final class HBaseProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private TableState(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -2687,22 +2687,22 @@ public final class HBaseProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableState_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableState_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2717,7 +2717,7 @@ public final class HBaseProtos {
* Protobuf enum {@code hbase.pb.TableState.State}
*/
public enum State
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>ENABLED = 0;</code>
*/
@@ -2776,27 +2776,27 @@ public final class HBaseProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<State>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<State>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
State> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<State>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<State>() {
public State findValueByNumber(int number) {
return State.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.getDescriptor().getEnumTypes().get(0);
}
@@ -2804,7 +2804,7 @@ public final class HBaseProtos {
private static final State[] VALUES = values();
public static State valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -2860,7 +2860,7 @@ public final class HBaseProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeEnum(1, state_);
@@ -2874,7 +2874,7 @@ public final class HBaseProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeEnumSize(1, state_);
}
size += unknownFields.getSerializedSize();
@@ -2919,61 +2919,61 @@ public final class HBaseProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -2991,7 +2991,7 @@ public final class HBaseProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -3003,15 +3003,15 @@ public final class HBaseProtos {
* Protobuf type {@code hbase.pb.TableState}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.TableState)
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableStateOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableState_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableState_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -3024,12 +3024,12 @@ public final class HBaseProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -3040,7 +3040,7 @@ public final class HBaseProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableState_descriptor;
}
@@ -3074,29 +3074,29 @@ public final class HBaseProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState)other);
} else {
@@ -3123,13 +3123,13 @@ public final class HBaseProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -3193,12 +3193,12 @@ public final class HBaseProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -3216,22 +3216,22 @@ public final class HBaseProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<TableState>
- PARSER = new com.google.protobuf.AbstractParser<TableState>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<TableState>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<TableState>() {
public TableState parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new TableState(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<TableState> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<TableState> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<TableState> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<TableState> getParserForType() {
return PARSER;
}
@@ -3243,7 +3243,7 @@ public final class HBaseProtos {
public interface ColumnFamilySchemaOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.ColumnFamilySchema)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required bytes name = 1;</code>
@@ -3252,7 +3252,7 @@ public final class HBaseProtos {
/**
* <code>required bytes name = 1;</code>
*/
- com.google.protobuf.ByteString getName();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getName();
/**
* <code>repeated .hbase.pb.BytesBytesPair attributes = 2;</code>
@@ -3312,32 +3312,32 @@ public final class HBaseProtos {
* Protobuf type {@code hbase.pb.ColumnFamilySchema}
*/
public static final class ColumnFamilySchema extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.ColumnFamilySchema)
ColumnFamilySchemaOrBuilder {
// Use ColumnFamilySchema.newBuilder() to construct.
- private ColumnFamilySchema(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private ColumnFamilySchema(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private ColumnFamilySchema() {
- name_ = com.google.protobuf.ByteString.EMPTY;
+ name_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
attributes_ = java.util.Collections.emptyList();
configuration_ = java.util.Collections.emptyList();
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private ColumnFamilySchema(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -3378,10 +3378,10 @@ public final class HBaseProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
@@ -3394,12 +3394,12 @@ public final class HBaseProtos {
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_ColumnFamilySchema_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_ColumnFamilySchema_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -3408,7 +3408,7 @@ public final class HBaseProtos {
private int bitField0_;
public static final int NAME_FIELD_NUMBER = 1;
- private com.google.protobuf.ByteString name_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString name_;
/**
* <code>required bytes name = 1;</code>
*/
@@ -3418,7 +3418,7 @@ public final class HBaseProtos {
/**
* <code>required bytes name = 1;</code>
*/
- public com.google.protobuf.ByteString getName() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getName() {
return name_;
}
@@ -3518,7 +3518,7 @@ public final class HBaseProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, name_);
@@ -3538,15 +3538,15 @@ public final class HBaseProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(1, name_);
}
for (int i = 0; i < attributes_.size(); i++) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(2, attributes_.get(i));
}
for (int i = 0; i < configuration_.size(); i++) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(3, configuration_.get(i));
}
size += unknownFields.getSerializedSize();
@@ -3604,61 +3604,61 @@ public final class HBaseProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -3676,7 +3676,7 @@ public final class HBaseProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -3690,15 +3690,15 @@ public final class HBaseProtos {
* Protobuf type {@code hbase.pb.ColumnFamilySchema}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.ColumnFamilySchema)
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbas
<TRUNCATED>
[05/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/SnapshotProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/SnapshotProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/SnapshotProtos.java
index 4d3b294..d40086b 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/SnapshotProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/SnapshotProtos.java
@@ -6,17 +6,17 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class SnapshotProtos {
private SnapshotProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
public interface SnapshotFileInfoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.SnapshotFileInfo)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required .hbase.pb.SnapshotFileInfo.Type type = 1;</code>
@@ -38,7 +38,7 @@ public final class SnapshotProtos {
/**
* <code>optional string hfile = 3;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getHfileBytes();
/**
@@ -52,7 +52,7 @@ public final class SnapshotProtos {
/**
* <code>optional string wal_server = 4;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getWalServerBytes();
/**
@@ -66,18 +66,18 @@ public final class SnapshotProtos {
/**
* <code>optional string wal_name = 5;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getWalNameBytes();
}
/**
* Protobuf type {@code hbase.pb.SnapshotFileInfo}
*/
public static final class SnapshotFileInfo extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.SnapshotFileInfo)
SnapshotFileInfoOrBuilder {
// Use SnapshotFileInfo.newBuilder() to construct.
- private SnapshotFileInfo(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private SnapshotFileInfo(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private SnapshotFileInfo() {
@@ -88,18 +88,18 @@ public final class SnapshotProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private SnapshotFileInfo(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -127,41 +127,41 @@ public final class SnapshotProtos {
break;
}
case 26: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000002;
hfile_ = bs;
break;
}
case 34: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000004;
walServer_ = bs;
break;
}
case 42: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000008;
walName_ = bs;
break;
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotFileInfo_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotFileInfo_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -172,7 +172,7 @@ public final class SnapshotProtos {
* Protobuf enum {@code hbase.pb.SnapshotFileInfo.Type}
*/
public enum Type
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>HFILE = 1;</code>
*/
@@ -213,27 +213,27 @@ public final class SnapshotProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<Type>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<Type>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
Type> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<Type>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<Type>() {
public Type findValueByNumber(int number) {
return Type.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo.getDescriptor().getEnumTypes().get(0);
}
@@ -241,7 +241,7 @@ public final class SnapshotProtos {
private static final Type[] VALUES = values();
public static Type valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -291,8 +291,8 @@ public final class SnapshotProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
hfile_ = s;
@@ -303,17 +303,17 @@ public final class SnapshotProtos {
/**
* <code>optional string hfile = 3;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getHfileBytes() {
java.lang.Object ref = hfile_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
hfile_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -333,8 +333,8 @@ public final class SnapshotProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
walServer_ = s;
@@ -345,17 +345,17 @@ public final class SnapshotProtos {
/**
* <code>optional string wal_server = 4;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getWalServerBytes() {
java.lang.Object ref = walServer_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
walServer_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -375,8 +375,8 @@ public final class SnapshotProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
walName_ = s;
@@ -387,17 +387,17 @@ public final class SnapshotProtos {
/**
* <code>optional string wal_name = 5;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getWalNameBytes() {
java.lang.Object ref = walName_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
walName_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -415,19 +415,19 @@ public final class SnapshotProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeEnum(1, type_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 3, hfile_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 3, hfile_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 4, walServer_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 4, walServer_);
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 5, walName_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 5, walName_);
}
unknownFields.writeTo(output);
}
@@ -438,17 +438,17 @@ public final class SnapshotProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeEnumSize(1, type_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(3, hfile_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(3, hfile_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(4, walServer_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(4, walServer_);
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(5, walName_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(5, walName_);
}
size += unknownFields.getSerializedSize();
memoizedSize = size;
@@ -519,61 +519,61 @@ public final class SnapshotProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -591,7 +591,7 @@ public final class SnapshotProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -599,15 +599,15 @@ public final class SnapshotProtos {
* Protobuf type {@code hbase.pb.SnapshotFileInfo}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.SnapshotFileInfo)
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfoOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotFileInfo_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotFileInfo_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -620,12 +620,12 @@ public final class SnapshotProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -642,7 +642,7 @@ public final class SnapshotProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotFileInfo_descriptor;
}
@@ -688,29 +688,29 @@ public final class SnapshotProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo)other);
} else {
@@ -752,13 +752,13 @@ public final class SnapshotProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -819,8 +819,8 @@ public final class SnapshotProtos {
public java.lang.String getHfile() {
java.lang.Object ref = hfile_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
hfile_ = s;
@@ -833,17 +833,17 @@ public final class SnapshotProtos {
/**
* <code>optional string hfile = 3;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getHfileBytes() {
java.lang.Object ref = hfile_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
hfile_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -872,7 +872,7 @@ public final class SnapshotProtos {
* <code>optional string hfile = 3;</code>
*/
public Builder setHfileBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -895,8 +895,8 @@ public final class SnapshotProtos {
public java.lang.String getWalServer() {
java.lang.Object ref = walServer_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
walServer_ = s;
@@ -909,17 +909,17 @@ public final class SnapshotProtos {
/**
* <code>optional string wal_server = 4;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getWalServerBytes() {
java.lang.Object ref = walServer_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
walServer_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -948,7 +948,7 @@ public final class SnapshotProtos {
* <code>optional string wal_server = 4;</code>
*/
public Builder setWalServerBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -971,8 +971,8 @@ public final class SnapshotProtos {
public java.lang.String getWalName() {
java.lang.Object ref = walName_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
walName_ = s;
@@ -985,17 +985,17 @@ public final class SnapshotProtos {
/**
* <code>optional string wal_name = 5;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getWalNameBytes() {
java.lang.Object ref = walName_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
walName_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -1024,7 +1024,7 @@ public final class SnapshotProtos {
* <code>optional string wal_name = 5;</code>
*/
public Builder setWalNameBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1034,12 +1034,12 @@ public final class SnapshotProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -1057,22 +1057,22 @@ public final class SnapshotProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<SnapshotFileInfo>
- PARSER = new com.google.protobuf.AbstractParser<SnapshotFileInfo>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<SnapshotFileInfo>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<SnapshotFileInfo>() {
public SnapshotFileInfo parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new SnapshotFileInfo(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<SnapshotFileInfo> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<SnapshotFileInfo> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<SnapshotFileInfo> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<SnapshotFileInfo> getParserForType() {
return PARSER;
}
@@ -1084,7 +1084,7 @@ public final class SnapshotProtos {
public interface SnapshotRegionManifestOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.SnapshotRegionManifest)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>optional int32 version = 1;</code>
@@ -1136,11 +1136,11 @@ public final class SnapshotProtos {
* Protobuf type {@code hbase.pb.SnapshotRegionManifest}
*/
public static final class SnapshotRegionManifest extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.SnapshotRegionManifest)
SnapshotRegionManifestOrBuilder {
// Use SnapshotRegionManifest.newBuilder() to construct.
- private SnapshotRegionManifest(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private SnapshotRegionManifest(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private SnapshotRegionManifest() {
@@ -1149,18 +1149,18 @@ public final class SnapshotProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private SnapshotRegionManifest(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -1205,10 +1205,10 @@ public final class SnapshotProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
@@ -1218,12 +1218,12 @@ public final class SnapshotProtos {
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotRegionManifest_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotRegionManifest_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1232,7 +1232,7 @@ public final class SnapshotProtos {
public interface StoreFileOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.SnapshotRegionManifest.StoreFile)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required string name = 1;</code>
@@ -1245,7 +1245,7 @@ public final class SnapshotProtos {
/**
* <code>required string name = 1;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getNameBytes();
/**
@@ -1282,11 +1282,11 @@ public final class SnapshotProtos {
* Protobuf type {@code hbase.pb.SnapshotRegionManifest.StoreFile}
*/
public static final class StoreFile extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.SnapshotRegionManifest.StoreFile)
StoreFileOrBuilder {
// Use StoreFile.newBuilder() to construct.
- private StoreFile(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private StoreFile(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private StoreFile() {
@@ -1295,18 +1295,18 @@ public final class SnapshotProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private StoreFile(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -1323,7 +1323,7 @@ public final class SnapshotProtos {
break;
}
case 10: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000001;
name_ = bs;
break;
@@ -1348,22 +1348,22 @@ public final class SnapshotProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotRegionManifest_StoreFile_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotRegionManifest_StoreFile_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1387,8 +1387,8 @@ public final class SnapshotProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
name_ = s;
@@ -1399,17 +1399,17 @@ public final class SnapshotProtos {
/**
* <code>required string name = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getNameBytes() {
java.lang.Object ref = name_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
name_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -1477,10 +1477,10 @@ public final class SnapshotProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 1, name_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, name_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
output.writeMessage(2, getReference());
@@ -1497,14 +1497,14 @@ public final class SnapshotProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, name_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, name_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(2, getReference());
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(3, fileSize_);
}
size += unknownFields.getSerializedSize();
@@ -1560,7 +1560,7 @@ public final class SnapshotProtos {
}
if (hasFileSize()) {
hash = (37 * hash) + FILE_SIZE_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getFileSize());
}
hash = (29 * hash) + unknownFields.hashCode();
@@ -1569,61 +1569,61 @@ public final class SnapshotProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -1641,7 +1641,7 @@ public final class SnapshotProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1649,15 +1649,15 @@ public final class SnapshotProtos {
* Protobuf type {@code hbase.pb.SnapshotRegionManifest.StoreFile}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.SnapshotRegionManifest.StoreFile)
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFileOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotRegionManifest_StoreFile_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotRegionManifest_StoreFile_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1670,12 +1670,12 @@ public final class SnapshotProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getReferenceFieldBuilder();
}
@@ -1695,7 +1695,7 @@ public final class SnapshotProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotRegionManifest_StoreFile_descriptor;
}
@@ -1741,29 +1741,29 @@ public final class SnapshotProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile)other);
} else {
@@ -1803,13 +1803,13 @@ public final class SnapshotProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -1834,8 +1834,8 @@ public final class SnapshotProtos {
public java.lang.String getName() {
java.lang.Object ref = name_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
name_ = s;
@@ -1848,17 +1848,17 @@ public final class SnapshotProtos {
/**
* <code>required string name = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getNameBytes() {
java.lang.Object ref = name_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
name_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -1887,7 +1887,7 @@ public final class SnapshotProtos {
* <code>required string name = 1;</code>
*/
public Builder setNameBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1898,7 +1898,7 @@ public final class SnapshotProtos {
}
private org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference reference_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference, org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.ReferenceOrBuilder> referenceBuilder_;
/**
* <code>optional .hbase.pb.Reference reference = 2;</code>
@@ -2001,11 +2001,11 @@ public final class SnapshotProtos {
/**
* <code>optional .hbase.pb.Reference reference = 2;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference, org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.ReferenceOrBuilder>
getReferenceFieldBuilder() {
if (referenceBuilder_ == null) {
- referenceBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ referenceBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference, org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.ReferenceOrBuilder>(
getReference(),
getParentForChildren(),
@@ -2063,12 +2063,12 @@ public final class SnapshotProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -2086,22 +2086,22 @@ public final class SnapshotProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<StoreFile>
- PARSER = new com.google.protobuf.AbstractParser<StoreFile>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<StoreFile>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<StoreFile>() {
public StoreFile parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new StoreFile(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<StoreFile> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<StoreFile> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<StoreFile> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<StoreFile> getParserForType() {
return PARSER;
}
@@ -2113,7 +2113,7 @@ public final class SnapshotProtos {
public interface FamilyFilesOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.SnapshotRegionManifest.FamilyFiles)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required bytes family_name = 1;</code>
@@ -2122,7 +2122,7 @@ public final class SnapshotProtos {
/**
* <code>required bytes family_name = 1;</code>
*/
- com.google.protobuf.ByteString getFamilyName();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamilyName();
/**
* <code>repeated .hbase.pb.SnapshotRegionManifest.StoreFile store_files = 2;</code>
@@ -2152,31 +2152,31 @@ public final class SnapshotProtos {
* Protobuf type {@code hbase.pb.SnapshotRegionManifest.FamilyFiles}
*/
public static final class FamilyFiles extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.SnapshotRegionManifest.FamilyFiles)
FamilyFilesOrBuilder {
// Use FamilyFiles.newBuilder() to construct.
- private FamilyFiles(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private FamilyFiles(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private FamilyFiles() {
- familyName_ = com.google.protobuf.ByteString.EMPTY;
+ familyName_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
storeFiles_ = java.util.Collections.emptyList();
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private FamilyFiles(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -2208,10 +2208,10 @@ public final class SnapshotProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
@@ -2221,12 +2221,12 @@ public final class SnapshotProtos {
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotRegionManifest_FamilyFiles_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotRegionManifest_FamilyFiles_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2235,7 +2235,7 @@ public final class SnapshotProtos {
private int bitField0_;
public static final int FAMILY_NAME_FIELD_NUMBER = 1;
- private com.google.protobuf.ByteString familyName_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString familyName_;
/**
* <code>required bytes family_name = 1;</code>
*/
@@ -2245,7 +2245,7 @@ public final class SnapshotProtos {
/**
* <code>required bytes family_name = 1;</code>
*/
- public com.google.protobuf.ByteString getFamilyName() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamilyName() {
return familyName_;
}
@@ -2304,7 +2304,7 @@ public final class SnapshotProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, familyName_);
@@ -2321,11 +2321,11 @@ public final class SnapshotProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(1, familyName_);
}
for (int i = 0; i < storeFiles_.size(); i++) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(2, storeFiles_.get(i));
}
size += unknownFields.getSerializedSize();
@@ -2377,61 +2377,61 @@ public final class SnapshotProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -2449,7 +2449,7 @@ public final class SnapshotProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -2457,15 +2457,15 @@ public final class SnapshotProtos {
* Protobuf type {@code hbase.pb.SnapshotRegionManifest.FamilyFiles}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.SnapshotRegionManifest.FamilyFiles)
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFilesOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotRegionManifest_FamilyFiles_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotRegionManifest_FamilyFiles_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2478,19 +2478,19 @@ public final class SnapshotProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getStoreFilesFieldBuilder();
}
}
public Builder clear() {
super.clear();
- familyName_ = com.google.protobuf.ByteString.EMPTY;
+ familyName_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000001);
if (storeFilesBuilder_ == null) {
storeFiles_ = java.util.Collections.emptyList();
@@ -2501,7 +2501,7 @@ public final class SnapshotProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotRegionManifest_FamilyFiles_descriptor;
}
@@ -2544,29 +2544,29 @@ public final class SnapshotProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles)other);
} else {
@@ -2599,7 +2599,7 @@ public final class SnapshotProtos {
storeFiles_ = other.storeFiles_;
bitField0_ = (bitField0_ & ~0x00000002);
storeFilesBuilder_ =
- com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getStoreFilesFieldBuilder() : null;
} else {
storeFilesBuilder_.addAllMessages(other.storeFiles_);
@@ -2624,13 +2624,13 @@ public final class SnapshotProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -2642,7 +2642,7 @@ public final class SnapshotProtos {
}
private int bitField0_;
- private com.google.protobuf.ByteString familyName_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString familyName_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>required bytes family_name = 1;</code>
*/
@@ -2652,13 +2652,13 @@ public final class SnapshotProtos {
/**
* <code>required bytes family_name = 1;</code>
*/
- public com.google.protobuf.ByteString getFamilyName() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamilyName() {
return familyName_;
}
/**
* <code>required bytes family_name = 1;</code>
*/
- public Builder setFamilyName(com.google.protobuf.ByteString value) {
+ public Builder setFamilyName(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -2686,7 +2686,7 @@ public final class SnapshotProtos {
}
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFileOrBuilder> storeFilesBuilder_;
/**
@@ -2818,7 +2818,7 @@ public final class SnapshotProtos {
java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile> values) {
if (storeFilesBuilder_ == null) {
ensureStoreFilesIsMutable();
- com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
values, storeFiles_);
onChanged();
} else {
@@ -2902,11 +2902,11 @@ public final class SnapshotProtos {
getStoreFilesBuilderList() {
return getStoreFilesFieldBuilder().getBuilderList();
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFileOrBuilder>
getStoreFilesFieldBuilder() {
if (storeFilesBuilder_ == null) {
- storeFilesBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3<
+ storeFilesBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFileOrBuilder>(
storeFiles_,
((bitField0_ & 0x00000002) == 0x00000002),
@@ -2917,12 +2917,12 @@ public final class SnapshotProtos {
return storeFilesBuilder_;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -2940,22 +2940,22 @@ public final class SnapshotProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<FamilyFiles>
- PARSER = new com.google.protobuf.AbstractParser<FamilyFiles>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<FamilyFiles>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<FamilyFiles>() {
public FamilyFiles parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new FamilyFiles(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<FamilyFiles> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<FamilyFiles> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<FamilyFiles> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<FamilyFiles> getParserForType() {
return PARSER;
}
@@ -3061,7 +3061,7 @@ public final class SnapshotProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
<TRUNCATED>
[02/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ZooKeeperProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ZooKeeperProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ZooKeeperProtos.java
index d7b5221..6baf845 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ZooKeeperProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ZooKeeperProtos.java
@@ -6,17 +6,17 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class ZooKeeperProtos {
private ZooKeeperProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
public interface MetaRegionServerOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.MetaRegionServer)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <pre>
@@ -93,11 +93,11 @@ public final class ZooKeeperProtos {
* Protobuf type {@code hbase.pb.MetaRegionServer}
*/
public static final class MetaRegionServer extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.MetaRegionServer)
MetaRegionServerOrBuilder {
// Use MetaRegionServer.newBuilder() to construct.
- private MetaRegionServer(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private MetaRegionServer(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private MetaRegionServer() {
@@ -106,18 +106,18 @@ public final class ZooKeeperProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private MetaRegionServer(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -164,22 +164,22 @@ public final class ZooKeeperProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.internal_static_hbase_pb_MetaRegionServer_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.internal_static_hbase_pb_MetaRegionServer_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -292,7 +292,7 @@ public final class ZooKeeperProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeMessage(1, getServer());
@@ -312,15 +312,15 @@ public final class ZooKeeperProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(1, getServer());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt32Size(2, rpcVersion_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeEnumSize(3, state_);
}
size += unknownFields.getSerializedSize();
@@ -383,61 +383,61 @@ public final class ZooKeeperProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -455,7 +455,7 @@ public final class ZooKeeperProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -468,15 +468,15 @@ public final class ZooKeeperProtos {
* Protobuf type {@code hbase.pb.MetaRegionServer}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.MetaRegionServer)
org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.MetaRegionServerOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.internal_static_hbase_pb_MetaRegionServer_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.internal_static_hbase_pb_MetaRegionServer_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -489,12 +489,12 @@ public final class ZooKeeperProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getServerFieldBuilder();
}
@@ -514,7 +514,7 @@ public final class ZooKeeperProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.internal_static_hbase_pb_MetaRegionServer_descriptor;
}
@@ -560,29 +560,29 @@ public final class ZooKeeperProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.MetaRegionServer) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.MetaRegionServer)other);
} else {
@@ -618,13 +618,13 @@ public final class ZooKeeperProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.MetaRegionServer parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.MetaRegionServer) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -637,7 +637,7 @@ public final class ZooKeeperProtos {
private int bitField0_;
private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName server_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverBuilder_;
/**
* <pre>
@@ -785,11 +785,11 @@ public final class ZooKeeperProtos {
*
* <code>required .hbase.pb.ServerName server = 1;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder>
getServerFieldBuilder() {
if (serverBuilder_ == null) {
- serverBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ serverBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder>(
getServer(),
getParentForChildren(),
@@ -907,12 +907,12 @@ public final class ZooKeeperProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -930,22 +930,22 @@ public final class ZooKeeperProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<MetaRegionServer>
- PARSER = new com.google.protobuf.AbstractParser<MetaRegionServer>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<MetaRegionServer>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<MetaRegionServer>() {
public MetaRegionServer parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new MetaRegionServer(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<MetaRegionServer> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<MetaRegionServer> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<MetaRegionServer> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<MetaRegionServer> getParserForType() {
return PARSER;
}
@@ -957,7 +957,7 @@ public final class ZooKeeperProtos {
public interface MasterOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.Master)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <pre>
@@ -1019,11 +1019,11 @@ public final class ZooKeeperProtos {
* Protobuf type {@code hbase.pb.Master}
*/
public static final class Master extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.Master)
MasterOrBuilder {
// Use Master.newBuilder() to construct.
- private Master(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private Master(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private Master() {
@@ -1032,18 +1032,18 @@ public final class ZooKeeperProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private Master(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -1084,22 +1084,22 @@ public final class ZooKeeperProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.internal_static_hbase_pb_Master_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.internal_static_hbase_pb_Master_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1196,7 +1196,7 @@ public final class ZooKeeperProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeMessage(1, getMaster());
@@ -1216,15 +1216,15 @@ public final class ZooKeeperProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(1, getMaster());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt32Size(2, rpcVersion_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt32Size(3, infoPort_);
}
size += unknownFields.getSerializedSize();
@@ -1288,61 +1288,61 @@ public final class ZooKeeperProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.Master parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.Master parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.Master parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.Master parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.Master parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.Master parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.Master parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.Master parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.Master parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.Master parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -1360,7 +1360,7 @@ public final class ZooKeeperProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1373,15 +1373,15 @@ public final class ZooKeeperProtos {
* Protobuf type {@code hbase.pb.Master}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.Master)
org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.MasterOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.internal_static_hbase_pb_Master_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.internal_static_hbase_pb_Master_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1394,12 +1394,12 @@ public final class ZooKeeperProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getMasterFieldBuilder();
}
@@ -1419,7 +1419,7 @@ public final class ZooKeeperProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.internal_static_hbase_pb_Master_descriptor;
}
@@ -1465,29 +1465,29 @@ public final class ZooKeeperProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.Master) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.Master)other);
} else {
@@ -1523,13 +1523,13 @@ public final class ZooKeeperProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.Master parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.Master) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -1542,7 +1542,7 @@ public final class ZooKeeperProtos {
private int bitField0_;
private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName master_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> masterBuilder_;
/**
* <pre>
@@ -1681,11 +1681,11 @@ public final class ZooKeeperProtos {
*
* <code>required .hbase.pb.ServerName master = 1;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder>
getMasterFieldBuilder() {
if (masterBuilder_ == null) {
- masterBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ masterBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder>(
getMaster(),
getParentForChildren(),
@@ -1775,12 +1775,12 @@ public final class ZooKeeperProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -1798,22 +1798,22 @@ public final class ZooKeeperProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<Master>
- PARSER = new com.google.protobuf.AbstractParser<Master>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Master>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Master>() {
public Master parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new Master(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<Master> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Master> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<Master> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Master> getParserForType() {
return PARSER;
}
@@ -1825,7 +1825,7 @@ public final class ZooKeeperProtos {
public interface ClusterUpOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.ClusterUp)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <pre>
@@ -1853,7 +1853,7 @@ public final class ZooKeeperProtos {
*
* <code>required string start_date = 1;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getStartDateBytes();
}
/**
@@ -1865,11 +1865,11 @@ public final class ZooKeeperProtos {
* Protobuf type {@code hbase.pb.ClusterUp}
*/
public static final class ClusterUp extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.ClusterUp)
ClusterUpOrBuilder {
// Use ClusterUp.newBuilder() to construct.
- private ClusterUp(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private ClusterUp(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private ClusterUp() {
@@ -1877,18 +1877,18 @@ public final class ZooKeeperProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private ClusterUp(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -1905,29 +1905,29 @@ public final class ZooKeeperProtos {
break;
}
case 10: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000001;
startDate_ = bs;
break;
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.internal_static_hbase_pb_ClusterUp_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.internal_static_hbase_pb_ClusterUp_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1961,8 +1961,8 @@ public final class ZooKeeperProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
startDate_ = s;
@@ -1978,17 +1978,17 @@ public final class ZooKeeperProtos {
*
* <code>required string start_date = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getStartDateBytes() {
java.lang.Object ref = startDate_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
startDate_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -2006,10 +2006,10 @@ public final class ZooKeeperProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 1, startDate_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, startDate_);
}
unknownFields.writeTo(output);
}
@@ -2020,7 +2020,7 @@ public final class ZooKeeperProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, startDate_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, startDate_);
}
size += unknownFields.getSerializedSize();
memoizedSize = size;
@@ -2065,61 +2065,61 @@ public final class ZooKeeperProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.ClusterUp parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.ClusterUp parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.ClusterUp parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.ClusterUp parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.ClusterUp parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.ClusterUp parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.ClusterUp parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.ClusterUp parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.ClusterUp parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.ClusterUp parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -2137,7 +2137,7 @@ public final class ZooKeeperProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -2150,15 +2150,15 @@ public final class ZooKeeperProtos {
* Protobuf type {@code hbase.pb.ClusterUp}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.ClusterUp)
org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.ClusterUpOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.internal_static_hbase_pb_ClusterUp_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.internal_static_hbase_pb_ClusterUp_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2171,12 +2171,12 @@ public final class ZooKeeperProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -2187,7 +2187,7 @@ public final class ZooKeeperProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.internal_static_hbase_pb_ClusterUp_descriptor;
}
@@ -2221,29 +2221,29 @@ public final class ZooKeeperProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.ClusterUp) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.ClusterUp)other);
} else {
@@ -2272,13 +2272,13 @@ public final class ZooKeeperProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.ClusterUp parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.ClusterUp) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -2313,8 +2313,8 @@ public final class ZooKeeperProtos {
public java.lang.String getStartDate() {
java.lang.Object ref = startDate_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
startDate_ = s;
@@ -2332,17 +2332,17 @@ public final class ZooKeeperProtos {
*
* <code>required string start_date = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getStartDateBytes() {
java.lang.Object ref = startDate_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
startDate_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -2386,7 +2386,7 @@ public final class ZooKeeperProtos {
* <code>required string start_date = 1;</code>
*/
public Builder setStartDateBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -2396,12 +2396,12 @@ public final class ZooKeeperProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -2419,22 +2419,22 @@ public final class ZooKeeperProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<ClusterUp>
- PARSER = new com.google.protobuf.AbstractParser<ClusterUp>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ClusterUp>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<ClusterUp>() {
public ClusterUp parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new ClusterUp(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<ClusterUp> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ClusterUp> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<ClusterUp> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ClusterUp> getParserForType() {
return PARSER;
}
@@ -2446,7 +2446,7 @@ public final class ZooKeeperProtos {
public interface SplitLogTaskOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.SplitLogTask)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required .hbase.pb.SplitLogTask.State state = 1;</code>
@@ -2489,11 +2489,11 @@ public final class ZooKeeperProtos {
* Protobuf type {@code hbase.pb.SplitLogTask}
*/
public static final class SplitLogTask extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.SplitLogTask)
SplitLogTaskOrBuilder {
// Use SplitLogTask.newBuilder() to construct.
- private SplitLogTask(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private SplitLogTask(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private SplitLogTask() {
@@ -2502,18 +2502,18 @@ public final class ZooKeeperProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private SplitLogTask(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -2566,22 +2566,22 @@ public final class ZooKeeperProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.internal_static_hbase_pb_SplitLogTask_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.internal_static_hbase_pb_SplitLogTask_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2592,7 +2592,7 @@ public final class ZooKeeperProtos {
* Protobuf enum {@code hbase.pb.SplitLogTask.State}
*/
public enum State
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>UNASSIGNED = 0;</code>
*/
@@ -2660,27 +2660,27 @@ public final class ZooKeeperProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<State>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<State>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
State> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<State>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<State>() {
public State findValueByNumber(int number) {
return State.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.SplitLogTask.getDescriptor().getEnumTypes().get(0);
}
@@ -2688,7 +2688,7 @@ public final class ZooKeeperProtos {
private static final State[] VALUES = values();
public static State valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -2709,7 +2709,7 @@ public final class ZooKeeperProtos {
* Protobuf enum {@code hbase.pb.SplitLogTask.RecoveryMode}
*/
public enum RecoveryMode
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>UNKNOWN = 0;</code>
*/
@@ -2759,27 +2759,27 @@ public final class ZooKeeperProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<RecoveryMode>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<RecoveryMode>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
RecoveryMode> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<RecoveryMode>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<RecoveryMode>() {
public RecoveryMode findValueByNumber(int number) {
return RecoveryMode.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.SplitLogTask.getDescriptor().getEnumTypes().get(1);
}
@@ -2787,7 +2787,7 @@ public final class ZooKeeperProtos {
private static final RecoveryMode[] VALUES = values();
public static RecoveryMode valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -2880,7 +2880,7 @@ public final class ZooKeeperProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeEnum(1, state_);
@@ -2900,15 +2900,15 @@ public final class ZooKeeperProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeEnumSize(1, state_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(2, getServerName());
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeEnumSize(3, mode_);
}
size += unknownFields.getSerializedSize();
@@ -2970,61 +2970,61 @@ public final class ZooKeeperProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.SplitLogTask parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.SplitLogTask parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.SplitLogTask parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.SplitLogTask parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.SplitLogTask parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.SplitLogTask parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.SplitLogTask parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.SplitLogTask parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.SplitLogTask parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.SplitLogTask parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -3042,7 +3042,7 @@ public final class ZooKeeperProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -3056,15 +3056,15 @@ public final class ZooKeeperProtos {
* Protobuf type {@code hbase.pb.SplitLogTask}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.SplitLogTask)
org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.SplitLogTaskOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.internal_static_hbase_pb_SplitLogTask_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.internal_static_hbase_pb_SplitLogTask_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -3077,12 +3077,12 @@ public final class ZooKeeperProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getServerNameFieldBuilder();
}
@@ -3102,7 +3102,7 @@ public final class ZooKeeperProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.internal_static_hbase_pb_SplitLogTask_descriptor;
}
@@ -3148,29 +3148,29 @@ public final class ZooKeeperProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.SplitLogTask) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.SplitLogTask)other);
} else {
@@ -3209,13 +3209,13 @@ public final class ZooKeeperProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.SplitLogTask parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.SplitLogTask) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -3264,7 +3264,7 @@ public final class ZooKeeperProtos {
}
private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName serverName_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_;
/**
* <code>required .hbase.pb.ServerName server_name = 2;</code>
@@ -3367,11 +3367,11 @@ public final class ZooKeeperProtos {
/**
* <code>required .hbase.pb.ServerName server_name = 2;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder>
getServerNameFieldBuilder() {
if (serverNameBuilder_ == null) {
- serverNameBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ serverNameBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder>(
getServerName(),
getParentForChildren(),
@@ -3417,12 +3417,12 @@ public final class ZooKeeperProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -3440,22 +3440,22 @@ public final class ZooKeeperProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<SplitLogTask>
- PARSER = new com.google.protobuf.AbstractParser<SplitLogTask>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<SplitLogTask>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<SplitLogTask>() {
public SplitLogTask parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new SplitLogTask(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<SplitLogTask> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<SplitLogTask> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<SplitLogTask> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<SplitLogTask> getParserForType() {
return PARSER;
}
@@ -3467,7 +3467,7 @@ public final class ZooKeeperProtos {
public interface DeprecatedTableStateOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.DeprecatedTableState)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <pre>
@@ -3500,11 +3500,11 @@ public final class ZooKeeperProtos {
* Protobuf type {@code hbase.pb.DeprecatedTableState}
*/
public static final class DeprecatedTableState extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.DeprecatedTableState)
DeprecatedTableStateOrBuilder {
// Use DeprecatedTableState.newBuilder() to construct.
- private DeprecatedTableState(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private DeprecatedTableState(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private DeprecatedTableState() {
@@ -3512,18 +3512,18 @@ public final class ZooKeeperProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private DeprecatedTableState(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.p
<TRUNCATED>
[19/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/EncryptionProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/EncryptionProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/EncryptionProtos.java
index 689d33f..b195b5a 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/EncryptionProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/EncryptionProtos.java
@@ -6,17 +6,17 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class EncryptionProtos {
private EncryptionProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
public interface WrappedKeyOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.WrappedKey)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required string algorithm = 1;</code>
@@ -29,7 +29,7 @@ public final class EncryptionProtos {
/**
* <code>required string algorithm = 1;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getAlgorithmBytes();
/**
@@ -48,7 +48,7 @@ public final class EncryptionProtos {
/**
* <code>required bytes data = 3;</code>
*/
- com.google.protobuf.ByteString getData();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getData();
/**
* <code>optional bytes iv = 4;</code>
@@ -57,7 +57,7 @@ public final class EncryptionProtos {
/**
* <code>optional bytes iv = 4;</code>
*/
- com.google.protobuf.ByteString getIv();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getIv();
/**
* <code>optional bytes hash = 5;</code>
@@ -66,40 +66,40 @@ public final class EncryptionProtos {
/**
* <code>optional bytes hash = 5;</code>
*/
- com.google.protobuf.ByteString getHash();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getHash();
}
/**
* Protobuf type {@code hbase.pb.WrappedKey}
*/
public static final class WrappedKey extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.WrappedKey)
WrappedKeyOrBuilder {
// Use WrappedKey.newBuilder() to construct.
- private WrappedKey(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private WrappedKey(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private WrappedKey() {
algorithm_ = "";
length_ = 0;
- data_ = com.google.protobuf.ByteString.EMPTY;
- iv_ = com.google.protobuf.ByteString.EMPTY;
- hash_ = com.google.protobuf.ByteString.EMPTY;
+ data_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
+ iv_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
+ hash_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private WrappedKey(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -116,7 +116,7 @@ public final class EncryptionProtos {
break;
}
case 10: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000001;
algorithm_ = bs;
break;
@@ -143,22 +143,22 @@ public final class EncryptionProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.EncryptionProtos.internal_static_hbase_pb_WrappedKey_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.EncryptionProtos.internal_static_hbase_pb_WrappedKey_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -182,8 +182,8 @@ public final class EncryptionProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
algorithm_ = s;
@@ -194,17 +194,17 @@ public final class EncryptionProtos {
/**
* <code>required string algorithm = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getAlgorithmBytes() {
java.lang.Object ref = algorithm_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
algorithm_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -224,7 +224,7 @@ public final class EncryptionProtos {
}
public static final int DATA_FIELD_NUMBER = 3;
- private com.google.protobuf.ByteString data_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data_;
/**
* <code>required bytes data = 3;</code>
*/
@@ -234,12 +234,12 @@ public final class EncryptionProtos {
/**
* <code>required bytes data = 3;</code>
*/
- public com.google.protobuf.ByteString getData() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getData() {
return data_;
}
public static final int IV_FIELD_NUMBER = 4;
- private com.google.protobuf.ByteString iv_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString iv_;
/**
* <code>optional bytes iv = 4;</code>
*/
@@ -249,12 +249,12 @@ public final class EncryptionProtos {
/**
* <code>optional bytes iv = 4;</code>
*/
- public com.google.protobuf.ByteString getIv() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getIv() {
return iv_;
}
public static final int HASH_FIELD_NUMBER = 5;
- private com.google.protobuf.ByteString hash_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString hash_;
/**
* <code>optional bytes hash = 5;</code>
*/
@@ -264,7 +264,7 @@ public final class EncryptionProtos {
/**
* <code>optional bytes hash = 5;</code>
*/
- public com.google.protobuf.ByteString getHash() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getHash() {
return hash_;
}
@@ -290,10 +290,10 @@ public final class EncryptionProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 1, algorithm_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, algorithm_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
output.writeUInt32(2, length_);
@@ -316,22 +316,22 @@ public final class EncryptionProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, algorithm_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, algorithm_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt32Size(2, length_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(3, data_);
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(4, iv_);
}
if (((bitField0_ & 0x00000010) == 0x00000010)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(5, hash_);
}
size += unknownFields.getSerializedSize();
@@ -413,61 +413,61 @@ public final class EncryptionProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.EncryptionProtos.WrappedKey parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.EncryptionProtos.WrappedKey parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.EncryptionProtos.WrappedKey parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.EncryptionProtos.WrappedKey parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.EncryptionProtos.WrappedKey parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.EncryptionProtos.WrappedKey parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.EncryptionProtos.WrappedKey parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.EncryptionProtos.WrappedKey parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.EncryptionProtos.WrappedKey parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.EncryptionProtos.WrappedKey parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -485,7 +485,7 @@ public final class EncryptionProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -493,15 +493,15 @@ public final class EncryptionProtos {
* Protobuf type {@code hbase.pb.WrappedKey}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.WrappedKey)
org.apache.hadoop.hbase.shaded.protobuf.generated.EncryptionProtos.WrappedKeyOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.EncryptionProtos.internal_static_hbase_pb_WrappedKey_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.EncryptionProtos.internal_static_hbase_pb_WrappedKey_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -514,12 +514,12 @@ public final class EncryptionProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -529,16 +529,16 @@ public final class EncryptionProtos {
bitField0_ = (bitField0_ & ~0x00000001);
length_ = 0;
bitField0_ = (bitField0_ & ~0x00000002);
- data_ = com.google.protobuf.ByteString.EMPTY;
+ data_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000004);
- iv_ = com.google.protobuf.ByteString.EMPTY;
+ iv_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000008);
- hash_ = com.google.protobuf.ByteString.EMPTY;
+ hash_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000010);
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.EncryptionProtos.internal_static_hbase_pb_WrappedKey_descriptor;
}
@@ -588,29 +588,29 @@ public final class EncryptionProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.EncryptionProtos.WrappedKey) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.EncryptionProtos.WrappedKey)other);
} else {
@@ -657,13 +657,13 @@ public final class EncryptionProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.EncryptionProtos.WrappedKey parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.EncryptionProtos.WrappedKey) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -688,8 +688,8 @@ public final class EncryptionProtos {
public java.lang.String getAlgorithm() {
java.lang.Object ref = algorithm_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
algorithm_ = s;
@@ -702,17 +702,17 @@ public final class EncryptionProtos {
/**
* <code>required string algorithm = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getAlgorithmBytes() {
java.lang.Object ref = algorithm_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
algorithm_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -741,7 +741,7 @@ public final class EncryptionProtos {
* <code>required string algorithm = 1;</code>
*/
public Builder setAlgorithmBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -783,7 +783,7 @@ public final class EncryptionProtos {
return this;
}
- private com.google.protobuf.ByteString data_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>required bytes data = 3;</code>
*/
@@ -793,13 +793,13 @@ public final class EncryptionProtos {
/**
* <code>required bytes data = 3;</code>
*/
- public com.google.protobuf.ByteString getData() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getData() {
return data_;
}
/**
* <code>required bytes data = 3;</code>
*/
- public Builder setData(com.google.protobuf.ByteString value) {
+ public Builder setData(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -818,7 +818,7 @@ public final class EncryptionProtos {
return this;
}
- private com.google.protobuf.ByteString iv_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString iv_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>optional bytes iv = 4;</code>
*/
@@ -828,13 +828,13 @@ public final class EncryptionProtos {
/**
* <code>optional bytes iv = 4;</code>
*/
- public com.google.protobuf.ByteString getIv() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getIv() {
return iv_;
}
/**
* <code>optional bytes iv = 4;</code>
*/
- public Builder setIv(com.google.protobuf.ByteString value) {
+ public Builder setIv(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -853,7 +853,7 @@ public final class EncryptionProtos {
return this;
}
- private com.google.protobuf.ByteString hash_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString hash_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>optional bytes hash = 5;</code>
*/
@@ -863,13 +863,13 @@ public final class EncryptionProtos {
/**
* <code>optional bytes hash = 5;</code>
*/
- public com.google.protobuf.ByteString getHash() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getHash() {
return hash_;
}
/**
* <code>optional bytes hash = 5;</code>
*/
- public Builder setHash(com.google.protobuf.ByteString value) {
+ public Builder setHash(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -888,12 +888,12 @@ public final class EncryptionProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -911,22 +911,22 @@ public final class EncryptionProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<WrappedKey>
- PARSER = new com.google.protobuf.AbstractParser<WrappedKey>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<WrappedKey>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<WrappedKey>() {
public WrappedKey parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new WrappedKey(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<WrappedKey> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<WrappedKey> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<WrappedKey> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<WrappedKey> getParserForType() {
return PARSER;
}
@@ -936,17 +936,17 @@ public final class EncryptionProtos {
}
- private static final com.google.protobuf.Descriptors.Descriptor
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
internal_static_hbase_pb_WrappedKey_descriptor;
private static final
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hbase_pb_WrappedKey_fieldAccessorTable;
- public static com.google.protobuf.Descriptors.FileDescriptor
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
getDescriptor() {
return descriptor;
}
- private static com.google.protobuf.Descriptors.FileDescriptor
+ private static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
descriptor;
static {
java.lang.String[] descriptorData = {
@@ -956,22 +956,22 @@ public final class EncryptionProtos {
"1org.apache.hadoop.hbase.shaded.protobuf" +
".generatedB\020EncryptionProtosH\001\240\001\001"
};
- com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
- new com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
- public com.google.protobuf.ExtensionRegistry assignDescriptors(
- com.google.protobuf.Descriptors.FileDescriptor root) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
descriptor = root;
return null;
}
};
- com.google.protobuf.Descriptors.FileDescriptor
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
.internalBuildGeneratedFileFrom(descriptorData,
- new com.google.protobuf.Descriptors.FileDescriptor[] {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
}, assigner);
internal_static_hbase_pb_WrappedKey_descriptor =
getDescriptor().getMessageTypes().get(0);
internal_static_hbase_pb_WrappedKey_fieldAccessorTable = new
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hbase_pb_WrappedKey_descriptor,
new java.lang.String[] { "Algorithm", "Length", "Data", "Iv", "Hash", });
}
[07/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/RegionNormalizerProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/RegionNormalizerProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/RegionNormalizerProtos.java
index 66b6990..0a58113 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/RegionNormalizerProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/RegionNormalizerProtos.java
@@ -6,17 +6,17 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class RegionNormalizerProtos {
private RegionNormalizerProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
public interface RegionNormalizerStateOrBuilder extends
// @@protoc_insertion_point(interface_extends:RegionNormalizerState)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>optional bool normalizer_on = 1;</code>
@@ -31,11 +31,11 @@ public final class RegionNormalizerProtos {
* Protobuf type {@code RegionNormalizerState}
*/
public static final class RegionNormalizerState extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:RegionNormalizerState)
RegionNormalizerStateOrBuilder {
// Use RegionNormalizerState.newBuilder() to construct.
- private RegionNormalizerState(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private RegionNormalizerState(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private RegionNormalizerState() {
@@ -43,18 +43,18 @@ public final class RegionNormalizerProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private RegionNormalizerState(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -77,22 +77,22 @@ public final class RegionNormalizerProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionNormalizerProtos.internal_static_RegionNormalizerState_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionNormalizerProtos.internal_static_RegionNormalizerState_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -125,7 +125,7 @@ public final class RegionNormalizerProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBool(1, normalizerOn_);
@@ -139,7 +139,7 @@ public final class RegionNormalizerProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBoolSize(1, normalizerOn_);
}
size += unknownFields.getSerializedSize();
@@ -177,7 +177,7 @@ public final class RegionNormalizerProtos {
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasNormalizerOn()) {
hash = (37 * hash) + NORMALIZER_ON_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean(
getNormalizerOn());
}
hash = (29 * hash) + unknownFields.hashCode();
@@ -186,61 +186,61 @@ public final class RegionNormalizerProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionNormalizerProtos.RegionNormalizerState parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionNormalizerProtos.RegionNormalizerState parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionNormalizerProtos.RegionNormalizerState parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionNormalizerProtos.RegionNormalizerState parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionNormalizerProtos.RegionNormalizerState parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionNormalizerProtos.RegionNormalizerState parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionNormalizerProtos.RegionNormalizerState parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionNormalizerProtos.RegionNormalizerState parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionNormalizerProtos.RegionNormalizerState parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RegionNormalizerProtos.RegionNormalizerState parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -258,7 +258,7 @@ public final class RegionNormalizerProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -266,15 +266,15 @@ public final class RegionNormalizerProtos {
* Protobuf type {@code RegionNormalizerState}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:RegionNormalizerState)
org.apache.hadoop.hbase.shaded.protobuf.generated.RegionNormalizerProtos.RegionNormalizerStateOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionNormalizerProtos.internal_static_RegionNormalizerState_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionNormalizerProtos.internal_static_RegionNormalizerState_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -287,12 +287,12 @@ public final class RegionNormalizerProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -303,7 +303,7 @@ public final class RegionNormalizerProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RegionNormalizerProtos.internal_static_RegionNormalizerState_descriptor;
}
@@ -337,29 +337,29 @@ public final class RegionNormalizerProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.RegionNormalizerProtos.RegionNormalizerState) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.RegionNormalizerProtos.RegionNormalizerState)other);
} else {
@@ -383,13 +383,13 @@ public final class RegionNormalizerProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.RegionNormalizerProtos.RegionNormalizerState parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.RegionNormalizerProtos.RegionNormalizerState) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -433,12 +433,12 @@ public final class RegionNormalizerProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -456,22 +456,22 @@ public final class RegionNormalizerProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<RegionNormalizerState>
- PARSER = new com.google.protobuf.AbstractParser<RegionNormalizerState>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RegionNormalizerState>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<RegionNormalizerState>() {
public RegionNormalizerState parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new RegionNormalizerState(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<RegionNormalizerState> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RegionNormalizerState> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<RegionNormalizerState> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RegionNormalizerState> getParserForType() {
return PARSER;
}
@@ -481,17 +481,17 @@ public final class RegionNormalizerProtos {
}
- private static final com.google.protobuf.Descriptors.Descriptor
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
internal_static_RegionNormalizerState_descriptor;
private static final
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_RegionNormalizerState_fieldAccessorTable;
- public static com.google.protobuf.Descriptors.FileDescriptor
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
getDescriptor() {
return descriptor;
}
- private static com.google.protobuf.Descriptors.FileDescriptor
+ private static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
descriptor;
static {
java.lang.String[] descriptorData = {
@@ -500,22 +500,22 @@ public final class RegionNormalizerProtos {
"apache.hadoop.hbase.shaded.protobuf.gene" +
"ratedB\026RegionNormalizerProtosH\001\240\001\001"
};
- com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
- new com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
- public com.google.protobuf.ExtensionRegistry assignDescriptors(
- com.google.protobuf.Descriptors.FileDescriptor root) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
descriptor = root;
return null;
}
};
- com.google.protobuf.Descriptors.FileDescriptor
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
.internalBuildGeneratedFileFrom(descriptorData,
- new com.google.protobuf.Descriptors.FileDescriptor[] {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
}, assigner);
internal_static_RegionNormalizerState_descriptor =
getDescriptor().getMessageTypes().get(0);
internal_static_RegionNormalizerState_fieldAccessorTable = new
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_RegionNormalizerState_descriptor,
new java.lang.String[] { "NormalizerOn", });
}
[24/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/CellProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/CellProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/CellProtos.java
index f049d50..60ca467 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/CellProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/CellProtos.java
@@ -6,13 +6,13 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class CellProtos {
private CellProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
/**
* <pre>
@@ -23,7 +23,7 @@ public final class CellProtos {
* Protobuf enum {@code hbase.pb.CellType}
*/
public enum CellType
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>MINIMUM = 0;</code>
*/
@@ -108,27 +108,27 @@ public final class CellProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<CellType>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<CellType>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
CellType> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<CellType>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<CellType>() {
public CellType findValueByNumber(int number) {
return CellType.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.getDescriptor().getEnumTypes().get(0);
}
@@ -136,7 +136,7 @@ public final class CellProtos {
private static final CellType[] VALUES = values();
public static CellType valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -155,7 +155,7 @@ public final class CellProtos {
public interface CellOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.Cell)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>optional bytes row = 1;</code>
@@ -164,7 +164,7 @@ public final class CellProtos {
/**
* <code>optional bytes row = 1;</code>
*/
- com.google.protobuf.ByteString getRow();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getRow();
/**
* <code>optional bytes family = 2;</code>
@@ -173,7 +173,7 @@ public final class CellProtos {
/**
* <code>optional bytes family = 2;</code>
*/
- com.google.protobuf.ByteString getFamily();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamily();
/**
* <code>optional bytes qualifier = 3;</code>
@@ -182,7 +182,7 @@ public final class CellProtos {
/**
* <code>optional bytes qualifier = 3;</code>
*/
- com.google.protobuf.ByteString getQualifier();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getQualifier();
/**
* <code>optional uint64 timestamp = 4;</code>
@@ -209,7 +209,7 @@ public final class CellProtos {
/**
* <code>optional bytes value = 6;</code>
*/
- com.google.protobuf.ByteString getValue();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getValue();
/**
* <code>optional bytes tags = 7;</code>
@@ -218,7 +218,7 @@ public final class CellProtos {
/**
* <code>optional bytes tags = 7;</code>
*/
- com.google.protobuf.ByteString getTags();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getTags();
}
/**
* <pre>
@@ -229,36 +229,36 @@ public final class CellProtos {
* Protobuf type {@code hbase.pb.Cell}
*/
public static final class Cell extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.Cell)
CellOrBuilder {
// Use Cell.newBuilder() to construct.
- private Cell(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private Cell(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private Cell() {
- row_ = com.google.protobuf.ByteString.EMPTY;
- family_ = com.google.protobuf.ByteString.EMPTY;
- qualifier_ = com.google.protobuf.ByteString.EMPTY;
+ row_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
+ family_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
+ qualifier_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
timestamp_ = 0L;
cellType_ = 0;
- value_ = com.google.protobuf.ByteString.EMPTY;
- tags_ = com.google.protobuf.ByteString.EMPTY;
+ value_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
+ tags_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private Cell(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -317,22 +317,22 @@ public final class CellProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.internal_static_hbase_pb_Cell_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.internal_static_hbase_pb_Cell_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -341,7 +341,7 @@ public final class CellProtos {
private int bitField0_;
public static final int ROW_FIELD_NUMBER = 1;
- private com.google.protobuf.ByteString row_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString row_;
/**
* <code>optional bytes row = 1;</code>
*/
@@ -351,12 +351,12 @@ public final class CellProtos {
/**
* <code>optional bytes row = 1;</code>
*/
- public com.google.protobuf.ByteString getRow() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getRow() {
return row_;
}
public static final int FAMILY_FIELD_NUMBER = 2;
- private com.google.protobuf.ByteString family_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString family_;
/**
* <code>optional bytes family = 2;</code>
*/
@@ -366,12 +366,12 @@ public final class CellProtos {
/**
* <code>optional bytes family = 2;</code>
*/
- public com.google.protobuf.ByteString getFamily() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamily() {
return family_;
}
public static final int QUALIFIER_FIELD_NUMBER = 3;
- private com.google.protobuf.ByteString qualifier_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString qualifier_;
/**
* <code>optional bytes qualifier = 3;</code>
*/
@@ -381,7 +381,7 @@ public final class CellProtos {
/**
* <code>optional bytes qualifier = 3;</code>
*/
- public com.google.protobuf.ByteString getQualifier() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getQualifier() {
return qualifier_;
}
@@ -417,7 +417,7 @@ public final class CellProtos {
}
public static final int VALUE_FIELD_NUMBER = 6;
- private com.google.protobuf.ByteString value_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value_;
/**
* <code>optional bytes value = 6;</code>
*/
@@ -427,12 +427,12 @@ public final class CellProtos {
/**
* <code>optional bytes value = 6;</code>
*/
- public com.google.protobuf.ByteString getValue() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getValue() {
return value_;
}
public static final int TAGS_FIELD_NUMBER = 7;
- private com.google.protobuf.ByteString tags_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString tags_;
/**
* <code>optional bytes tags = 7;</code>
*/
@@ -442,7 +442,7 @@ public final class CellProtos {
/**
* <code>optional bytes tags = 7;</code>
*/
- public com.google.protobuf.ByteString getTags() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getTags() {
return tags_;
}
@@ -456,7 +456,7 @@ public final class CellProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, row_);
@@ -488,31 +488,31 @@ public final class CellProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(1, row_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(2, family_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(3, qualifier_);
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(4, timestamp_);
}
if (((bitField0_ & 0x00000010) == 0x00000010)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeEnumSize(5, cellType_);
}
if (((bitField0_ & 0x00000020) == 0x00000020)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(6, value_);
}
if (((bitField0_ & 0x00000040) == 0x00000040)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(7, tags_);
}
size += unknownFields.getSerializedSize();
@@ -591,7 +591,7 @@ public final class CellProtos {
}
if (hasTimestamp()) {
hash = (37 * hash) + TIMESTAMP_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getTimestamp());
}
if (hasCellType()) {
@@ -612,61 +612,61 @@ public final class CellProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.Cell parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.Cell parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.Cell parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.Cell parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.Cell parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.Cell parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.Cell parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.Cell parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.Cell parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.Cell parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -684,7 +684,7 @@ public final class CellProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -697,15 +697,15 @@ public final class CellProtos {
* Protobuf type {@code hbase.pb.Cell}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.Cell)
org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.CellOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.internal_static_hbase_pb_Cell_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.internal_static_hbase_pb_Cell_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -718,35 +718,35 @@ public final class CellProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
public Builder clear() {
super.clear();
- row_ = com.google.protobuf.ByteString.EMPTY;
+ row_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000001);
- family_ = com.google.protobuf.ByteString.EMPTY;
+ family_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000002);
- qualifier_ = com.google.protobuf.ByteString.EMPTY;
+ qualifier_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000004);
timestamp_ = 0L;
bitField0_ = (bitField0_ & ~0x00000008);
cellType_ = 0;
bitField0_ = (bitField0_ & ~0x00000010);
- value_ = com.google.protobuf.ByteString.EMPTY;
+ value_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000020);
- tags_ = com.google.protobuf.ByteString.EMPTY;
+ tags_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000040);
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.internal_static_hbase_pb_Cell_descriptor;
}
@@ -804,29 +804,29 @@ public final class CellProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.Cell) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.Cell)other);
} else {
@@ -868,13 +868,13 @@ public final class CellProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.Cell parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.Cell) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -886,7 +886,7 @@ public final class CellProtos {
}
private int bitField0_;
- private com.google.protobuf.ByteString row_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString row_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>optional bytes row = 1;</code>
*/
@@ -896,13 +896,13 @@ public final class CellProtos {
/**
* <code>optional bytes row = 1;</code>
*/
- public com.google.protobuf.ByteString getRow() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getRow() {
return row_;
}
/**
* <code>optional bytes row = 1;</code>
*/
- public Builder setRow(com.google.protobuf.ByteString value) {
+ public Builder setRow(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -921,7 +921,7 @@ public final class CellProtos {
return this;
}
- private com.google.protobuf.ByteString family_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString family_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>optional bytes family = 2;</code>
*/
@@ -931,13 +931,13 @@ public final class CellProtos {
/**
* <code>optional bytes family = 2;</code>
*/
- public com.google.protobuf.ByteString getFamily() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamily() {
return family_;
}
/**
* <code>optional bytes family = 2;</code>
*/
- public Builder setFamily(com.google.protobuf.ByteString value) {
+ public Builder setFamily(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -956,7 +956,7 @@ public final class CellProtos {
return this;
}
- private com.google.protobuf.ByteString qualifier_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString qualifier_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>optional bytes qualifier = 3;</code>
*/
@@ -966,13 +966,13 @@ public final class CellProtos {
/**
* <code>optional bytes qualifier = 3;</code>
*/
- public com.google.protobuf.ByteString getQualifier() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getQualifier() {
return qualifier_;
}
/**
* <code>optional bytes qualifier = 3;</code>
*/
- public Builder setQualifier(com.google.protobuf.ByteString value) {
+ public Builder setQualifier(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1059,7 +1059,7 @@ public final class CellProtos {
return this;
}
- private com.google.protobuf.ByteString value_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>optional bytes value = 6;</code>
*/
@@ -1069,13 +1069,13 @@ public final class CellProtos {
/**
* <code>optional bytes value = 6;</code>
*/
- public com.google.protobuf.ByteString getValue() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getValue() {
return value_;
}
/**
* <code>optional bytes value = 6;</code>
*/
- public Builder setValue(com.google.protobuf.ByteString value) {
+ public Builder setValue(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1094,7 +1094,7 @@ public final class CellProtos {
return this;
}
- private com.google.protobuf.ByteString tags_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString tags_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>optional bytes tags = 7;</code>
*/
@@ -1104,13 +1104,13 @@ public final class CellProtos {
/**
* <code>optional bytes tags = 7;</code>
*/
- public com.google.protobuf.ByteString getTags() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getTags() {
return tags_;
}
/**
* <code>optional bytes tags = 7;</code>
*/
- public Builder setTags(com.google.protobuf.ByteString value) {
+ public Builder setTags(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1129,12 +1129,12 @@ public final class CellProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -1152,22 +1152,22 @@ public final class CellProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<Cell>
- PARSER = new com.google.protobuf.AbstractParser<Cell>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Cell>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Cell>() {
public Cell parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new Cell(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<Cell> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Cell> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<Cell> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Cell> getParserForType() {
return PARSER;
}
@@ -1179,7 +1179,7 @@ public final class CellProtos {
public interface KeyValueOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.KeyValue)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required bytes row = 1;</code>
@@ -1188,7 +1188,7 @@ public final class CellProtos {
/**
* <code>required bytes row = 1;</code>
*/
- com.google.protobuf.ByteString getRow();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getRow();
/**
* <code>required bytes family = 2;</code>
@@ -1197,7 +1197,7 @@ public final class CellProtos {
/**
* <code>required bytes family = 2;</code>
*/
- com.google.protobuf.ByteString getFamily();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamily();
/**
* <code>required bytes qualifier = 3;</code>
@@ -1206,7 +1206,7 @@ public final class CellProtos {
/**
* <code>required bytes qualifier = 3;</code>
*/
- com.google.protobuf.ByteString getQualifier();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getQualifier();
/**
* <code>optional uint64 timestamp = 4;</code>
@@ -1233,7 +1233,7 @@ public final class CellProtos {
/**
* <code>optional bytes value = 6;</code>
*/
- com.google.protobuf.ByteString getValue();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getValue();
/**
* <code>optional bytes tags = 7;</code>
@@ -1242,7 +1242,7 @@ public final class CellProtos {
/**
* <code>optional bytes tags = 7;</code>
*/
- com.google.protobuf.ByteString getTags();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getTags();
}
/**
* <pre>
@@ -1254,36 +1254,36 @@ public final class CellProtos {
* Protobuf type {@code hbase.pb.KeyValue}
*/
public static final class KeyValue extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.KeyValue)
KeyValueOrBuilder {
// Use KeyValue.newBuilder() to construct.
- private KeyValue(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private KeyValue(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private KeyValue() {
- row_ = com.google.protobuf.ByteString.EMPTY;
- family_ = com.google.protobuf.ByteString.EMPTY;
- qualifier_ = com.google.protobuf.ByteString.EMPTY;
+ row_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
+ family_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
+ qualifier_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
timestamp_ = 0L;
keyType_ = 0;
- value_ = com.google.protobuf.ByteString.EMPTY;
- tags_ = com.google.protobuf.ByteString.EMPTY;
+ value_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
+ tags_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private KeyValue(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -1342,22 +1342,22 @@ public final class CellProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.internal_static_hbase_pb_KeyValue_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.internal_static_hbase_pb_KeyValue_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1366,7 +1366,7 @@ public final class CellProtos {
private int bitField0_;
public static final int ROW_FIELD_NUMBER = 1;
- private com.google.protobuf.ByteString row_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString row_;
/**
* <code>required bytes row = 1;</code>
*/
@@ -1376,12 +1376,12 @@ public final class CellProtos {
/**
* <code>required bytes row = 1;</code>
*/
- public com.google.protobuf.ByteString getRow() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getRow() {
return row_;
}
public static final int FAMILY_FIELD_NUMBER = 2;
- private com.google.protobuf.ByteString family_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString family_;
/**
* <code>required bytes family = 2;</code>
*/
@@ -1391,12 +1391,12 @@ public final class CellProtos {
/**
* <code>required bytes family = 2;</code>
*/
- public com.google.protobuf.ByteString getFamily() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamily() {
return family_;
}
public static final int QUALIFIER_FIELD_NUMBER = 3;
- private com.google.protobuf.ByteString qualifier_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString qualifier_;
/**
* <code>required bytes qualifier = 3;</code>
*/
@@ -1406,7 +1406,7 @@ public final class CellProtos {
/**
* <code>required bytes qualifier = 3;</code>
*/
- public com.google.protobuf.ByteString getQualifier() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getQualifier() {
return qualifier_;
}
@@ -1442,7 +1442,7 @@ public final class CellProtos {
}
public static final int VALUE_FIELD_NUMBER = 6;
- private com.google.protobuf.ByteString value_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value_;
/**
* <code>optional bytes value = 6;</code>
*/
@@ -1452,12 +1452,12 @@ public final class CellProtos {
/**
* <code>optional bytes value = 6;</code>
*/
- public com.google.protobuf.ByteString getValue() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getValue() {
return value_;
}
public static final int TAGS_FIELD_NUMBER = 7;
- private com.google.protobuf.ByteString tags_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString tags_;
/**
* <code>optional bytes tags = 7;</code>
*/
@@ -1467,7 +1467,7 @@ public final class CellProtos {
/**
* <code>optional bytes tags = 7;</code>
*/
- public com.google.protobuf.ByteString getTags() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getTags() {
return tags_;
}
@@ -1493,7 +1493,7 @@ public final class CellProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, row_);
@@ -1525,31 +1525,31 @@ public final class CellProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(1, row_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(2, family_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(3, qualifier_);
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(4, timestamp_);
}
if (((bitField0_ & 0x00000010) == 0x00000010)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeEnumSize(5, keyType_);
}
if (((bitField0_ & 0x00000020) == 0x00000020)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(6, value_);
}
if (((bitField0_ & 0x00000040) == 0x00000040)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(7, tags_);
}
size += unknownFields.getSerializedSize();
@@ -1628,7 +1628,7 @@ public final class CellProtos {
}
if (hasTimestamp()) {
hash = (37 * hash) + TIMESTAMP_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getTimestamp());
}
if (hasKeyType()) {
@@ -1649,61 +1649,61 @@ public final class CellProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.KeyValue parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.KeyValue parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.KeyValue parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.KeyValue parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.KeyValue parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.KeyValue parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.KeyValue parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.KeyValue parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.KeyValue parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.KeyValue parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -1721,7 +1721,7 @@ public final class CellProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1735,15 +1735,15 @@ public final class CellProtos {
* Protobuf type {@code hbase.pb.KeyValue}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.KeyValue)
org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.KeyValueOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.internal_static_hbase_pb_KeyValue_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.internal_static_hbase_pb_KeyValue_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1756,35 +1756,35 @@ public final class CellProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
public Builder clear() {
super.clear();
- row_ = com.google.protobuf.ByteString.EMPTY;
+ row_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000001);
- family_ = com.google.protobuf.ByteString.EMPTY;
+ family_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000002);
- qualifier_ = com.google.protobuf.ByteString.EMPTY;
+ qualifier_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000004);
timestamp_ = 0L;
bitField0_ = (bitField0_ & ~0x00000008);
keyType_ = 0;
bitField0_ = (bitField0_ & ~0x00000010);
- value_ = com.google.protobuf.ByteString.EMPTY;
+ value_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000020);
- tags_ = com.google.protobuf.ByteString.EMPTY;
+ tags_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000040);
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.internal_static_hbase_pb_KeyValue_descriptor;
}
@@ -1842,29 +1842,29 @@ public final class CellProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.KeyValue) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.KeyValue)other);
} else {
@@ -1915,13 +1915,13 @@ public final class CellProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.KeyValue parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.CellProtos.KeyValue) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -1933,7 +1933,7 @@ public final class CellProtos {
}
private int bitField0_;
- private com.google.protobuf.ByteString row_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString row_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>required bytes row = 1;</code>
*/
@@ -1943,13 +1943,13 @@ public final class CellProtos {
/**
* <code>required bytes row = 1;</code>
*/
- public com.google.protobuf.ByteString getRow() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getRow() {
return row_;
}
/**
* <code>required bytes row = 1;</code>
*/
- public Builder setRow(com.google.protobuf.ByteString value) {
+ public Builder setRow(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1968,7 +1968,7 @@ public final class CellProtos {
return this;
}
- private com.google.protobuf.ByteString family_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString family_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>required bytes family = 2;</code>
*/
@@ -1978,13 +1978,13 @@ public final class CellProtos {
/**
* <code>required bytes family = 2;</code>
*/
- public com.google.protobuf.ByteString getFamily() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamily() {
return family_;
}
/**
* <code>required bytes family = 2;</code>
*/
- public Builder setFamily(com.google.protobuf.ByteString value) {
+ public Builder setFamily(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -2003,7 +2003,7 @@ public final class CellProtos {
return this;
}
- private com.google.protobuf.ByteString qualifier_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString qualifier_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>required bytes qualifier = 3;</code>
*/
@@ -2013,13 +2013,13 @@ public final class CellProtos {
/**
* <code>required bytes qualifier = 3;</code>
*/
- public com.google.protobuf.ByteString getQualifier() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getQualifier() {
return qualifier_;
}
/**
* <code>required bytes qualifier = 3;</code>
*/
- public Builder setQualifier(com.google.protobuf.ByteString value) {
+ public Builder setQualifier(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -2106,7 +2106,7 @@ public final class CellProtos {
return this;
}
- private com.google.protobuf.ByteString value_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>optional bytes value = 6;</code>
*/
@@ -2116,13 +2116,13 @@ public final class CellProtos {
/**
* <code>optional bytes value = 6;</code>
*/
- public com.google.protobuf.ByteString getValue() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getValue() {
return value_;
}
/**
* <code>optional bytes value = 6;</code>
*/
- public Builder setValue(com.google.protobuf.ByteString value) {
+ public Builder setValue(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -2141,7 +2141,7 @@ public final class CellProtos {
return this;
}
- private com.google.protobuf.ByteString tags_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString tags_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>optional bytes tags = 7;</code>
*/
@@ -2151,13 +2151,13 @@ public final class CellProtos {
/**
* <code>optional bytes tags = 7;</code>
*/
- public com.google.protobuf.ByteString getTags() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getTags() {
return tags_;
}
/**
* <code>optional bytes tags = 7;</code>
*/
- public Builder setTags(com.google.protobuf.ByteString value) {
+ public Builder setTags(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -2176,12 +2176,12 @@ public final class CellProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -2199,22 +2199,22 @@ public final class CellProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<KeyValue>
- PARSER = new com.google.protobuf.AbstractParser<KeyValue>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<KeyValue>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<KeyValue>() {
public KeyValue parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new KeyValue(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<KeyValue> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<KeyValue> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<KeyValue> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<KeyValue> getParserForType() {
return PARSER;
}
@@ -2224,22 +2224,22 @@ public final class CellProtos {
}
- private static final com.google.protobuf.Descriptors.Descriptor
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
internal_static_hbase_pb_Cell_descriptor;
private static final
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hbase_pb_Cell_fieldAccessorTable;
- private static final com.google.protobuf.Descriptors.Descriptor
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
internal_static_hbase_pb_KeyValue_descriptor;
private static final
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hbase_pb_KeyValue_fieldAccessorTable;
- public static com.google.protobuf.Descriptors.FileDescriptor
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
getDescriptor() {
return descriptor;
}
- private static com.google.protobuf.Descriptors.FileDescriptor
+ private static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
descriptor;
static {
java.lang.String[] descriptorData = {
@@ -2257,28 +2257,28 @@ public final class CellProtos {
"haded.protobuf.generatedB\nCellProtosH\001\240\001" +
"\001"
};
- com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
- new com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
- public com.google.protobuf.ExtensionRegistry assignDescriptors(
- com.google.protobuf.Descriptors.FileDescriptor root) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
descriptor = root;
return null;
}
};
- com.google.protobuf.Descriptors.FileDescriptor
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
.internalBuildGeneratedFileFrom(descriptorData,
- new com.google.protobuf.Descriptors.FileDescriptor[] {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
}, assigner);
internal_static_hbase_pb_Cell_descriptor =
getDescriptor().getMessageTypes().get(0);
internal_static_hbase_pb_Cell_fieldAccessorTable = new
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hbase_pb_Cell_descriptor,
new java.lang.String[] { "Row", "Family", "Qualifier", "Timestamp", "CellType", "Value", "Tags", });
internal_static_hbase_pb_KeyValue_descriptor =
getDescriptor().getMessageTypes().get(1);
internal_static_hbase_pb_KeyValue_fieldAccessorTable = new
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hbase_pb_KeyValue_descriptor,
new java.lang.String[] { "Row", "Family", "Qualifier", "Timestamp", "KeyType", "Value", "Tags", });
}
[18/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ErrorHandlingProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ErrorHandlingProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ErrorHandlingProtos.java
index d696e90..ccb7e86 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ErrorHandlingProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ErrorHandlingProtos.java
@@ -6,17 +6,17 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class ErrorHandlingProtos {
private ErrorHandlingProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
public interface StackTraceElementMessageOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.StackTraceElementMessage)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>optional string declaring_class = 1;</code>
@@ -29,7 +29,7 @@ public final class ErrorHandlingProtos {
/**
* <code>optional string declaring_class = 1;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getDeclaringClassBytes();
/**
@@ -43,7 +43,7 @@ public final class ErrorHandlingProtos {
/**
* <code>optional string method_name = 2;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getMethodNameBytes();
/**
@@ -57,7 +57,7 @@ public final class ErrorHandlingProtos {
/**
* <code>optional string file_name = 3;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getFileNameBytes();
/**
@@ -79,11 +79,11 @@ public final class ErrorHandlingProtos {
* Protobuf type {@code hbase.pb.StackTraceElementMessage}
*/
public static final class StackTraceElementMessage extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.StackTraceElementMessage)
StackTraceElementMessageOrBuilder {
// Use StackTraceElementMessage.newBuilder() to construct.
- private StackTraceElementMessage(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private StackTraceElementMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private StackTraceElementMessage() {
@@ -94,18 +94,18 @@ public final class ErrorHandlingProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private StackTraceElementMessage(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -122,19 +122,19 @@ public final class ErrorHandlingProtos {
break;
}
case 10: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000001;
declaringClass_ = bs;
break;
}
case 18: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000002;
methodName_ = bs;
break;
}
case 26: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000004;
fileName_ = bs;
break;
@@ -146,22 +146,22 @@ public final class ErrorHandlingProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.internal_static_hbase_pb_StackTraceElementMessage_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.internal_static_hbase_pb_StackTraceElementMessage_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -185,8 +185,8 @@ public final class ErrorHandlingProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
declaringClass_ = s;
@@ -197,17 +197,17 @@ public final class ErrorHandlingProtos {
/**
* <code>optional string declaring_class = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getDeclaringClassBytes() {
java.lang.Object ref = declaringClass_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
declaringClass_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -227,8 +227,8 @@ public final class ErrorHandlingProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
methodName_ = s;
@@ -239,17 +239,17 @@ public final class ErrorHandlingProtos {
/**
* <code>optional string method_name = 2;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getMethodNameBytes() {
java.lang.Object ref = methodName_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
methodName_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -269,8 +269,8 @@ public final class ErrorHandlingProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
fileName_ = s;
@@ -281,17 +281,17 @@ public final class ErrorHandlingProtos {
/**
* <code>optional string file_name = 3;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getFileNameBytes() {
java.lang.Object ref = fileName_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
fileName_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -320,16 +320,16 @@ public final class ErrorHandlingProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 1, declaringClass_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, declaringClass_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 2, methodName_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 2, methodName_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 3, fileName_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 3, fileName_);
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
output.writeInt32(4, lineNumber_);
@@ -343,16 +343,16 @@ public final class ErrorHandlingProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, declaringClass_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, declaringClass_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(2, methodName_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(2, methodName_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(3, fileName_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(3, fileName_);
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeInt32Size(4, lineNumber_);
}
size += unknownFields.getSerializedSize();
@@ -425,61 +425,61 @@ public final class ErrorHandlingProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessage parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessage parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessage parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessage parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessage parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessage parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessage parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessage parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessage parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessage parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -497,7 +497,7 @@ public final class ErrorHandlingProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -511,15 +511,15 @@ public final class ErrorHandlingProtos {
* Protobuf type {@code hbase.pb.StackTraceElementMessage}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.StackTraceElementMessage)
org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessageOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.internal_static_hbase_pb_StackTraceElementMessage_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.internal_static_hbase_pb_StackTraceElementMessage_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -532,12 +532,12 @@ public final class ErrorHandlingProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -554,7 +554,7 @@ public final class ErrorHandlingProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.internal_static_hbase_pb_StackTraceElementMessage_descriptor;
}
@@ -600,29 +600,29 @@ public final class ErrorHandlingProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessage) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessage)other);
} else {
@@ -661,13 +661,13 @@ public final class ErrorHandlingProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessage parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessage) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -692,8 +692,8 @@ public final class ErrorHandlingProtos {
public java.lang.String getDeclaringClass() {
java.lang.Object ref = declaringClass_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
declaringClass_ = s;
@@ -706,17 +706,17 @@ public final class ErrorHandlingProtos {
/**
* <code>optional string declaring_class = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getDeclaringClassBytes() {
java.lang.Object ref = declaringClass_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
declaringClass_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -745,7 +745,7 @@ public final class ErrorHandlingProtos {
* <code>optional string declaring_class = 1;</code>
*/
public Builder setDeclaringClassBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -768,8 +768,8 @@ public final class ErrorHandlingProtos {
public java.lang.String getMethodName() {
java.lang.Object ref = methodName_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
methodName_ = s;
@@ -782,17 +782,17 @@ public final class ErrorHandlingProtos {
/**
* <code>optional string method_name = 2;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getMethodNameBytes() {
java.lang.Object ref = methodName_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
methodName_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -821,7 +821,7 @@ public final class ErrorHandlingProtos {
* <code>optional string method_name = 2;</code>
*/
public Builder setMethodNameBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -844,8 +844,8 @@ public final class ErrorHandlingProtos {
public java.lang.String getFileName() {
java.lang.Object ref = fileName_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
fileName_ = s;
@@ -858,17 +858,17 @@ public final class ErrorHandlingProtos {
/**
* <code>optional string file_name = 3;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getFileNameBytes() {
java.lang.Object ref = fileName_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
fileName_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -897,7 +897,7 @@ public final class ErrorHandlingProtos {
* <code>optional string file_name = 3;</code>
*/
public Builder setFileNameBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -939,12 +939,12 @@ public final class ErrorHandlingProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -962,22 +962,22 @@ public final class ErrorHandlingProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<StackTraceElementMessage>
- PARSER = new com.google.protobuf.AbstractParser<StackTraceElementMessage>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<StackTraceElementMessage>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<StackTraceElementMessage>() {
public StackTraceElementMessage parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new StackTraceElementMessage(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<StackTraceElementMessage> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<StackTraceElementMessage> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<StackTraceElementMessage> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<StackTraceElementMessage> getParserForType() {
return PARSER;
}
@@ -989,7 +989,7 @@ public final class ErrorHandlingProtos {
public interface GenericExceptionMessageOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.GenericExceptionMessage)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>optional string class_name = 1;</code>
@@ -1002,7 +1002,7 @@ public final class ErrorHandlingProtos {
/**
* <code>optional string class_name = 1;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getClassNameBytes();
/**
@@ -1016,7 +1016,7 @@ public final class ErrorHandlingProtos {
/**
* <code>optional string message = 2;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getMessageBytes();
/**
@@ -1026,7 +1026,7 @@ public final class ErrorHandlingProtos {
/**
* <code>optional bytes error_info = 3;</code>
*/
- com.google.protobuf.ByteString getErrorInfo();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getErrorInfo();
/**
* <code>repeated .hbase.pb.StackTraceElementMessage trace = 4;</code>
@@ -1064,33 +1064,33 @@ public final class ErrorHandlingProtos {
* Protobuf type {@code hbase.pb.GenericExceptionMessage}
*/
public static final class GenericExceptionMessage extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.GenericExceptionMessage)
GenericExceptionMessageOrBuilder {
// Use GenericExceptionMessage.newBuilder() to construct.
- private GenericExceptionMessage(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private GenericExceptionMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private GenericExceptionMessage() {
className_ = "";
message_ = "";
- errorInfo_ = com.google.protobuf.ByteString.EMPTY;
+ errorInfo_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
trace_ = java.util.Collections.emptyList();
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private GenericExceptionMessage(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -1107,13 +1107,13 @@ public final class ErrorHandlingProtos {
break;
}
case 10: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000001;
className_ = bs;
break;
}
case 18: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000002;
message_ = bs;
break;
@@ -1134,10 +1134,10 @@ public final class ErrorHandlingProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000008) == 0x00000008)) {
@@ -1147,12 +1147,12 @@ public final class ErrorHandlingProtos {
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.internal_static_hbase_pb_GenericExceptionMessage_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.internal_static_hbase_pb_GenericExceptionMessage_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1176,8 +1176,8 @@ public final class ErrorHandlingProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
className_ = s;
@@ -1188,17 +1188,17 @@ public final class ErrorHandlingProtos {
/**
* <code>optional string class_name = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getClassNameBytes() {
java.lang.Object ref = className_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
className_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -1218,8 +1218,8 @@ public final class ErrorHandlingProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
message_ = s;
@@ -1230,22 +1230,22 @@ public final class ErrorHandlingProtos {
/**
* <code>optional string message = 2;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getMessageBytes() {
java.lang.Object ref = message_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
message_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
public static final int ERROR_INFO_FIELD_NUMBER = 3;
- private com.google.protobuf.ByteString errorInfo_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString errorInfo_;
/**
* <code>optional bytes error_info = 3;</code>
*/
@@ -1255,7 +1255,7 @@ public final class ErrorHandlingProtos {
/**
* <code>optional bytes error_info = 3;</code>
*/
- public com.google.protobuf.ByteString getErrorInfo() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getErrorInfo() {
return errorInfo_;
}
@@ -1304,13 +1304,13 @@ public final class ErrorHandlingProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 1, className_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, className_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 2, message_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 2, message_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
output.writeBytes(3, errorInfo_);
@@ -1327,17 +1327,17 @@ public final class ErrorHandlingProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, className_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, className_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(2, message_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(2, message_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(3, errorInfo_);
}
for (int i = 0; i < trace_.size(); i++) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(4, trace_.get(i));
}
size += unknownFields.getSerializedSize();
@@ -1407,61 +1407,61 @@ public final class ErrorHandlingProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -1479,7 +1479,7 @@ public final class ErrorHandlingProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1495,15 +1495,15 @@ public final class ErrorHandlingProtos {
* Protobuf type {@code hbase.pb.GenericExceptionMessage}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.GenericExceptionMessage)
org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessageOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.internal_static_hbase_pb_GenericExceptionMessage_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.internal_static_hbase_pb_GenericExceptionMessage_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1516,12 +1516,12 @@ public final class ErrorHandlingProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getTraceFieldBuilder();
}
@@ -1532,7 +1532,7 @@ public final class ErrorHandlingProtos {
bitField0_ = (bitField0_ & ~0x00000001);
message_ = "";
bitField0_ = (bitField0_ & ~0x00000002);
- errorInfo_ = com.google.protobuf.ByteString.EMPTY;
+ errorInfo_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000004);
if (traceBuilder_ == null) {
trace_ = java.util.Collections.emptyList();
@@ -1543,7 +1543,7 @@ public final class ErrorHandlingProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.internal_static_hbase_pb_GenericExceptionMessage_descriptor;
}
@@ -1594,29 +1594,29 @@ public final class ErrorHandlingProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage)other);
} else {
@@ -1659,7 +1659,7 @@ public final class ErrorHandlingProtos {
trace_ = other.trace_;
bitField0_ = (bitField0_ & ~0x00000008);
traceBuilder_ =
- com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getTraceFieldBuilder() : null;
} else {
traceBuilder_.addAllMessages(other.trace_);
@@ -1676,13 +1676,13 @@ public final class ErrorHandlingProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -1707,8 +1707,8 @@ public final class ErrorHandlingProtos {
public java.lang.String getClassName() {
java.lang.Object ref = className_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
className_ = s;
@@ -1721,17 +1721,17 @@ public final class ErrorHandlingProtos {
/**
* <code>optional string class_name = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getClassNameBytes() {
java.lang.Object ref = className_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
className_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -1760,7 +1760,7 @@ public final class ErrorHandlingProtos {
* <code>optional string class_name = 1;</code>
*/
public Builder setClassNameBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1783,8 +1783,8 @@ public final class ErrorHandlingProtos {
public java.lang.String getMessage() {
java.lang.Object ref = message_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
message_ = s;
@@ -1797,17 +1797,17 @@ public final class ErrorHandlingProtos {
/**
* <code>optional string message = 2;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getMessageBytes() {
java.lang.Object ref = message_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
message_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -1836,7 +1836,7 @@ public final class ErrorHandlingProtos {
* <code>optional string message = 2;</code>
*/
public Builder setMessageBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1846,7 +1846,7 @@ public final class ErrorHandlingProtos {
return this;
}
- private com.google.protobuf.ByteString errorInfo_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString errorInfo_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>optional bytes error_info = 3;</code>
*/
@@ -1856,13 +1856,13 @@ public final class ErrorHandlingProtos {
/**
* <code>optional bytes error_info = 3;</code>
*/
- public com.google.protobuf.ByteString getErrorInfo() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getErrorInfo() {
return errorInfo_;
}
/**
* <code>optional bytes error_info = 3;</code>
*/
- public Builder setErrorInfo(com.google.protobuf.ByteString value) {
+ public Builder setErrorInfo(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1890,7 +1890,7 @@ public final class ErrorHandlingProtos {
}
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessage, org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessage.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessageOrBuilder> traceBuilder_;
/**
@@ -2022,7 +2022,7 @@ public final class ErrorHandlingProtos {
java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessage> values) {
if (traceBuilder_ == null) {
ensureTraceIsMutable();
- com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
values, trace_);
onChanged();
} else {
@@ -2106,11 +2106,11 @@ public final class ErrorHandlingProtos {
getTraceBuilderList() {
return getTraceFieldBuilder().getBuilderList();
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessage, org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessage.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessageOrBuilder>
getTraceFieldBuilder() {
if (traceBuilder_ == null) {
- traceBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3<
+ traceBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessage, org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessage.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessageOrBuilder>(
trace_,
((bitField0_ & 0x00000008) == 0x00000008),
@@ -2121,12 +2121,12 @@ public final class ErrorHandlingProtos {
return traceBuilder_;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -2144,22 +2144,22 @@ public final class ErrorHandlingProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<GenericExceptionMessage>
- PARSER = new com.google.protobuf.AbstractParser<GenericExceptionMessage>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<GenericExceptionMessage>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<GenericExceptionMessage>() {
public GenericExceptionMessage parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new GenericExceptionMessage(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<GenericExceptionMessage> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<GenericExceptionMessage> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<GenericExceptionMessage> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<GenericExceptionMessage> getParserForType() {
return PARSER;
}
@@ -2171,7 +2171,7 @@ public final class ErrorHandlingProtos {
public interface ForeignExceptionMessageOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.ForeignExceptionMessage)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>optional string source = 1;</code>
@@ -2184,7 +2184,7 @@ public final class ErrorHandlingProtos {
/**
* <code>optional string source = 1;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getSourceBytes();
/**
@@ -2210,11 +2210,11 @@ public final class ErrorHandlingProtos {
* Protobuf type {@code hbase.pb.ForeignExceptionMessage}
*/
public static final class ForeignExceptionMessage extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.ForeignExceptionMessage)
ForeignExceptionMessageOrBuilder {
// Use ForeignExceptionMessage.newBuilder() to construct.
- private ForeignExceptionMessage(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private ForeignExceptionMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private ForeignExceptionMessage() {
@@ -2222,18 +2222,18 @@ public final class ErrorHandlingProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private ForeignExceptionMessage(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -2250,7 +2250,7 @@ public final class ErrorHandlingProtos {
break;
}
case 10: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000001;
source_ = bs;
break;
@@ -2270,22 +2270,22 @@ public final class ErrorHandlingProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.internal_static_hbase_pb_ForeignExceptionMessage_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.internal_static_hbase_pb_ForeignExceptionMessage_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2309,8 +2309,8 @@ public final class ErrorHandlingProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
source_ = s;
@@ -2321,17 +2321,17 @@ public final class ErrorHandlingProtos {
/**
* <code>optional string source = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getSourceBytes() {
java.lang.Object ref = source_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
source_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -2366,10 +2366,10 @@ public final class ErrorHandlingProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 1, source_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, source_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
output.writeMessage(2, getGenericException());
@@ -2383,10 +2383,10 @@ public final class ErrorHandlingProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, source_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, source_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(2, getGenericException());
}
size += unknownFields.getSerializedSize();
@@ -2441,61 +2441,61 @@ public final class ErrorHandlingProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -2513,7 +2513,7 @@ public final class ErrorHandlingProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -2527,15 +2527,15 @@ public final class ErrorHandlingProtos {
* Protobuf type {@code hbase.pb.ForeignExceptionMessage}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.ForeignExceptionMessage)
org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessageOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.internal_static_hbase_pb_ForeignExceptionMessage_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.internal_static_hbase_pb_ForeignExceptionMessage_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2548,12 +2548,12 @@ public final class ErrorHandlingProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getGenericExceptionFieldBuilder();
}
@@ -2571,7 +2571,7 @@ public final class ErrorHandlingProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.internal_static_hbase_pb_ForeignExceptionMessage_descriptor;
}
@@ -2613,29 +2613,29 @@ public final class ErrorHandlingProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage)other);
} else {
@@ -2664,13 +2664,13 @@ public final class ErrorHandlingProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -2695,8 +2695,8 @@ public final class ErrorHandlingProtos {
public java.lang.String getSource() {
java.lang.Object ref = source_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
source_ = s;
@@ -2709,17 +2709,17 @@ public final class ErrorHandlingProtos {
/**
* <code>optional string source = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getSourceBytes() {
java.lang.Object ref = source_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
source_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -2748,7 +2748,7 @@ public final class ErrorHandlingProtos {
* <code>optional string source = 1;</code>
*/
public Builder setSourceBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -2759,7 +2759,7 @@ public final class ErrorHandlingProtos {
}
private org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage genericException_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage, org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessageOrBuilder> genericExceptionBuilder_;
/**
* <code>optional .hbase.pb.GenericExceptionMessage generic_exception = 2;</code>
@@ -2862,11 +2862,11 @@ public final class ErrorHandlingProtos {
/**
* <code>optional .hbase.pb.GenericExceptionMessage generic_exception = 2;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage, org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessageOrBuilder>
getGenericExceptionFieldBuilder() {
if (genericExceptionBuilder_ == null) {
- genericExceptionBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ genericExceptionBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage, org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessageOrBuilder>(
getGenericException(),
getParentForChildren(),
@@ -2876,12 +2876,12 @@ public final class ErrorHandlingProtos {
return genericExceptionBuilder_;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -2899,22 +2899,22 @@ public final class ErrorHandlingProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final c
<TRUNCATED>
[03/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/WALProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/WALProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/WALProtos.java
index e032be7..db73e41 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/WALProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/WALProtos.java
@@ -6,19 +6,19 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class WALProtos {
private WALProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
/**
* Protobuf enum {@code hbase.pb.ScopeType}
*/
public enum ScopeType
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>REPLICATION_SCOPE_LOCAL = 0;</code>
*/
@@ -68,27 +68,27 @@ public final class WALProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<ScopeType>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<ScopeType>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
ScopeType> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<ScopeType>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<ScopeType>() {
public ScopeType findValueByNumber(int number) {
return ScopeType.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.getDescriptor().getEnumTypes().get(0);
}
@@ -96,7 +96,7 @@ public final class WALProtos {
private static final ScopeType[] VALUES = values();
public static ScopeType valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -115,7 +115,7 @@ public final class WALProtos {
public interface WALHeaderOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.WALHeader)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>optional bool has_compression = 1;</code>
@@ -133,7 +133,7 @@ public final class WALProtos {
/**
* <code>optional bytes encryption_key = 2;</code>
*/
- com.google.protobuf.ByteString getEncryptionKey();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getEncryptionKey();
/**
* <code>optional bool has_tag_compression = 3;</code>
@@ -155,7 +155,7 @@ public final class WALProtos {
/**
* <code>optional string writer_cls_name = 4;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getWriterClsNameBytes();
/**
@@ -169,41 +169,41 @@ public final class WALProtos {
/**
* <code>optional string cell_codec_cls_name = 5;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getCellCodecClsNameBytes();
}
/**
* Protobuf type {@code hbase.pb.WALHeader}
*/
public static final class WALHeader extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.WALHeader)
WALHeaderOrBuilder {
// Use WALHeader.newBuilder() to construct.
- private WALHeader(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private WALHeader(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private WALHeader() {
hasCompression_ = false;
- encryptionKey_ = com.google.protobuf.ByteString.EMPTY;
+ encryptionKey_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
hasTagCompression_ = false;
writerClsName_ = "";
cellCodecClsName_ = "";
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private WALHeader(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -235,35 +235,35 @@ public final class WALProtos {
break;
}
case 34: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000008;
writerClsName_ = bs;
break;
}
case 42: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000010;
cellCodecClsName_ = bs;
break;
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.internal_static_hbase_pb_WALHeader_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.internal_static_hbase_pb_WALHeader_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -287,7 +287,7 @@ public final class WALProtos {
}
public static final int ENCRYPTION_KEY_FIELD_NUMBER = 2;
- private com.google.protobuf.ByteString encryptionKey_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString encryptionKey_;
/**
* <code>optional bytes encryption_key = 2;</code>
*/
@@ -297,7 +297,7 @@ public final class WALProtos {
/**
* <code>optional bytes encryption_key = 2;</code>
*/
- public com.google.protobuf.ByteString getEncryptionKey() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getEncryptionKey() {
return encryptionKey_;
}
@@ -332,8 +332,8 @@ public final class WALProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
writerClsName_ = s;
@@ -344,17 +344,17 @@ public final class WALProtos {
/**
* <code>optional string writer_cls_name = 4;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getWriterClsNameBytes() {
java.lang.Object ref = writerClsName_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
writerClsName_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -374,8 +374,8 @@ public final class WALProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
cellCodecClsName_ = s;
@@ -386,17 +386,17 @@ public final class WALProtos {
/**
* <code>optional string cell_codec_cls_name = 5;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getCellCodecClsNameBytes() {
java.lang.Object ref = cellCodecClsName_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
cellCodecClsName_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -410,7 +410,7 @@ public final class WALProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBool(1, hasCompression_);
@@ -422,10 +422,10 @@ public final class WALProtos {
output.writeBool(3, hasTagCompression_);
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 4, writerClsName_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 4, writerClsName_);
}
if (((bitField0_ & 0x00000010) == 0x00000010)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 5, cellCodecClsName_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 5, cellCodecClsName_);
}
unknownFields.writeTo(output);
}
@@ -436,22 +436,22 @@ public final class WALProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBoolSize(1, hasCompression_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(2, encryptionKey_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBoolSize(3, hasTagCompression_);
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(4, writerClsName_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(4, writerClsName_);
}
if (((bitField0_ & 0x00000010) == 0x00000010)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(5, cellCodecClsName_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(5, cellCodecClsName_);
}
size += unknownFields.getSerializedSize();
memoizedSize = size;
@@ -508,7 +508,7 @@ public final class WALProtos {
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasHasCompression()) {
hash = (37 * hash) + HAS_COMPRESSION_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean(
getHasCompression());
}
if (hasEncryptionKey()) {
@@ -517,7 +517,7 @@ public final class WALProtos {
}
if (hasHasTagCompression()) {
hash = (37 * hash) + HAS_TAG_COMPRESSION_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean(
getHasTagCompression());
}
if (hasWriterClsName()) {
@@ -534,61 +534,61 @@ public final class WALProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALHeader parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALHeader parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALHeader parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALHeader parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALHeader parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALHeader parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALHeader parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALHeader parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALHeader parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALHeader parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -606,7 +606,7 @@ public final class WALProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -614,15 +614,15 @@ public final class WALProtos {
* Protobuf type {@code hbase.pb.WALHeader}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.WALHeader)
org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALHeaderOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.internal_static_hbase_pb_WALHeader_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.internal_static_hbase_pb_WALHeader_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -635,12 +635,12 @@ public final class WALProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -648,7 +648,7 @@ public final class WALProtos {
super.clear();
hasCompression_ = false;
bitField0_ = (bitField0_ & ~0x00000001);
- encryptionKey_ = com.google.protobuf.ByteString.EMPTY;
+ encryptionKey_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000002);
hasTagCompression_ = false;
bitField0_ = (bitField0_ & ~0x00000004);
@@ -659,7 +659,7 @@ public final class WALProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.internal_static_hbase_pb_WALHeader_descriptor;
}
@@ -709,29 +709,29 @@ public final class WALProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALHeader) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALHeader)other);
} else {
@@ -771,13 +771,13 @@ public final class WALProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALHeader parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALHeader) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -821,7 +821,7 @@ public final class WALProtos {
return this;
}
- private com.google.protobuf.ByteString encryptionKey_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString encryptionKey_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>optional bytes encryption_key = 2;</code>
*/
@@ -831,13 +831,13 @@ public final class WALProtos {
/**
* <code>optional bytes encryption_key = 2;</code>
*/
- public com.google.protobuf.ByteString getEncryptionKey() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getEncryptionKey() {
return encryptionKey_;
}
/**
* <code>optional bytes encryption_key = 2;</code>
*/
- public Builder setEncryptionKey(com.google.protobuf.ByteString value) {
+ public Builder setEncryptionKey(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -901,8 +901,8 @@ public final class WALProtos {
public java.lang.String getWriterClsName() {
java.lang.Object ref = writerClsName_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
writerClsName_ = s;
@@ -915,17 +915,17 @@ public final class WALProtos {
/**
* <code>optional string writer_cls_name = 4;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getWriterClsNameBytes() {
java.lang.Object ref = writerClsName_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
writerClsName_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -954,7 +954,7 @@ public final class WALProtos {
* <code>optional string writer_cls_name = 4;</code>
*/
public Builder setWriterClsNameBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -977,8 +977,8 @@ public final class WALProtos {
public java.lang.String getCellCodecClsName() {
java.lang.Object ref = cellCodecClsName_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
cellCodecClsName_ = s;
@@ -991,17 +991,17 @@ public final class WALProtos {
/**
* <code>optional string cell_codec_cls_name = 5;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getCellCodecClsNameBytes() {
java.lang.Object ref = cellCodecClsName_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
cellCodecClsName_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -1030,7 +1030,7 @@ public final class WALProtos {
* <code>optional string cell_codec_cls_name = 5;</code>
*/
public Builder setCellCodecClsNameBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1040,12 +1040,12 @@ public final class WALProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -1063,22 +1063,22 @@ public final class WALProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<WALHeader>
- PARSER = new com.google.protobuf.AbstractParser<WALHeader>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<WALHeader>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<WALHeader>() {
public WALHeader parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new WALHeader(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<WALHeader> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<WALHeader> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<WALHeader> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<WALHeader> getParserForType() {
return PARSER;
}
@@ -1090,7 +1090,7 @@ public final class WALProtos {
public interface WALKeyOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.WALKey)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required bytes encoded_region_name = 1;</code>
@@ -1099,7 +1099,7 @@ public final class WALProtos {
/**
* <code>required bytes encoded_region_name = 1;</code>
*/
- com.google.protobuf.ByteString getEncodedRegionName();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getEncodedRegionName();
/**
* <code>required bytes table_name = 2;</code>
@@ -1108,7 +1108,7 @@ public final class WALProtos {
/**
* <code>required bytes table_name = 2;</code>
*/
- com.google.protobuf.ByteString getTableName();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getTableName();
/**
* <code>required uint64 log_sequence_number = 3;</code>
@@ -1280,16 +1280,16 @@ public final class WALProtos {
* Protobuf type {@code hbase.pb.WALKey}
*/
public static final class WALKey extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.WALKey)
WALKeyOrBuilder {
// Use WALKey.newBuilder() to construct.
- private WALKey(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private WALKey(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private WALKey() {
- encodedRegionName_ = com.google.protobuf.ByteString.EMPTY;
- tableName_ = com.google.protobuf.ByteString.EMPTY;
+ encodedRegionName_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
+ tableName_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
logSequenceNumber_ = 0L;
writeTime_ = 0L;
scopes_ = java.util.Collections.emptyList();
@@ -1301,18 +1301,18 @@ public final class WALProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private WALKey(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -1401,10 +1401,10 @@ public final class WALProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000020) == 0x00000020)) {
@@ -1417,12 +1417,12 @@ public final class WALProtos {
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.internal_static_hbase_pb_WALKey_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.internal_static_hbase_pb_WALKey_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1431,7 +1431,7 @@ public final class WALProtos {
private int bitField0_;
public static final int ENCODED_REGION_NAME_FIELD_NUMBER = 1;
- private com.google.protobuf.ByteString encodedRegionName_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString encodedRegionName_;
/**
* <code>required bytes encoded_region_name = 1;</code>
*/
@@ -1441,12 +1441,12 @@ public final class WALProtos {
/**
* <code>required bytes encoded_region_name = 1;</code>
*/
- public com.google.protobuf.ByteString getEncodedRegionName() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getEncodedRegionName() {
return encodedRegionName_;
}
public static final int TABLE_NAME_FIELD_NUMBER = 2;
- private com.google.protobuf.ByteString tableName_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString tableName_;
/**
* <code>required bytes table_name = 2;</code>
*/
@@ -1456,7 +1456,7 @@ public final class WALProtos {
/**
* <code>required bytes table_name = 2;</code>
*/
- public com.google.protobuf.ByteString getTableName() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getTableName() {
return tableName_;
}
@@ -1731,7 +1731,7 @@ public final class WALProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, encodedRegionName_);
@@ -1775,47 +1775,47 @@ public final class WALProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(1, encodedRegionName_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(2, tableName_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(3, logSequenceNumber_);
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(4, writeTime_);
}
if (((bitField0_ & 0x00000010) == 0x00000010)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(5, getClusterId());
}
for (int i = 0; i < scopes_.size(); i++) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(6, scopes_.get(i));
}
if (((bitField0_ & 0x00000020) == 0x00000020)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt32Size(7, followingKvCount_);
}
for (int i = 0; i < clusterIds_.size(); i++) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(8, clusterIds_.get(i));
}
if (((bitField0_ & 0x00000040) == 0x00000040)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(9, nonceGroup_);
}
if (((bitField0_ & 0x00000080) == 0x00000080)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(10, nonce_);
}
if (((bitField0_ & 0x00000100) == 0x00000100)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(11, origSequenceNumber_);
}
size += unknownFields.getSerializedSize();
@@ -1905,12 +1905,12 @@ public final class WALProtos {
}
if (hasLogSequenceNumber()) {
hash = (37 * hash) + LOG_SEQUENCE_NUMBER_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getLogSequenceNumber());
}
if (hasWriteTime()) {
hash = (37 * hash) + WRITE_TIME_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getWriteTime());
}
if (hasClusterId()) {
@@ -1931,17 +1931,17 @@ public final class WALProtos {
}
if (hasNonceGroup()) {
hash = (37 * hash) + NONCEGROUP_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getNonceGroup());
}
if (hasNonce()) {
hash = (37 * hash) + NONCE_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getNonce());
}
if (hasOrigSequenceNumber()) {
hash = (37 * hash) + ORIG_SEQUENCE_NUMBER_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getOrigSequenceNumber());
}
hash = (29 * hash) + unknownFields.hashCode();
@@ -1950,61 +1950,61 @@ public final class WALProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALKey parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALKey parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALKey parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALKey parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALKey parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALKey parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALKey parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALKey parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALKey parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALKey parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -2022,7 +2022,7 @@ public final class WALProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -2035,15 +2035,15 @@ public final class WALProtos {
* Protobuf type {@code hbase.pb.WALKey}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.WALKey)
org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALKeyOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.internal_static_hbase_pb_WALKey_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.internal_static_hbase_pb_WALKey_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2056,12 +2056,12 @@ public final class WALProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getClusterIdFieldBuilder();
getScopesFieldBuilder();
@@ -2070,9 +2070,9 @@ public final class WALProtos {
}
public Builder clear() {
super.clear();
- encodedRegionName_ = com.google.protobuf.ByteString.EMPTY;
+ encodedRegionName_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000001);
- tableName_ = com.google.protobuf.ByteString.EMPTY;
+ tableName_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000002);
logSequenceNumber_ = 0L;
bitField0_ = (bitField0_ & ~0x00000004);
@@ -2107,7 +2107,7 @@ public final class WALProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.internal_static_hbase_pb_WALKey_descriptor;
}
@@ -2195,29 +2195,29 @@ public final class WALProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALKey) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALKey)other);
} else {
@@ -2262,7 +2262,7 @@ public final class WALProtos {
scopes_ = other.scopes_;
bitField0_ = (bitField0_ & ~0x00000020);
scopesBuilder_ =
- com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getScopesFieldBuilder() : null;
} else {
scopesBuilder_.addAllMessages(other.scopes_);
@@ -2291,7 +2291,7 @@ public final class WALProtos {
clusterIds_ = other.clusterIds_;
bitField0_ = (bitField0_ & ~0x00000080);
clusterIdsBuilder_ =
- com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getClusterIdsFieldBuilder() : null;
} else {
clusterIdsBuilder_.addAllMessages(other.clusterIds_);
@@ -2344,13 +2344,13 @@ public final class WALProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALKey parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALKey) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -2362,7 +2362,7 @@ public final class WALProtos {
}
private int bitField0_;
- private com.google.protobuf.ByteString encodedRegionName_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString encodedRegionName_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>required bytes encoded_region_name = 1;</code>
*/
@@ -2372,13 +2372,13 @@ public final class WALProtos {
/**
* <code>required bytes encoded_region_name = 1;</code>
*/
- public com.google.protobuf.ByteString getEncodedRegionName() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getEncodedRegionName() {
return encodedRegionName_;
}
/**
* <code>required bytes encoded_region_name = 1;</code>
*/
- public Builder setEncodedRegionName(com.google.protobuf.ByteString value) {
+ public Builder setEncodedRegionName(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -2397,7 +2397,7 @@ public final class WALProtos {
return this;
}
- private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString tableName_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>required bytes table_name = 2;</code>
*/
@@ -2407,13 +2407,13 @@ public final class WALProtos {
/**
* <code>required bytes table_name = 2;</code>
*/
- public com.google.protobuf.ByteString getTableName() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getTableName() {
return tableName_;
}
/**
* <code>required bytes table_name = 2;</code>
*/
- public Builder setTableName(com.google.protobuf.ByteString value) {
+ public Builder setTableName(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -2497,7 +2497,7 @@ public final class WALProtos {
}
private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.UUID clusterId_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.UUID, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.UUID.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.UUIDOrBuilder> clusterIdBuilder_;
/**
* <pre>
@@ -2663,11 +2663,11 @@ public final class WALProtos {
*
* <code>optional .hbase.pb.UUID cluster_id = 5 [deprecated = true];</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.UUID, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.UUID.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.UUIDOrBuilder>
getClusterIdFieldBuilder() {
if (clusterIdBuilder_ == null) {
- clusterIdBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ clusterIdBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.UUID, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.UUID.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.UUIDOrBuilder>(
getClusterId(),
getParentForChildren(),
@@ -2686,7 +2686,7 @@ public final class WALProtos {
}
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScope, org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScope.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScopeOrBuilder> scopesBuilder_;
/**
@@ -2818,7 +2818,7 @@ public final class WALProtos {
java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScope> values) {
if (scopesBuilder_ == null) {
ensureScopesIsMutable();
- com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
values, scopes_);
onChanged();
} else {
@@ -2902,11 +2902,11 @@ public final class WALProtos {
getScopesBuilderList() {
return getScopesFieldBuilder().getBuilderList();
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScope, org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScope.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScopeOrBuilder>
getScopesFieldBuilder() {
if (scopesBuilder_ == null) {
- scopesBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3<
+ scopesBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScope, org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScope.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScopeOrBuilder>(
scopes_,
((bitField0_ & 0x00000020) == 0x00000020),
@@ -2958,7 +2958,7 @@ public final class WALProtos {
}
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.UUID, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.UUID.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.UUIDOrBuilder> clusterIdsBuilder_;
/**
@@ -3140,7 +3140,7 @@ public final class WALProtos {
java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.UUID> values) {
if (clusterIdsBuilder_ == null) {
ensureClusterIdsIsMutable();
- com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
values, clusterIds_);
onChanged();
} else {
@@ -3264,11 +3264,11 @@ public final class WALProtos {
getClusterIdsBuilderList() {
return getClusterIdsFieldBuilder().getBuilderList();
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.UUID, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.UUID.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.UUIDOrBuilder>
getClusterIdsFieldBuilder() {
if (clusterIdsBuilder_ == null) {
- clusterIdsBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3<
+ clusterIdsBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.UUID, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.UUID.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.UUIDOrBuilder>(
clusterIds_,
((bitField0_ & 0x00000080) == 0x00000080),
@@ -3375,12 +3375,12 @@ public final class WALProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -3398,22 +3398,22 @@ public final class WALProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<WALKey>
- PARSER = new com.google.protobuf.AbstractParser<WALKey>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<WALKey>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<WALKey>() {
public WALKey parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new WALKey(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<WALKey> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<WALKey> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<WALKey> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<WALKey> getParserForType() {
return PARSER;
}
@@ -3425,7 +3425,7 @@ public final class WALProtos {
public interface FamilyScopeOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.FamilyScope)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required bytes family = 1;</code>
@@ -3434,7 +3434,7 @@ public final class WALProtos {
/**
* <code>required bytes family = 1;</code>
*/
- com.google.protobuf.ByteString getFamily();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamily();
/**
* <code>required .hbase.pb.ScopeType scope_type = 2;</code>
@@ -3449,31 +3449,31 @@ public final class WALProtos {
* Protobuf type {@code hbase.pb.FamilyScope}
*/
public static final class FamilyScope extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.FamilyScope)
FamilyScopeOrBuilder {
// Use FamilyScope.newBuilder() to construct.
- private FamilyScope(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private FamilyScope(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private FamilyScope() {
- family_ = com.google.protobuf.ByteString.EMPTY;
+ family_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
scopeType_ = 0;
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private FamilyScope(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -3507,22 +3507,22 @@ public final class WALProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.internal_static_hbase_pb_FamilyScope_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.internal_static_hbase_pb_FamilyScope_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -3531,7 +3531,7 @@ public final class WALProtos {
private int bitField0_;
public static final int FAMILY_FIELD_NUMBER = 1;
- private com.google.protobuf.ByteString family_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString family_;
/**
* <code>required bytes family = 1;</code>
*/
@@ -3541,7 +3541,7 @@ public final class WALProtos {
/**
* <code>required bytes family = 1;</code>
*/
- public com.google.protobuf.ByteString getFamily() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamily() {
return family_;
}
@@ -3579,7 +3579,7 @@ public final class WALProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, family_);
@@ -3596,11 +3596,11 @@ public final class WALProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(1, family_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeEnumSize(2, scopeType_);
}
size += unknownFields.getSerializedSize();
@@ -3654,61 +3654,61 @@ public final class WALProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScope parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScope parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScope parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScope parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScope parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScope parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScope parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScope parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScope parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScope parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -3726,7 +3726,7 @@ public final class WALProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -3734,15 +3734,15 @@ public final class WALProtos {
* Protobuf type {@code hbase.pb.FamilyScope}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.FamilyScope)
org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScopeOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.internal_static_hbase_pb_FamilyScope_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.internal_static_hbase_pb_FamilyScope_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -3755,25 +3755,25 @@ public final class WALProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
public Builder clear() {
super.clear();
- family_ = com.google.protobuf.ByteString.EMPTY;
+ family_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000001);
scopeType_ = 0;
bitField0_ = (bitField0_ & ~0x00000002);
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.internal_static_hbase_pb_FamilyScope_descriptor;
}
@@ -3811,29 +3811,29 @@ public final class WALProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScope) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScope)other);
} else {
@@ -3866,13 +3866,13 @@ public final class WALProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScope parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FamilyScope) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -3884,7 +3884,7 @@ public final class WALProtos {
}
private int bitField0_;
- private com.google.protobuf.ByteString family_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString family_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>required bytes family = 1;</code>
*/
@@ -3894,13 +3894,13 @@ public final class WALProtos {
/**
* <code>required bytes family = 1;</code>
*/
- public com.google.protobuf.ByteString getFamily() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamily() {
return family_;
}
/**
* <code>required bytes family = 1;</code>
*/
- public Builder setFamily(com.google.protobuf.ByteString value) {
+ public Builder setFamily(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -3955,12 +3955,12 @@ public final class WALProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -3978,22 +3978,22 @@ public final class WALProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<FamilyScope>
- PARSER = new com.google.protobuf.AbstractParser<FamilyScope>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<FamilyScope>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<FamilyScope>() {
public FamilyScope parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new FamilyScope(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<FamilyScope> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<FamilyScope> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<FamilyScope> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<FamilyScope> getParserForType() {
return PARSER;
}
@@ -4005,7 +4005,7 @@ public final class WALProtos {
public interface CompactionDescriptorOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.CompactionDescriptor)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <pre>
@@ -4022,7 +4022,7 @@ public final class WALProtos {
*
* <code>required bytes table_name = 1;</code>
*/
- com.google.protobuf.ByteString getTableName();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getTableName();
/**
* <code>required bytes encoded_region_name = 2;</code>
@@ -4031,7 +4031,7 @@ public final class WALProtos {
/**
* <code>required bytes encoded_region_name = 2;</code>
*/
- com.google.protobuf.ByteString getEncodedRegionName();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getEncodedRegionName();
/**
* <code>required bytes family_name = 3;</code>
@@ -4040,7 +4040,7 @@ public final class WALProtos {
/**
* <code>required bytes family_name = 3;</code>
*/
- com.google.protobuf.ByteString getFamilyName();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamilyName();
/**
* <pre>
@@ -4074,7 +4074,7 @@ public final class WALProtos {
*
* <code>
<TRUNCATED>
[16/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FilterProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FilterProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FilterProtos.java
index 7204813..b329e45 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FilterProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FilterProtos.java
@@ -6,17 +6,17 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class FilterProtos {
private FilterProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
public interface FilterOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.Filter)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required string name = 1;</code>
@@ -29,7 +29,7 @@ public final class FilterProtos {
/**
* <code>required string name = 1;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getNameBytes();
/**
@@ -39,37 +39,37 @@ public final class FilterProtos {
/**
* <code>optional bytes serialized_filter = 2;</code>
*/
- com.google.protobuf.ByteString getSerializedFilter();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getSerializedFilter();
}
/**
* Protobuf type {@code hbase.pb.Filter}
*/
public static final class Filter extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.Filter)
FilterOrBuilder {
// Use Filter.newBuilder() to construct.
- private Filter(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private Filter(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private Filter() {
name_ = "";
- serializedFilter_ = com.google.protobuf.ByteString.EMPTY;
+ serializedFilter_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private Filter(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -86,7 +86,7 @@ public final class FilterProtos {
break;
}
case 10: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000001;
name_ = bs;
break;
@@ -98,22 +98,22 @@ public final class FilterProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_Filter_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_Filter_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -137,8 +137,8 @@ public final class FilterProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
name_ = s;
@@ -149,22 +149,22 @@ public final class FilterProtos {
/**
* <code>required string name = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getNameBytes() {
java.lang.Object ref = name_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
name_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
public static final int SERIALIZED_FILTER_FIELD_NUMBER = 2;
- private com.google.protobuf.ByteString serializedFilter_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString serializedFilter_;
/**
* <code>optional bytes serialized_filter = 2;</code>
*/
@@ -174,7 +174,7 @@ public final class FilterProtos {
/**
* <code>optional bytes serialized_filter = 2;</code>
*/
- public com.google.protobuf.ByteString getSerializedFilter() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getSerializedFilter() {
return serializedFilter_;
}
@@ -192,10 +192,10 @@ public final class FilterProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 1, name_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, name_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
output.writeBytes(2, serializedFilter_);
@@ -209,10 +209,10 @@ public final class FilterProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, name_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, name_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(2, serializedFilter_);
}
size += unknownFields.getSerializedSize();
@@ -267,61 +267,61 @@ public final class FilterProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -339,7 +339,7 @@ public final class FilterProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -347,15 +347,15 @@ public final class FilterProtos {
* Protobuf type {@code hbase.pb.Filter}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.Filter)
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.FilterOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_Filter_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_Filter_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -368,12 +368,12 @@ public final class FilterProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -381,12 +381,12 @@ public final class FilterProtos {
super.clear();
name_ = "";
bitField0_ = (bitField0_ & ~0x00000001);
- serializedFilter_ = com.google.protobuf.ByteString.EMPTY;
+ serializedFilter_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000002);
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_Filter_descriptor;
}
@@ -424,29 +424,29 @@ public final class FilterProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter)other);
} else {
@@ -478,13 +478,13 @@ public final class FilterProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -509,8 +509,8 @@ public final class FilterProtos {
public java.lang.String getName() {
java.lang.Object ref = name_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
name_ = s;
@@ -523,17 +523,17 @@ public final class FilterProtos {
/**
* <code>required string name = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getNameBytes() {
java.lang.Object ref = name_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
name_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -562,7 +562,7 @@ public final class FilterProtos {
* <code>required string name = 1;</code>
*/
public Builder setNameBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -572,7 +572,7 @@ public final class FilterProtos {
return this;
}
- private com.google.protobuf.ByteString serializedFilter_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString serializedFilter_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>optional bytes serialized_filter = 2;</code>
*/
@@ -582,13 +582,13 @@ public final class FilterProtos {
/**
* <code>optional bytes serialized_filter = 2;</code>
*/
- public com.google.protobuf.ByteString getSerializedFilter() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getSerializedFilter() {
return serializedFilter_;
}
/**
* <code>optional bytes serialized_filter = 2;</code>
*/
- public Builder setSerializedFilter(com.google.protobuf.ByteString value) {
+ public Builder setSerializedFilter(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -607,12 +607,12 @@ public final class FilterProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -630,22 +630,22 @@ public final class FilterProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<Filter>
- PARSER = new com.google.protobuf.AbstractParser<Filter>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Filter>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Filter>() {
public Filter parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new Filter(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<Filter> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Filter> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<Filter> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Filter> getParserForType() {
return PARSER;
}
@@ -657,7 +657,7 @@ public final class FilterProtos {
public interface ColumnCountGetFilterOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.ColumnCountGetFilter)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required int32 limit = 1;</code>
@@ -672,11 +672,11 @@ public final class FilterProtos {
* Protobuf type {@code hbase.pb.ColumnCountGetFilter}
*/
public static final class ColumnCountGetFilter extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.ColumnCountGetFilter)
ColumnCountGetFilterOrBuilder {
// Use ColumnCountGetFilter.newBuilder() to construct.
- private ColumnCountGetFilter(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private ColumnCountGetFilter(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private ColumnCountGetFilter() {
@@ -684,18 +684,18 @@ public final class FilterProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private ColumnCountGetFilter(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -718,22 +718,22 @@ public final class FilterProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnCountGetFilter_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnCountGetFilter_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -770,7 +770,7 @@ public final class FilterProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeInt32(1, limit_);
@@ -784,7 +784,7 @@ public final class FilterProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeInt32Size(1, limit_);
}
size += unknownFields.getSerializedSize();
@@ -830,61 +830,61 @@ public final class FilterProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -902,7 +902,7 @@ public final class FilterProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -910,15 +910,15 @@ public final class FilterProtos {
* Protobuf type {@code hbase.pb.ColumnCountGetFilter}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.ColumnCountGetFilter)
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilterOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnCountGetFilter_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnCountGetFilter_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -931,12 +931,12 @@ public final class FilterProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -947,7 +947,7 @@ public final class FilterProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnCountGetFilter_descriptor;
}
@@ -981,29 +981,29 @@ public final class FilterProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter)other);
} else {
@@ -1030,13 +1030,13 @@ public final class FilterProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -1080,12 +1080,12 @@ public final class FilterProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -1103,22 +1103,22 @@ public final class FilterProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<ColumnCountGetFilter>
- PARSER = new com.google.protobuf.AbstractParser<ColumnCountGetFilter>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ColumnCountGetFilter>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<ColumnCountGetFilter>() {
public ColumnCountGetFilter parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new ColumnCountGetFilter(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<ColumnCountGetFilter> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ColumnCountGetFilter> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<ColumnCountGetFilter> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ColumnCountGetFilter> getParserForType() {
return PARSER;
}
@@ -1130,7 +1130,7 @@ public final class FilterProtos {
public interface ColumnPaginationFilterOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.ColumnPaginationFilter)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required int32 limit = 1;</code>
@@ -1157,38 +1157,38 @@ public final class FilterProtos {
/**
* <code>optional bytes column_offset = 3;</code>
*/
- com.google.protobuf.ByteString getColumnOffset();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getColumnOffset();
}
/**
* Protobuf type {@code hbase.pb.ColumnPaginationFilter}
*/
public static final class ColumnPaginationFilter extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.ColumnPaginationFilter)
ColumnPaginationFilterOrBuilder {
// Use ColumnPaginationFilter.newBuilder() to construct.
- private ColumnPaginationFilter(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private ColumnPaginationFilter(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private ColumnPaginationFilter() {
limit_ = 0;
offset_ = 0;
- columnOffset_ = com.google.protobuf.ByteString.EMPTY;
+ columnOffset_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private ColumnPaginationFilter(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -1221,22 +1221,22 @@ public final class FilterProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPaginationFilter_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPaginationFilter_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1275,7 +1275,7 @@ public final class FilterProtos {
}
public static final int COLUMN_OFFSET_FIELD_NUMBER = 3;
- private com.google.protobuf.ByteString columnOffset_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString columnOffset_;
/**
* <code>optional bytes column_offset = 3;</code>
*/
@@ -1285,7 +1285,7 @@ public final class FilterProtos {
/**
* <code>optional bytes column_offset = 3;</code>
*/
- public com.google.protobuf.ByteString getColumnOffset() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getColumnOffset() {
return columnOffset_;
}
@@ -1303,7 +1303,7 @@ public final class FilterProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeInt32(1, limit_);
@@ -1323,15 +1323,15 @@ public final class FilterProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeInt32Size(1, limit_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeInt32Size(2, offset_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(3, columnOffset_);
}
size += unknownFields.getSerializedSize();
@@ -1395,61 +1395,61 @@ public final class FilterProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -1467,7 +1467,7 @@ public final class FilterProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1475,15 +1475,15 @@ public final class FilterProtos {
* Protobuf type {@code hbase.pb.ColumnPaginationFilter}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.ColumnPaginationFilter)
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilterOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPaginationFilter_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPaginationFilter_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1496,12 +1496,12 @@ public final class FilterProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -1511,12 +1511,12 @@ public final class FilterProtos {
bitField0_ = (bitField0_ & ~0x00000001);
offset_ = 0;
bitField0_ = (bitField0_ & ~0x00000002);
- columnOffset_ = com.google.protobuf.ByteString.EMPTY;
+ columnOffset_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000004);
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPaginationFilter_descriptor;
}
@@ -1558,29 +1558,29 @@ public final class FilterProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter)other);
} else {
@@ -1613,13 +1613,13 @@ public final class FilterProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -1695,7 +1695,7 @@ public final class FilterProtos {
return this;
}
- private com.google.protobuf.ByteString columnOffset_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString columnOffset_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>optional bytes column_offset = 3;</code>
*/
@@ -1705,13 +1705,13 @@ public final class FilterProtos {
/**
* <code>optional bytes column_offset = 3;</code>
*/
- public com.google.protobuf.ByteString getColumnOffset() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getColumnOffset() {
return columnOffset_;
}
/**
* <code>optional bytes column_offset = 3;</code>
*/
- public Builder setColumnOffset(com.google.protobuf.ByteString value) {
+ public Builder setColumnOffset(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1730,12 +1730,12 @@ public final class FilterProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -1753,22 +1753,22 @@ public final class FilterProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<ColumnPaginationFilter>
- PARSER = new com.google.protobuf.AbstractParser<ColumnPaginationFilter>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ColumnPaginationFilter>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<ColumnPaginationFilter>() {
public ColumnPaginationFilter parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new ColumnPaginationFilter(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<ColumnPaginationFilter> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ColumnPaginationFilter> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<ColumnPaginationFilter> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ColumnPaginationFilter> getParserForType() {
return PARSER;
}
@@ -1780,7 +1780,7 @@ public final class FilterProtos {
public interface ColumnPrefixFilterOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.ColumnPrefixFilter)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required bytes prefix = 1;</code>
@@ -1789,36 +1789,36 @@ public final class FilterProtos {
/**
* <code>required bytes prefix = 1;</code>
*/
- com.google.protobuf.ByteString getPrefix();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getPrefix();
}
/**
* Protobuf type {@code hbase.pb.ColumnPrefixFilter}
*/
public static final class ColumnPrefixFilter extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.ColumnPrefixFilter)
ColumnPrefixFilterOrBuilder {
// Use ColumnPrefixFilter.newBuilder() to construct.
- private ColumnPrefixFilter(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private ColumnPrefixFilter(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private ColumnPrefixFilter() {
- prefix_ = com.google.protobuf.ByteString.EMPTY;
+ prefix_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private ColumnPrefixFilter(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -1841,22 +1841,22 @@ public final class FilterProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPrefixFilter_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPrefixFilter_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1865,7 +1865,7 @@ public final class FilterProtos {
private int bitField0_;
public static final int PREFIX_FIELD_NUMBER = 1;
- private com.google.protobuf.ByteString prefix_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString prefix_;
/**
* <code>required bytes prefix = 1;</code>
*/
@@ -1875,7 +1875,7 @@ public final class FilterProtos {
/**
* <code>required bytes prefix = 1;</code>
*/
- public com.google.protobuf.ByteString getPrefix() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getPrefix() {
return prefix_;
}
@@ -1893,7 +1893,7 @@ public final class FilterProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, prefix_);
@@ -1907,7 +1907,7 @@ public final class FilterProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(1, prefix_);
}
size += unknownFields.getSerializedSize();
@@ -1953,61 +1953,61 @@ public final class FilterProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -2025,7 +2025,7 @@ public final class FilterProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -2033,15 +2033,15 @@ public final class FilterProtos {
* Protobuf type {@code hbase.pb.ColumnPrefixFilter}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.ColumnPrefixFilter)
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilterOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPrefixFilter_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPrefixFilter_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2054,23 +2054,23 @@ public final class FilterProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
public Builder clear() {
super.clear();
- prefix_ = com.google.protobuf.ByteString.EMPTY;
+ prefix_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000001);
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPrefixFilter_descriptor;
}
@@ -2104,29 +2104,29 @@ public final class FilterProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter)other);
} else {
@@ -2153,13 +2153,13 @@ public final class FilterProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -2171,7 +2171,7 @@ public final class FilterProtos {
}
private int bitField0_;
- private com.google.protobuf.ByteString prefix_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString prefix_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>required bytes prefix = 1;</code>
*/
@@ -2181,13 +2181,13 @@ public final class FilterProtos {
/**
* <code>required bytes prefix = 1;</code>
*/
- public com.google.protobuf.ByteString getPrefix() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getPrefix() {
return prefix_;
}
/**
* <code>required bytes prefix = 1;</code>
*/
- public Builder setPrefix(com.google.protobuf.ByteString value) {
+ public Builder setPrefix(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -2206,12 +2206,12 @@ public final class FilterProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -2229,22 +2229,22 @@ public final class FilterProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<ColumnPrefixFilter>
- PARSER = new com.google.protobuf.AbstractParser<ColumnPrefixFilter>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ColumnPrefixFilter>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<ColumnPrefixFilter>() {
public ColumnPrefixFilter parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new ColumnPrefixFilter(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<ColumnPrefixFilter> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ColumnPrefixFilter> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<ColumnPrefixFilter> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ColumnPrefixFilter> getParserForType() {
return PARSER;
}
@@ -2256,7 +2256,7 @@ public final class FilterProtos {
public interface ColumnRangeFilterOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.ColumnRangeFilter)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>optional bytes min_column = 1;</code>
@@ -2265,7 +2265,7 @@ public final class FilterProtos {
/**
* <code>optional bytes min_column = 1;</code>
*/
- com.google.protobuf.ByteString getMinColumn();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getMinColumn();
/**
* <code>optional bool min_column_inclusive = 2;</code>
@@ -2283,7 +2283,7 @@ public final class FilterProtos {
/**
* <code>optional bytes max_column = 3;</code>
*/
- com.google.protobuf.ByteString getMaxColumn();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getMaxColumn();
/**
* <code>optional bool max_column_inclusive = 4;</code>
@@ -2298,33 +2298,33 @@ public final class FilterProtos {
* Protobuf type {@code hbase.pb.ColumnRangeFilter}
*/
public static final class ColumnRangeFilter extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.ColumnRangeFilter)
ColumnRangeFilterOrBuilder {
// Use ColumnRangeFilter.newBuilder() to construct.
- private ColumnRangeFilter(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private ColumnRangeFilter(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private ColumnRangeFilter() {
- minColumn_ = com.google.protobuf.ByteString.EMPTY;
+ minColumn_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
minColumnInclusive_ = false;
- maxColumn_ = com.google.protobuf.ByteString.EMPTY;
+ maxColumn_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
maxColumnInclusive_ = false;
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private ColumnRangeFilter(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -2362,22 +2362,22 @@ public final class FilterProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnRangeFilter_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnRangeFilter_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2386,7 +2386,7 @@ public final class FilterProtos {
private int bitField0_;
public static final int MIN_COLUMN_FIELD_NUMBER = 1;
- private com.google.protobuf.ByteString minColumn_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString minColumn_;
/**
* <code>optional bytes min_column = 1;</code>
*/
@@ -2396,7 +2396,7 @@ public final class FilterProtos {
/**
* <code>optional bytes min_column = 1;</code>
*/
- public com.google.protobuf.ByteString getMinColumn() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getMinColumn() {
return minColumn_;
}
@@ -2416,7 +2416,7 @@ public final class FilterProtos {
}
public static final int MAX_COLUMN_FIELD_NUMBER = 3;
- private com.google.protobuf.ByteString maxColumn_;
+ private org.apache.hadoop.hbase.shaded.
<TRUNCATED>
[12/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProcedureProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProcedureProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProcedureProtos.java
index bc8d40b..9942cd2 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProcedureProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProcedureProtos.java
@@ -6,19 +6,19 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class MasterProcedureProtos {
private MasterProcedureProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
/**
* Protobuf enum {@code hbase.pb.CreateTableState}
*/
public enum CreateTableState
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>CREATE_TABLE_PRE_OPERATION = 1;</code>
*/
@@ -95,27 +95,27 @@ public final class MasterProcedureProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<CreateTableState>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<CreateTableState>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
CreateTableState> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<CreateTableState>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<CreateTableState>() {
public CreateTableState findValueByNumber(int number) {
return CreateTableState.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.getDescriptor().getEnumTypes().get(0);
}
@@ -123,7 +123,7 @@ public final class MasterProcedureProtos {
private static final CreateTableState[] VALUES = values();
public static CreateTableState valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -144,7 +144,7 @@ public final class MasterProcedureProtos {
* Protobuf enum {@code hbase.pb.ModifyTableState}
*/
public enum ModifyTableState
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>MODIFY_TABLE_PREPARE = 1;</code>
*/
@@ -230,27 +230,27 @@ public final class MasterProcedureProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<ModifyTableState>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<ModifyTableState>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
ModifyTableState> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<ModifyTableState>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<ModifyTableState>() {
public ModifyTableState findValueByNumber(int number) {
return ModifyTableState.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.getDescriptor().getEnumTypes().get(1);
}
@@ -258,7 +258,7 @@ public final class MasterProcedureProtos {
private static final ModifyTableState[] VALUES = values();
public static ModifyTableState valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -279,7 +279,7 @@ public final class MasterProcedureProtos {
* Protobuf enum {@code hbase.pb.TruncateTableState}
*/
public enum TruncateTableState
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>TRUNCATE_TABLE_PRE_OPERATION = 1;</code>
*/
@@ -365,27 +365,27 @@ public final class MasterProcedureProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<TruncateTableState>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<TruncateTableState>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
TruncateTableState> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<TruncateTableState>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<TruncateTableState>() {
public TruncateTableState findValueByNumber(int number) {
return TruncateTableState.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.getDescriptor().getEnumTypes().get(2);
}
@@ -393,7 +393,7 @@ public final class MasterProcedureProtos {
private static final TruncateTableState[] VALUES = values();
public static TruncateTableState valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -414,7 +414,7 @@ public final class MasterProcedureProtos {
* Protobuf enum {@code hbase.pb.DeleteTableState}
*/
public enum DeleteTableState
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>DELETE_TABLE_PRE_OPERATION = 1;</code>
*/
@@ -491,27 +491,27 @@ public final class MasterProcedureProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<DeleteTableState>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<DeleteTableState>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
DeleteTableState> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<DeleteTableState>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<DeleteTableState>() {
public DeleteTableState findValueByNumber(int number) {
return DeleteTableState.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.getDescriptor().getEnumTypes().get(3);
}
@@ -519,7 +519,7 @@ public final class MasterProcedureProtos {
private static final DeleteTableState[] VALUES = values();
public static DeleteTableState valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -540,7 +540,7 @@ public final class MasterProcedureProtos {
* Protobuf enum {@code hbase.pb.CreateNamespaceState}
*/
public enum CreateNamespaceState
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>CREATE_NAMESPACE_PREPARE = 1;</code>
*/
@@ -608,27 +608,27 @@ public final class MasterProcedureProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<CreateNamespaceState>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<CreateNamespaceState>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
CreateNamespaceState> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<CreateNamespaceState>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<CreateNamespaceState>() {
public CreateNamespaceState findValueByNumber(int number) {
return CreateNamespaceState.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.getDescriptor().getEnumTypes().get(4);
}
@@ -636,7 +636,7 @@ public final class MasterProcedureProtos {
private static final CreateNamespaceState[] VALUES = values();
public static CreateNamespaceState valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -657,7 +657,7 @@ public final class MasterProcedureProtos {
* Protobuf enum {@code hbase.pb.ModifyNamespaceState}
*/
public enum ModifyNamespaceState
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>MODIFY_NAMESPACE_PREPARE = 1;</code>
*/
@@ -707,27 +707,27 @@ public final class MasterProcedureProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<ModifyNamespaceState>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<ModifyNamespaceState>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
ModifyNamespaceState> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<ModifyNamespaceState>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<ModifyNamespaceState>() {
public ModifyNamespaceState findValueByNumber(int number) {
return ModifyNamespaceState.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.getDescriptor().getEnumTypes().get(5);
}
@@ -735,7 +735,7 @@ public final class MasterProcedureProtos {
private static final ModifyNamespaceState[] VALUES = values();
public static ModifyNamespaceState valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -756,7 +756,7 @@ public final class MasterProcedureProtos {
* Protobuf enum {@code hbase.pb.DeleteNamespaceState}
*/
public enum DeleteNamespaceState
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>DELETE_NAMESPACE_PREPARE = 1;</code>
*/
@@ -824,27 +824,27 @@ public final class MasterProcedureProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<DeleteNamespaceState>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<DeleteNamespaceState>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
DeleteNamespaceState> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<DeleteNamespaceState>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<DeleteNamespaceState>() {
public DeleteNamespaceState findValueByNumber(int number) {
return DeleteNamespaceState.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.getDescriptor().getEnumTypes().get(6);
}
@@ -852,7 +852,7 @@ public final class MasterProcedureProtos {
private static final DeleteNamespaceState[] VALUES = values();
public static DeleteNamespaceState valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -873,7 +873,7 @@ public final class MasterProcedureProtos {
* Protobuf enum {@code hbase.pb.AddColumnFamilyState}
*/
public enum AddColumnFamilyState
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>ADD_COLUMN_FAMILY_PREPARE = 1;</code>
*/
@@ -941,27 +941,27 @@ public final class MasterProcedureProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<AddColumnFamilyState>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<AddColumnFamilyState>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
AddColumnFamilyState> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<AddColumnFamilyState>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<AddColumnFamilyState>() {
public AddColumnFamilyState findValueByNumber(int number) {
return AddColumnFamilyState.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.getDescriptor().getEnumTypes().get(7);
}
@@ -969,7 +969,7 @@ public final class MasterProcedureProtos {
private static final AddColumnFamilyState[] VALUES = values();
public static AddColumnFamilyState valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -990,7 +990,7 @@ public final class MasterProcedureProtos {
* Protobuf enum {@code hbase.pb.ModifyColumnFamilyState}
*/
public enum ModifyColumnFamilyState
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>MODIFY_COLUMN_FAMILY_PREPARE = 1;</code>
*/
@@ -1058,27 +1058,27 @@ public final class MasterProcedureProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<ModifyColumnFamilyState>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<ModifyColumnFamilyState>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
ModifyColumnFamilyState> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<ModifyColumnFamilyState>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<ModifyColumnFamilyState>() {
public ModifyColumnFamilyState findValueByNumber(int number) {
return ModifyColumnFamilyState.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.getDescriptor().getEnumTypes().get(8);
}
@@ -1086,7 +1086,7 @@ public final class MasterProcedureProtos {
private static final ModifyColumnFamilyState[] VALUES = values();
public static ModifyColumnFamilyState valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -1107,7 +1107,7 @@ public final class MasterProcedureProtos {
* Protobuf enum {@code hbase.pb.DeleteColumnFamilyState}
*/
public enum DeleteColumnFamilyState
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>DELETE_COLUMN_FAMILY_PREPARE = 1;</code>
*/
@@ -1184,27 +1184,27 @@ public final class MasterProcedureProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<DeleteColumnFamilyState>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<DeleteColumnFamilyState>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
DeleteColumnFamilyState> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<DeleteColumnFamilyState>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<DeleteColumnFamilyState>() {
public DeleteColumnFamilyState findValueByNumber(int number) {
return DeleteColumnFamilyState.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.getDescriptor().getEnumTypes().get(9);
}
@@ -1212,7 +1212,7 @@ public final class MasterProcedureProtos {
private static final DeleteColumnFamilyState[] VALUES = values();
public static DeleteColumnFamilyState valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -1233,7 +1233,7 @@ public final class MasterProcedureProtos {
* Protobuf enum {@code hbase.pb.EnableTableState}
*/
public enum EnableTableState
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>ENABLE_TABLE_PREPARE = 1;</code>
*/
@@ -1310,27 +1310,27 @@ public final class MasterProcedureProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<EnableTableState>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<EnableTableState>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
EnableTableState> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<EnableTableState>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<EnableTableState>() {
public EnableTableState findValueByNumber(int number) {
return EnableTableState.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.getDescriptor().getEnumTypes().get(10);
}
@@ -1338,7 +1338,7 @@ public final class MasterProcedureProtos {
private static final EnableTableState[] VALUES = values();
public static EnableTableState valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -1359,7 +1359,7 @@ public final class MasterProcedureProtos {
* Protobuf enum {@code hbase.pb.DisableTableState}
*/
public enum DisableTableState
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>DISABLE_TABLE_PREPARE = 1;</code>
*/
@@ -1436,27 +1436,27 @@ public final class MasterProcedureProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<DisableTableState>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<DisableTableState>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
DisableTableState> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<DisableTableState>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<DisableTableState>() {
public DisableTableState findValueByNumber(int number) {
return DisableTableState.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.getDescriptor().getEnumTypes().get(11);
}
@@ -1464,7 +1464,7 @@ public final class MasterProcedureProtos {
private static final DisableTableState[] VALUES = values();
public static DisableTableState valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -1485,7 +1485,7 @@ public final class MasterProcedureProtos {
* Protobuf enum {@code hbase.pb.CloneSnapshotState}
*/
public enum CloneSnapshotState
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>CLONE_SNAPSHOT_PRE_OPERATION = 1;</code>
*/
@@ -1562,27 +1562,27 @@ public final class MasterProcedureProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<CloneSnapshotState>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<CloneSnapshotState>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
CloneSnapshotState> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<CloneSnapshotState>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<CloneSnapshotState>() {
public CloneSnapshotState findValueByNumber(int number) {
return CloneSnapshotState.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.getDescriptor().getEnumTypes().get(12);
}
@@ -1590,7 +1590,7 @@ public final class MasterProcedureProtos {
private static final CloneSnapshotState[] VALUES = values();
public static CloneSnapshotState valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -1611,7 +1611,7 @@ public final class MasterProcedureProtos {
* Protobuf enum {@code hbase.pb.RestoreSnapshotState}
*/
public enum RestoreSnapshotState
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>RESTORE_SNAPSHOT_PRE_OPERATION = 1;</code>
*/
@@ -1670,27 +1670,27 @@ public final class MasterProcedureProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<RestoreSnapshotState>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<RestoreSnapshotState>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
RestoreSnapshotState> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<RestoreSnapshotState>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<RestoreSnapshotState>() {
public RestoreSnapshotState findValueByNumber(int number) {
return RestoreSnapshotState.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.getDescriptor().getEnumTypes().get(13);
}
@@ -1698,7 +1698,7 @@ public final class MasterProcedureProtos {
private static final RestoreSnapshotState[] VALUES = values();
public static RestoreSnapshotState valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -1719,7 +1719,7 @@ public final class MasterProcedureProtos {
* Protobuf enum {@code hbase.pb.DispatchMergingRegionsState}
*/
public enum DispatchMergingRegionsState
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>DISPATCH_MERGING_REGIONS_PREPARE = 1;</code>
*/
@@ -1787,27 +1787,27 @@ public final class MasterProcedureProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<DispatchMergingRegionsState>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<DispatchMergingRegionsState>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
DispatchMergingRegionsState> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<DispatchMergingRegionsState>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<DispatchMergingRegionsState>() {
public DispatchMergingRegionsState findValueByNumber(int number) {
return DispatchMergingRegionsState.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.getDescriptor().getEnumTypes().get(14);
}
@@ -1815,7 +1815,7 @@ public final class MasterProcedureProtos {
private static final DispatchMergingRegionsState[] VALUES = values();
public static DispatchMergingRegionsState valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -1836,7 +1836,7 @@ public final class MasterProcedureProtos {
* Protobuf enum {@code hbase.pb.ServerCrashState}
*/
public enum ServerCrashState
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>SERVER_CRASH_START = 1;</code>
*/
@@ -1948,27 +1948,27 @@ public final class MasterProcedureProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<ServerCrashState>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<ServerCrashState>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
ServerCrashState> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<ServerCrashState>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<ServerCrashState>() {
public ServerCrashState findValueByNumber(int number) {
return ServerCrashState.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.getDescriptor().getEnumTypes().get(15);
}
@@ -1976,7 +1976,7 @@ public final class MasterProcedureProtos {
private static final ServerCrashState[] VALUES = values();
public static ServerCrashState valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -1995,7 +1995,7 @@ public final class MasterProcedureProtos {
public interface CreateTableStateDataOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.CreateTableStateData)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required .hbase.pb.UserInformation user_info = 1;</code>
@@ -2051,11 +2051,11 @@ public final class MasterProcedureProtos {
* Protobuf type {@code hbase.pb.CreateTableStateData}
*/
public static final class CreateTableStateData extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.CreateTableStateData)
CreateTableStateDataOrBuilder {
// Use CreateTableStateData.newBuilder() to construct.
- private CreateTableStateData(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private CreateTableStateData(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private CreateTableStateData() {
@@ -2063,18 +2063,18 @@ public final class MasterProcedureProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private CreateTableStateData(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -2127,10 +2127,10 @@ public final class MasterProcedureProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
@@ -2140,12 +2140,12 @@ public final class MasterProcedureProtos {
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.internal_static_hbase_pb_CreateTableStateData_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.internal_static_hbase_pb_CreateTableStateData_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2262,7 +2262,7 @@ public final class MasterProcedureProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeMessage(1, getUserInfo());
@@ -2282,15 +2282,15 @@ public final class MasterProcedureProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(1, getUserInfo());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(2, getTableSchema());
}
for (int i = 0; i < regionInfo_.size(); i++) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(3, regionInfo_.get(i));
}
size += unknownFields.getSerializedSize();
@@ -2351,61 +2351,61 @@ public final class MasterProcedureProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CreateTableStateData parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CreateTableStateData parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CreateTableStateData parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CreateTableStateData parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CreateTableStateData parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CreateTableStateData parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CreateTableStateData parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CreateTableStateData parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CreateTableStateData parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CreateTableStateData parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -2423,7 +2423,7 @@ public final class MasterProcedureProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -2431,15 +2431,15 @@ public final class MasterProcedureProtos {
* Protobuf type {@code hbase.pb.CreateTableStateData}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.CreateTableStateData)
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CreateTableStateDataOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.internal_static_hbase_pb_CreateTableStateData_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.internal_static_hbase_pb_CreateTableStateData_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2452,12 +2452,12 @@ public final class MasterProcedureProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getUserInfoFieldBuilder();
getTableSchemaFieldBuilder();
@@ -2487,7 +2487,7 @@ public final class MasterProcedureProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.internal_static_hbase_pb_CreateTableStateData_descriptor;
}
@@ -2542,29 +2542,29 @@ public final class MasterProcedureProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CreateTableStateData) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CreateTableStateData)other);
} else {
@@ -2600,7 +2600,7 @@ public final class MasterProcedureProtos {
regionInfo_ = other.regionInfo_;
bitField0_ = (bitField0_ & ~0x00000004);
regionInfoBuilder_ =
- com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getRegionInfoFieldBuilder() : null;
} else {
regionInfoBuilder_.addAllMessages(other.regionInfo_);
@@ -2634,13 +2634,13 @@ public final class MasterProcedureProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CreateTableStateData parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CreateTableStateData) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -2653,7 +2653,7 @@ public final class MasterProcedureProtos {
private int bitField0_;
private org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation userInfo_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationOrBuilder> userInfoBuilder_;
/**
* <code>required .hbase.pb.UserInformation user_info = 1;</code>
@@ -2756,11 +2756,11 @@ public final class MasterProcedureProtos {
/**
* <code>required .hbase.pb.UserInformation user_info = 1;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationOrBuilder>
getUserInfoFieldBuilder() {
if (userInfoBuilder_ == null) {
- userInfoBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ userInfoBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationOrBuilder>(
getUserInfo(),
getParentForChildren(),
@@ -2771,7 +2771,7 @@ public final class MasterProcedureProtos {
}
private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema tableSchema_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> tableSchemaBuilder_;
/**
* <code>required .hbase.pb.TableSchema table_schema = 2;</code>
@@ -2874,11 +2874,11 @@ public final class MasterProcedureProtos {
/**
* <code>required .hbase.pb.TableSchema table_schema = 2;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder>
getTableSchemaFieldBuilder() {
if (tableSchemaBuilder_ == null) {
- tableSchemaBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ tableSchemaBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder>(
getTableSchema(),
getParentForChildren(),
@@ -2897,7 +2897,7 @@ public final class MasterProcedureProtos {
}
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder> regionInfoBuilder_;
/**
@@ -3029,7 +3029,7 @@ public final class MasterProcedureProtos {
java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo> values) {
if (regionInfoBuilder_ == null) {
ensureRegionInfoIsMutable();
- com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
values, regionInfo_);
onChanged();
} else {
@@ -3113,11 +3113,11 @@ public final class MasterProcedureProtos {
getRegionInfoBuilderList() {
return getRegionInfoFieldBuilder().getBuilderList();
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder>
getRegionInfoFieldBuilder() {
if (regionInfoBuilder_ == null) {
- regionInfoBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3<
+ regionInfoBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder>(
regionInfo_,
((bitField0_ & 0x00000004) == 0x00000004),
@@ -3128,12 +3128,12 @@ public final class MasterProcedureProtos {
return regionInfoBuilder_;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -3151,22 +3151,22 @@ public final class MasterProcedureProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<CreateTableStateData>
- PARSER = new com.google.protobuf.AbstractParser<CreateTableStateData>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<CreateTableStateData>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<CreateTableStateData>() {
public CreateTableStateData parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new CreateTableStateData(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<CreateTableStateData> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<CreateTableStateData> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<CreateTableStateData> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<CreateTableStateData> getParserForType() {
return PARSER;
}
@@ -3178,7 +3178,7 @@ public final class MasterProcedureProtos {
public interface ModifyTableStateDataOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.ModifyTableStateData)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required .hbase.pb.UserInformation user_info = 1;</code>
@@ -3232,11 +3232,11 @@ public final class MasterProcedureProtos {
* Protobuf type {@code hbase.pb.ModifyTableStateData}
*/
public static final class ModifyTableStateData extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.ModifyTableStateData)
ModifyTableStateDataOrBuilder {
// Use ModifyTableStateData.newBuilder() to construct.
- private ModifyTableStateData(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private ModifyTableStateData(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private ModifyTableStateData() {
@@ -3244,18 +3244,18 @@ public final class MasterProcedureProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private ModifyTableStateData(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -3317,22 +3317,22 @@ public final class MasterProcedureProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.internal_static_hbase_pb_ModifyTableStateData_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.internal_static_hbase_pb_ModifyTableStateData_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -3454,7 +3454,7 @@ public final class MasterProcedureProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeMessage(1, getUserInfo());
@@ -3477,19 +3477,19 @@ public final class MasterProcedureProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(1, getUserInfo());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(2, getUnmodifiedTableSchema());
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(3, getModifiedTableSchema());
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBoolSize(4, deleteColumnFamilyInModify_);
}
size += unknownFields.getSerializedSize();
@@ -3554,7 +3554,7 @@ public final class MasterProcedureProtos {
}
if (hasDeleteColumnFamilyInModify()) {
hash = (37 * hash) + DELETE_COLUMN_FAMILY_IN_MODIFY_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean(
getDeleteColumnFamilyInModify());
}
hash = (29 * hash) + unknownFields.hashCode();
@@ -3563,61 +3563,61 @@ public final class MasterProcedureProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStateData parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStateData parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStateData parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStateData parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStateData parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStateData parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStateData parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStateData parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStateData parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStateData parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -3635,7 +3635,7 @@ public final class MasterProcedureProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -3643,15 +3643,15 @@ public final class MasterProcedureProtos {
* Protobuf type {@code hbase.pb.ModifyTableStateData}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.ModifyTableStateData)
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStateDataOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.internal_static_hbase_pb_ModifyTableStateData_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.internal_static_hbase_pb_ModifyTableStateData_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -3664,12 +3664,12 @@ public final class MasterProcedureProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getUserInfoFieldBuilder();
getUnmodifiedTableSchemaFieldBuilder();
@@ -3701,7 +3701,7 @@ public final class MasterProcedureProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.internal_static_hbase_pb_ModifyTableStateData_descriptor;
}
@@ -3759,29 +3759,29 @@ public final class MasterProcedureProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStateData) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStateData)other);
} else {
@@ -3834,13 +3834,13 @@ public final class MasterProcedureProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStateData parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStateData) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -3853,7 +3853,7 @@ public final class MasterProcedureProtos {
private int bitField0_;
private org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation userInfo_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationOrBuilder> userInfoBuilder_;
/**
* <code>required .hbase.pb.UserInformation user_info = 1;</code>
@@ -3956,11 +3956,11 @@ public final class MasterProcedureProtos {
/**
* <code>required .hbase.pb.UserInformation user_info = 1;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationOrBuilder>
getUserInfoFieldBuilder() {
if (userInfoBuilder_ == null) {
- userInfoBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ userInfoBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationOrBuilder>(
getUserInfo(),
getParentForChildren(),
@@ -3971,7 +3971,7 @@ public final class MasterProcedureProtos {
}
private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema unmodifiedTableSchema_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> unmodifiedTableSchemaBuilder_;
/**
* <code>optional .hbase.pb.TableSchema unmodified_table_schema = 2;</code>
@@ -4074,11 +4074,11 @@ public final class MasterProcedureProtos {
/**
* <code>optional .hbase.pb.TableSchema unmodified_table_schema = 2;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache
<TRUNCATED>
[13/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MapReduceProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MapReduceProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MapReduceProtos.java
index b5abf42..09605a2 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MapReduceProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MapReduceProtos.java
@@ -6,17 +6,17 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class MapReduceProtos {
private MapReduceProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
public interface ScanMetricsOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.ScanMetrics)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>repeated .hbase.pb.NameInt64Pair metrics = 1;</code>
@@ -46,11 +46,11 @@ public final class MapReduceProtos {
* Protobuf type {@code hbase.pb.ScanMetrics}
*/
public static final class ScanMetrics extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.ScanMetrics)
ScanMetricsOrBuilder {
// Use ScanMetrics.newBuilder() to construct.
- private ScanMetrics(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private ScanMetrics(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private ScanMetrics() {
@@ -58,18 +58,18 @@ public final class MapReduceProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private ScanMetrics(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -96,10 +96,10 @@ public final class MapReduceProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
@@ -109,12 +109,12 @@ public final class MapReduceProtos {
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.internal_static_hbase_pb_ScanMetrics_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.internal_static_hbase_pb_ScanMetrics_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -166,7 +166,7 @@ public final class MapReduceProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
for (int i = 0; i < metrics_.size(); i++) {
output.writeMessage(1, metrics_.get(i));
@@ -180,7 +180,7 @@ public final class MapReduceProtos {
size = 0;
for (int i = 0; i < metrics_.size(); i++) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(1, metrics_.get(i));
}
size += unknownFields.getSerializedSize();
@@ -223,61 +223,61 @@ public final class MapReduceProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.ScanMetrics parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.ScanMetrics parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.ScanMetrics parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.ScanMetrics parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.ScanMetrics parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.ScanMetrics parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.ScanMetrics parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.ScanMetrics parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.ScanMetrics parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.ScanMetrics parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -295,7 +295,7 @@ public final class MapReduceProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -303,15 +303,15 @@ public final class MapReduceProtos {
* Protobuf type {@code hbase.pb.ScanMetrics}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.ScanMetrics)
org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.ScanMetricsOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.internal_static_hbase_pb_ScanMetrics_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.internal_static_hbase_pb_ScanMetrics_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -324,12 +324,12 @@ public final class MapReduceProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getMetricsFieldBuilder();
}
@@ -345,7 +345,7 @@ public final class MapReduceProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.internal_static_hbase_pb_ScanMetrics_descriptor;
}
@@ -382,29 +382,29 @@ public final class MapReduceProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.ScanMetrics) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.ScanMetrics)other);
} else {
@@ -434,7 +434,7 @@ public final class MapReduceProtos {
metrics_ = other.metrics_;
bitField0_ = (bitField0_ & ~0x00000001);
metricsBuilder_ =
- com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getMetricsFieldBuilder() : null;
} else {
metricsBuilder_.addAllMessages(other.metrics_);
@@ -451,13 +451,13 @@ public final class MapReduceProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.ScanMetrics parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.ScanMetrics) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -478,7 +478,7 @@ public final class MapReduceProtos {
}
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameInt64Pair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameInt64Pair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameInt64PairOrBuilder> metricsBuilder_;
/**
@@ -610,7 +610,7 @@ public final class MapReduceProtos {
java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameInt64Pair> values) {
if (metricsBuilder_ == null) {
ensureMetricsIsMutable();
- com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
values, metrics_);
onChanged();
} else {
@@ -694,11 +694,11 @@ public final class MapReduceProtos {
getMetricsBuilderList() {
return getMetricsFieldBuilder().getBuilderList();
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameInt64Pair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameInt64Pair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameInt64PairOrBuilder>
getMetricsFieldBuilder() {
if (metricsBuilder_ == null) {
- metricsBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3<
+ metricsBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameInt64Pair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameInt64Pair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameInt64PairOrBuilder>(
metrics_,
((bitField0_ & 0x00000001) == 0x00000001),
@@ -709,12 +709,12 @@ public final class MapReduceProtos {
return metricsBuilder_;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -732,22 +732,22 @@ public final class MapReduceProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<ScanMetrics>
- PARSER = new com.google.protobuf.AbstractParser<ScanMetrics>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ScanMetrics>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<ScanMetrics>() {
public ScanMetrics parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new ScanMetrics(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<ScanMetrics> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ScanMetrics> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<ScanMetrics> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ScanMetrics> getParserForType() {
return PARSER;
}
@@ -759,7 +759,7 @@ public final class MapReduceProtos {
public interface TableSnapshotRegionSplitOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.TableSnapshotRegionSplit)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>repeated string locations = 2;</code>
@@ -777,7 +777,7 @@ public final class MapReduceProtos {
/**
* <code>repeated string locations = 2;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getLocationsBytes(int index);
/**
@@ -810,30 +810,30 @@ public final class MapReduceProtos {
* Protobuf type {@code hbase.pb.TableSnapshotRegionSplit}
*/
public static final class TableSnapshotRegionSplit extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.TableSnapshotRegionSplit)
TableSnapshotRegionSplitOrBuilder {
// Use TableSnapshotRegionSplit.newBuilder() to construct.
- private TableSnapshotRegionSplit(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private TableSnapshotRegionSplit(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private TableSnapshotRegionSplit() {
- locations_ = com.google.protobuf.LazyStringArrayList.EMPTY;
+ locations_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private TableSnapshotRegionSplit(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -850,9 +850,9 @@ public final class MapReduceProtos {
break;
}
case 18: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
- locations_ = new com.google.protobuf.LazyStringArrayList();
+ locations_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList();
mutable_bitField0_ |= 0x00000001;
}
locations_.add(bs);
@@ -886,10 +886,10 @@ public final class MapReduceProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
@@ -899,12 +899,12 @@ public final class MapReduceProtos {
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.internal_static_hbase_pb_TableSnapshotRegionSplit_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.internal_static_hbase_pb_TableSnapshotRegionSplit_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -913,11 +913,11 @@ public final class MapReduceProtos {
private int bitField0_;
public static final int LOCATIONS_FIELD_NUMBER = 2;
- private com.google.protobuf.LazyStringList locations_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringList locations_;
/**
* <code>repeated string locations = 2;</code>
*/
- public com.google.protobuf.ProtocolStringList
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolStringList
getLocationsList() {
return locations_;
}
@@ -936,7 +936,7 @@ public final class MapReduceProtos {
/**
* <code>repeated string locations = 2;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getLocationsBytes(int index) {
return locations_.getByteString(index);
}
@@ -1005,10 +1005,10 @@ public final class MapReduceProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
for (int i = 0; i < locations_.size(); i++) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 2, locations_.getRaw(i));
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 2, locations_.getRaw(i));
}
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeMessage(3, getTable());
@@ -1033,11 +1033,11 @@ public final class MapReduceProtos {
size += 1 * getLocationsList().size();
}
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(3, getTable());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(4, getRegion());
}
size += unknownFields.getSerializedSize();
@@ -1098,61 +1098,61 @@ public final class MapReduceProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.TableSnapshotRegionSplit parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.TableSnapshotRegionSplit parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.TableSnapshotRegionSplit parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.TableSnapshotRegionSplit parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.TableSnapshotRegionSplit parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.TableSnapshotRegionSplit parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.TableSnapshotRegionSplit parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.TableSnapshotRegionSplit parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.TableSnapshotRegionSplit parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.TableSnapshotRegionSplit parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -1170,7 +1170,7 @@ public final class MapReduceProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1178,15 +1178,15 @@ public final class MapReduceProtos {
* Protobuf type {@code hbase.pb.TableSnapshotRegionSplit}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.TableSnapshotRegionSplit)
org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.TableSnapshotRegionSplitOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.internal_static_hbase_pb_TableSnapshotRegionSplit_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.internal_static_hbase_pb_TableSnapshotRegionSplit_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1199,12 +1199,12 @@ public final class MapReduceProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getTableFieldBuilder();
getRegionFieldBuilder();
@@ -1212,7 +1212,7 @@ public final class MapReduceProtos {
}
public Builder clear() {
super.clear();
- locations_ = com.google.protobuf.LazyStringArrayList.EMPTY;
+ locations_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
bitField0_ = (bitField0_ & ~0x00000001);
if (tableBuilder_ == null) {
table_ = null;
@@ -1229,7 +1229,7 @@ public final class MapReduceProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.internal_static_hbase_pb_TableSnapshotRegionSplit_descriptor;
}
@@ -1280,29 +1280,29 @@ public final class MapReduceProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.TableSnapshotRegionSplit) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.TableSnapshotRegionSplit)other);
} else {
@@ -1349,13 +1349,13 @@ public final class MapReduceProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.TableSnapshotRegionSplit parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MapReduceProtos.TableSnapshotRegionSplit) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -1367,17 +1367,17 @@ public final class MapReduceProtos {
}
private int bitField0_;
- private com.google.protobuf.LazyStringList locations_ = com.google.protobuf.LazyStringArrayList.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringList locations_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
private void ensureLocationsIsMutable() {
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
- locations_ = new com.google.protobuf.LazyStringArrayList(locations_);
+ locations_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList(locations_);
bitField0_ |= 0x00000001;
}
}
/**
* <code>repeated string locations = 2;</code>
*/
- public com.google.protobuf.ProtocolStringList
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolStringList
getLocationsList() {
return locations_.getUnmodifiableView();
}
@@ -1396,7 +1396,7 @@ public final class MapReduceProtos {
/**
* <code>repeated string locations = 2;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getLocationsBytes(int index) {
return locations_.getByteString(index);
}
@@ -1432,7 +1432,7 @@ public final class MapReduceProtos {
public Builder addAllLocations(
java.lang.Iterable<java.lang.String> values) {
ensureLocationsIsMutable();
- com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
values, locations_);
onChanged();
return this;
@@ -1441,7 +1441,7 @@ public final class MapReduceProtos {
* <code>repeated string locations = 2;</code>
*/
public Builder clearLocations() {
- locations_ = com.google.protobuf.LazyStringArrayList.EMPTY;
+ locations_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
return this;
@@ -1450,7 +1450,7 @@ public final class MapReduceProtos {
* <code>repeated string locations = 2;</code>
*/
public Builder addLocationsBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1461,7 +1461,7 @@ public final class MapReduceProtos {
}
private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema table_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> tableBuilder_;
/**
* <code>optional .hbase.pb.TableSchema table = 3;</code>
@@ -1564,11 +1564,11 @@ public final class MapReduceProtos {
/**
* <code>optional .hbase.pb.TableSchema table = 3;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder>
getTableFieldBuilder() {
if (tableBuilder_ == null) {
- tableBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ tableBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder>(
getTable(),
getParentForChildren(),
@@ -1579,7 +1579,7 @@ public final class MapReduceProtos {
}
private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo region_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder> regionBuilder_;
/**
* <code>optional .hbase.pb.RegionInfo region = 4;</code>
@@ -1682,11 +1682,11 @@ public final class MapReduceProtos {
/**
* <code>optional .hbase.pb.RegionInfo region = 4;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder>
getRegionFieldBuilder() {
if (regionBuilder_ == null) {
- regionBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ regionBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder>(
getRegion(),
getParentForChildren(),
@@ -1696,12 +1696,12 @@ public final class MapReduceProtos {
return regionBuilder_;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -1719,22 +1719,22 @@ public final class MapReduceProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<TableSnapshotRegionSplit>
- PARSER = new com.google.protobuf.AbstractParser<TableSnapshotRegionSplit>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<TableSnapshotRegionSplit>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<TableSnapshotRegionSplit>() {
public TableSnapshotRegionSplit parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new TableSnapshotRegionSplit(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<TableSnapshotRegionSplit> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<TableSnapshotRegionSplit> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<TableSnapshotRegionSplit> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<TableSnapshotRegionSplit> getParserForType() {
return PARSER;
}
@@ -1744,22 +1744,22 @@ public final class MapReduceProtos {
}
- private static final com.google.protobuf.Descriptors.Descriptor
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
internal_static_hbase_pb_ScanMetrics_descriptor;
private static final
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hbase_pb_ScanMetrics_fieldAccessorTable;
- private static final com.google.protobuf.Descriptors.Descriptor
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
internal_static_hbase_pb_TableSnapshotRegionSplit_descriptor;
private static final
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hbase_pb_TableSnapshotRegionSplit_fieldAccessorTable;
- public static com.google.protobuf.Descriptors.FileDescriptor
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
getDescriptor() {
return descriptor;
}
- private static com.google.protobuf.Descriptors.FileDescriptor
+ private static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
descriptor;
static {
java.lang.String[] descriptorData = {
@@ -1772,29 +1772,29 @@ public final class MapReduceProtos {
"p.hbase.shaded.protobuf.generatedB\017MapRe" +
"duceProtosH\001\240\001\001"
};
- com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
- new com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
- public com.google.protobuf.ExtensionRegistry assignDescriptors(
- com.google.protobuf.Descriptors.FileDescriptor root) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
descriptor = root;
return null;
}
};
- com.google.protobuf.Descriptors.FileDescriptor
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
.internalBuildGeneratedFileFrom(descriptorData,
- new com.google.protobuf.Descriptors.FileDescriptor[] {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.getDescriptor(),
}, assigner);
internal_static_hbase_pb_ScanMetrics_descriptor =
getDescriptor().getMessageTypes().get(0);
internal_static_hbase_pb_ScanMetrics_fieldAccessorTable = new
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hbase_pb_ScanMetrics_descriptor,
new java.lang.String[] { "Metrics", });
internal_static_hbase_pb_TableSnapshotRegionSplit_descriptor =
getDescriptor().getMessageTypes().get(1);
internal_static_hbase_pb_TableSnapshotRegionSplit_fieldAccessorTable = new
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hbase_pb_TableSnapshotRegionSplit_descriptor,
new java.lang.String[] { "Locations", "Table", "Region", });
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.getDescriptor();
[27/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestProtos.java
index 0fdd1fb..36a15c9 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestProtos.java
@@ -6,44 +6,44 @@ package org.apache.hadoop.hbase.shaded.ipc.protobuf.generated;
public final class TestProtos {
private TestProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
public interface EmptyRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:EmptyRequestProto)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
}
/**
* Protobuf type {@code EmptyRequestProto}
*/
public static final class EmptyRequestProto extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:EmptyRequestProto)
EmptyRequestProtoOrBuilder {
// Use EmptyRequestProto.newBuilder() to construct.
- private EmptyRequestProto(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private EmptyRequestProto(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private EmptyRequestProto() {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private EmptyRequestProto(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -61,22 +61,22 @@ public final class TestProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EmptyRequestProto_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EmptyRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -93,7 +93,7 @@ public final class TestProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
unknownFields.writeTo(output);
}
@@ -137,61 +137,61 @@ public final class TestProtos {
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -209,7 +209,7 @@ public final class TestProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -217,15 +217,15 @@ public final class TestProtos {
* Protobuf type {@code EmptyRequestProto}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:EmptyRequestProto)
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProtoOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EmptyRequestProto_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EmptyRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -238,12 +238,12 @@ public final class TestProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -252,7 +252,7 @@ public final class TestProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EmptyRequestProto_descriptor;
}
@@ -279,29 +279,29 @@ public final class TestProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto) {
return mergeFrom((org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto)other);
} else {
@@ -322,13 +322,13 @@ public final class TestProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -339,12 +339,12 @@ public final class TestProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -362,22 +362,22 @@ public final class TestProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<EmptyRequestProto>
- PARSER = new com.google.protobuf.AbstractParser<EmptyRequestProto>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EmptyRequestProto>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<EmptyRequestProto>() {
public EmptyRequestProto parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new EmptyRequestProto(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<EmptyRequestProto> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EmptyRequestProto> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<EmptyRequestProto> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EmptyRequestProto> getParserForType() {
return PARSER;
}
@@ -389,34 +389,34 @@ public final class TestProtos {
public interface EmptyResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:EmptyResponseProto)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
}
/**
* Protobuf type {@code EmptyResponseProto}
*/
public static final class EmptyResponseProto extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:EmptyResponseProto)
EmptyResponseProtoOrBuilder {
// Use EmptyResponseProto.newBuilder() to construct.
- private EmptyResponseProto(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private EmptyResponseProto(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private EmptyResponseProto() {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private EmptyResponseProto(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -434,22 +434,22 @@ public final class TestProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EmptyResponseProto_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EmptyResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -466,7 +466,7 @@ public final class TestProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
unknownFields.writeTo(output);
}
@@ -510,61 +510,61 @@ public final class TestProtos {
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -582,7 +582,7 @@ public final class TestProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -590,15 +590,15 @@ public final class TestProtos {
* Protobuf type {@code EmptyResponseProto}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:EmptyResponseProto)
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProtoOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EmptyResponseProto_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EmptyResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -611,12 +611,12 @@ public final class TestProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -625,7 +625,7 @@ public final class TestProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EmptyResponseProto_descriptor;
}
@@ -652,29 +652,29 @@ public final class TestProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto) {
return mergeFrom((org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto)other);
} else {
@@ -695,13 +695,13 @@ public final class TestProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -712,12 +712,12 @@ public final class TestProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -735,22 +735,22 @@ public final class TestProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<EmptyResponseProto>
- PARSER = new com.google.protobuf.AbstractParser<EmptyResponseProto>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EmptyResponseProto>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<EmptyResponseProto>() {
public EmptyResponseProto parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new EmptyResponseProto(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<EmptyResponseProto> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EmptyResponseProto> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<EmptyResponseProto> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EmptyResponseProto> getParserForType() {
return PARSER;
}
@@ -762,7 +762,7 @@ public final class TestProtos {
public interface EchoRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:EchoRequestProto)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required string message = 1;</code>
@@ -775,18 +775,18 @@ public final class TestProtos {
/**
* <code>required string message = 1;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getMessageBytes();
}
/**
* Protobuf type {@code EchoRequestProto}
*/
public static final class EchoRequestProto extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:EchoRequestProto)
EchoRequestProtoOrBuilder {
// Use EchoRequestProto.newBuilder() to construct.
- private EchoRequestProto(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private EchoRequestProto(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private EchoRequestProto() {
@@ -794,18 +794,18 @@ public final class TestProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private EchoRequestProto(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -822,29 +822,29 @@ public final class TestProtos {
break;
}
case 10: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000001;
message_ = bs;
break;
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EchoRequestProto_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EchoRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -868,8 +868,8 @@ public final class TestProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
message_ = s;
@@ -880,17 +880,17 @@ public final class TestProtos {
/**
* <code>required string message = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getMessageBytes() {
java.lang.Object ref = message_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
message_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -908,10 +908,10 @@ public final class TestProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 1, message_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, message_);
}
unknownFields.writeTo(output);
}
@@ -922,7 +922,7 @@ public final class TestProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, message_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, message_);
}
size += unknownFields.getSerializedSize();
memoizedSize = size;
@@ -967,61 +967,61 @@ public final class TestProtos {
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -1039,7 +1039,7 @@ public final class TestProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1047,15 +1047,15 @@ public final class TestProtos {
* Protobuf type {@code EchoRequestProto}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:EchoRequestProto)
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProtoOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EchoRequestProto_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EchoRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1068,12 +1068,12 @@ public final class TestProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -1084,7 +1084,7 @@ public final class TestProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EchoRequestProto_descriptor;
}
@@ -1118,29 +1118,29 @@ public final class TestProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto) {
return mergeFrom((org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto)other);
} else {
@@ -1169,13 +1169,13 @@ public final class TestProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -1200,8 +1200,8 @@ public final class TestProtos {
public java.lang.String getMessage() {
java.lang.Object ref = message_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
message_ = s;
@@ -1214,17 +1214,17 @@ public final class TestProtos {
/**
* <code>required string message = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getMessageBytes() {
java.lang.Object ref = message_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
message_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -1253,7 +1253,7 @@ public final class TestProtos {
* <code>required string message = 1;</code>
*/
public Builder setMessageBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1263,12 +1263,12 @@ public final class TestProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -1286,22 +1286,22 @@ public final class TestProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<EchoRequestProto>
- PARSER = new com.google.protobuf.AbstractParser<EchoRequestProto>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EchoRequestProto>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<EchoRequestProto>() {
public EchoRequestProto parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new EchoRequestProto(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<EchoRequestProto> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EchoRequestProto> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<EchoRequestProto> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EchoRequestProto> getParserForType() {
return PARSER;
}
@@ -1313,7 +1313,7 @@ public final class TestProtos {
public interface EchoResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:EchoResponseProto)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required string message = 1;</code>
@@ -1326,18 +1326,18 @@ public final class TestProtos {
/**
* <code>required string message = 1;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getMessageBytes();
}
/**
* Protobuf type {@code EchoResponseProto}
*/
public static final class EchoResponseProto extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:EchoResponseProto)
EchoResponseProtoOrBuilder {
// Use EchoResponseProto.newBuilder() to construct.
- private EchoResponseProto(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private EchoResponseProto(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private EchoResponseProto() {
@@ -1345,18 +1345,18 @@ public final class TestProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private EchoResponseProto(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -1373,29 +1373,29 @@ public final class TestProtos {
break;
}
case 10: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000001;
message_ = bs;
break;
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EchoResponseProto_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EchoResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1419,8 +1419,8 @@ public final class TestProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
message_ = s;
@@ -1431,17 +1431,17 @@ public final class TestProtos {
/**
* <code>required string message = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getMessageBytes() {
java.lang.Object ref = message_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
message_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -1459,10 +1459,10 @@ public final class TestProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 1, message_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, message_);
}
unknownFields.writeTo(output);
}
@@ -1473,7 +1473,7 @@ public final class TestProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, message_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, message_);
}
size += unknownFields.getSerializedSize();
memoizedSize = size;
@@ -1518,61 +1518,61 @@ public final class TestProtos {
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -1590,7 +1590,7 @@ public final class TestProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1598,15 +1598,15 @@ public final class TestProtos {
* Protobuf type {@code EchoResponseProto}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:EchoResponseProto)
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProtoOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EchoResponseProto_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EchoResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1619,12 +1619,12 @@ public final class TestProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -1635,7 +1635,7 @@ public final class TestProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EchoResponseProto_descriptor;
}
@@ -1669,29 +1669,29 @@ public final class TestProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto) {
return mergeFrom((org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto)other);
} else {
@@ -1720,13 +1720,13 @@ public final class TestProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -1751,8 +1751,8 @@ public final class TestProtos {
public java.lang.String getMessage() {
java.lang.Object ref = message_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
message_ = s;
@@ -1765,17 +1765,17 @@ public final class TestProtos {
/**
* <code>required string message = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getMessageBytes() {
java.lang.Object ref = message_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
message_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -1804,7 +1804,7 @@ public final class TestProtos {
* <code>required string message = 1;</code>
*/
public Builder setMessageBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1814,12 +1814,12 @@ public final class TestProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -1837,22 +1837,22 @@ public final class TestProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<EchoResponseProto>
- PARSER = new com.google.protobuf.AbstractParser<EchoResponseProto>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EchoResponseProto>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<EchoResponseProto>() {
public EchoResponseProto parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new EchoResponseProto(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<EchoResponseProto> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EchoResponseProto> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<EchoResponseProto> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EchoResponseProto> getParserForType() {
return PARSER;
}
@@ -1864,7 +1864,7 @@ public final class TestProtos {
public interface PauseRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:PauseRequestProto)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required uint32 ms = 1;</code>
@@ -1879,11 +1879,11 @@ public final class TestProtos {
* Protobuf type {@code PauseRequestProto}
*/
public static final class PauseRequestProto extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:PauseRequestProto)
PauseRequestProtoOrBuilder {
// Use PauseRequestProto.newBuilder() to construct.
- private PauseRequestProto(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private PauseRequestProto(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private PauseRequestProto() {
@@ -1891,18 +1891,18 @@ public final class TestProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private PauseRequestProto(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -1925,22 +1925,22 @@ public final class TestProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_PauseRequestProto_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_PauseRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1977,7 +1977,7 @@ public final class TestProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeUInt32(1, ms_);
@@ -1991,7 +1991,7 @@ public final class TestProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt32Size(1, ms_);
}
size += unknownFields.getSerializedSize();
@@ -2037,61 +2037,61 @@ public final class TestProtos {
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -2109,7 +2109,7 @@ public final class TestProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -2117,15 +2117,15 @@ public final class TestProtos {
* Protobuf type {@code PauseRequestProto}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:PauseRequestProto)
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProtoOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_PauseRequestProto_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Fi
<TRUNCATED>
[08/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/RPCProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/RPCProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/RPCProtos.java
index ead61ed..3c2bec0 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/RPCProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/RPCProtos.java
@@ -6,17 +6,17 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class RPCProtos {
private RPCProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
public interface UserInformationOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.UserInformation)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required string effective_user = 1;</code>
@@ -29,7 +29,7 @@ public final class RPCProtos {
/**
* <code>required string effective_user = 1;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getEffectiveUserBytes();
/**
@@ -43,7 +43,7 @@ public final class RPCProtos {
/**
* <code>optional string real_user = 2;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getRealUserBytes();
}
/**
@@ -54,11 +54,11 @@ public final class RPCProtos {
* Protobuf type {@code hbase.pb.UserInformation}
*/
public static final class UserInformation extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.UserInformation)
UserInformationOrBuilder {
// Use UserInformation.newBuilder() to construct.
- private UserInformation(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private UserInformation(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private UserInformation() {
@@ -67,18 +67,18 @@ public final class RPCProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private UserInformation(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -95,35 +95,35 @@ public final class RPCProtos {
break;
}
case 10: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000001;
effectiveUser_ = bs;
break;
}
case 18: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000002;
realUser_ = bs;
break;
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_UserInformation_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_UserInformation_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -147,8 +147,8 @@ public final class RPCProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
effectiveUser_ = s;
@@ -159,17 +159,17 @@ public final class RPCProtos {
/**
* <code>required string effective_user = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getEffectiveUserBytes() {
java.lang.Object ref = effectiveUser_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
effectiveUser_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -189,8 +189,8 @@ public final class RPCProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
realUser_ = s;
@@ -201,17 +201,17 @@ public final class RPCProtos {
/**
* <code>optional string real_user = 2;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getRealUserBytes() {
java.lang.Object ref = realUser_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
realUser_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -229,13 +229,13 @@ public final class RPCProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 1, effectiveUser_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, effectiveUser_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 2, realUser_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 2, realUser_);
}
unknownFields.writeTo(output);
}
@@ -246,10 +246,10 @@ public final class RPCProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, effectiveUser_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, effectiveUser_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(2, realUser_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(2, realUser_);
}
size += unknownFields.getSerializedSize();
memoizedSize = size;
@@ -303,61 +303,61 @@ public final class RPCProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -375,7 +375,7 @@ public final class RPCProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -387,15 +387,15 @@ public final class RPCProtos {
* Protobuf type {@code hbase.pb.UserInformation}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.UserInformation)
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_UserInformation_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_UserInformation_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -408,12 +408,12 @@ public final class RPCProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -426,7 +426,7 @@ public final class RPCProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_UserInformation_descriptor;
}
@@ -464,29 +464,29 @@ public final class RPCProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation)other);
} else {
@@ -520,13 +520,13 @@ public final class RPCProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -551,8 +551,8 @@ public final class RPCProtos {
public java.lang.String getEffectiveUser() {
java.lang.Object ref = effectiveUser_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
effectiveUser_ = s;
@@ -565,17 +565,17 @@ public final class RPCProtos {
/**
* <code>required string effective_user = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getEffectiveUserBytes() {
java.lang.Object ref = effectiveUser_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
effectiveUser_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -604,7 +604,7 @@ public final class RPCProtos {
* <code>required string effective_user = 1;</code>
*/
public Builder setEffectiveUserBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -627,8 +627,8 @@ public final class RPCProtos {
public java.lang.String getRealUser() {
java.lang.Object ref = realUser_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
realUser_ = s;
@@ -641,17 +641,17 @@ public final class RPCProtos {
/**
* <code>optional string real_user = 2;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getRealUserBytes() {
java.lang.Object ref = realUser_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
realUser_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -680,7 +680,7 @@ public final class RPCProtos {
* <code>optional string real_user = 2;</code>
*/
public Builder setRealUserBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -690,12 +690,12 @@ public final class RPCProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -713,22 +713,22 @@ public final class RPCProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<UserInformation>
- PARSER = new com.google.protobuf.AbstractParser<UserInformation>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<UserInformation>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<UserInformation>() {
public UserInformation parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new UserInformation(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<UserInformation> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<UserInformation> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<UserInformation> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<UserInformation> getParserForType() {
return PARSER;
}
@@ -740,7 +740,7 @@ public final class RPCProtos {
public interface ConnectionHeaderOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.ConnectionHeader)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>optional .hbase.pb.UserInformation user_info = 1;</code>
@@ -766,7 +766,7 @@ public final class RPCProtos {
/**
* <code>optional string service_name = 2;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getServiceNameBytes();
/**
@@ -795,7 +795,7 @@ public final class RPCProtos {
*
* <code>optional string cell_block_codec_class = 3;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getCellBlockCodecClassBytes();
/**
@@ -824,7 +824,7 @@ public final class RPCProtos {
*
* <code>optional string cell_block_compressor_class = 4;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getCellBlockCompressorClassBytes();
/**
@@ -848,11 +848,11 @@ public final class RPCProtos {
* Protobuf type {@code hbase.pb.ConnectionHeader}
*/
public static final class ConnectionHeader extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.ConnectionHeader)
ConnectionHeaderOrBuilder {
// Use ConnectionHeader.newBuilder() to construct.
- private ConnectionHeader(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private ConnectionHeader(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private ConnectionHeader() {
@@ -862,18 +862,18 @@ public final class RPCProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private ConnectionHeader(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -903,19 +903,19 @@ public final class RPCProtos {
break;
}
case 18: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000002;
serviceName_ = bs;
break;
}
case 26: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000004;
cellBlockCodecClass_ = bs;
break;
}
case 34: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000008;
cellBlockCompressorClass_ = bs;
break;
@@ -935,22 +935,22 @@ public final class RPCProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_ConnectionHeader_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_ConnectionHeader_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -995,8 +995,8 @@ public final class RPCProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
serviceName_ = s;
@@ -1007,17 +1007,17 @@ public final class RPCProtos {
/**
* <code>optional string service_name = 2;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getServiceNameBytes() {
java.lang.Object ref = serviceName_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
serviceName_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -1047,8 +1047,8 @@ public final class RPCProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
cellBlockCodecClass_ = s;
@@ -1064,17 +1064,17 @@ public final class RPCProtos {
*
* <code>optional string cell_block_codec_class = 3;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getCellBlockCodecClassBytes() {
java.lang.Object ref = cellBlockCodecClass_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
cellBlockCodecClass_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -1104,8 +1104,8 @@ public final class RPCProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
cellBlockCompressorClass_ = s;
@@ -1121,17 +1121,17 @@ public final class RPCProtos {
*
* <code>optional string cell_block_compressor_class = 4;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getCellBlockCompressorClassBytes() {
java.lang.Object ref = cellBlockCompressorClass_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
cellBlockCompressorClass_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -1178,19 +1178,19 @@ public final class RPCProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeMessage(1, getUserInfo());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 2, serviceName_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 2, serviceName_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 3, cellBlockCodecClass_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 3, cellBlockCodecClass_);
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 4, cellBlockCompressorClass_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 4, cellBlockCompressorClass_);
}
if (((bitField0_ & 0x00000010) == 0x00000010)) {
output.writeMessage(5, getVersionInfo());
@@ -1204,20 +1204,20 @@ public final class RPCProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(1, getUserInfo());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(2, serviceName_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(2, serviceName_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(3, cellBlockCodecClass_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(3, cellBlockCodecClass_);
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(4, cellBlockCompressorClass_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(4, cellBlockCompressorClass_);
}
if (((bitField0_ & 0x00000010) == 0x00000010)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(5, getVersionInfo());
}
size += unknownFields.getSerializedSize();
@@ -1299,61 +1299,61 @@ public final class RPCProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -1371,7 +1371,7 @@ public final class RPCProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1383,15 +1383,15 @@ public final class RPCProtos {
* Protobuf type {@code hbase.pb.ConnectionHeader}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.ConnectionHeader)
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeaderOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_ConnectionHeader_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_ConnectionHeader_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1404,12 +1404,12 @@ public final class RPCProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getUserInfoFieldBuilder();
getVersionInfoFieldBuilder();
@@ -1438,7 +1438,7 @@ public final class RPCProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_ConnectionHeader_descriptor;
}
@@ -1496,29 +1496,29 @@ public final class RPCProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader)other);
} else {
@@ -1570,13 +1570,13 @@ public final class RPCProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -1589,7 +1589,7 @@ public final class RPCProtos {
private int bitField0_;
private org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation userInfo_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationOrBuilder> userInfoBuilder_;
/**
* <code>optional .hbase.pb.UserInformation user_info = 1;</code>
@@ -1692,11 +1692,11 @@ public final class RPCProtos {
/**
* <code>optional .hbase.pb.UserInformation user_info = 1;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationOrBuilder>
getUserInfoFieldBuilder() {
if (userInfoBuilder_ == null) {
- userInfoBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ userInfoBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationOrBuilder>(
getUserInfo(),
getParentForChildren(),
@@ -1719,8 +1719,8 @@ public final class RPCProtos {
public java.lang.String getServiceName() {
java.lang.Object ref = serviceName_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
serviceName_ = s;
@@ -1733,17 +1733,17 @@ public final class RPCProtos {
/**
* <code>optional string service_name = 2;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getServiceNameBytes() {
java.lang.Object ref = serviceName_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
serviceName_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -1772,7 +1772,7 @@ public final class RPCProtos {
* <code>optional string service_name = 2;</code>
*/
public Builder setServiceNameBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1805,8 +1805,8 @@ public final class RPCProtos {
public java.lang.String getCellBlockCodecClass() {
java.lang.Object ref = cellBlockCodecClass_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
cellBlockCodecClass_ = s;
@@ -1824,17 +1824,17 @@ public final class RPCProtos {
*
* <code>optional string cell_block_codec_class = 3;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getCellBlockCodecClassBytes() {
java.lang.Object ref = cellBlockCodecClass_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
cellBlockCodecClass_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -1878,7 +1878,7 @@ public final class RPCProtos {
* <code>optional string cell_block_codec_class = 3;</code>
*/
public Builder setCellBlockCodecClassBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1911,8 +1911,8 @@ public final class RPCProtos {
public java.lang.String getCellBlockCompressorClass() {
java.lang.Object ref = cellBlockCompressorClass_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
cellBlockCompressorClass_ = s;
@@ -1930,17 +1930,17 @@ public final class RPCProtos {
*
* <code>optional string cell_block_compressor_class = 4;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getCellBlockCompressorClassBytes() {
java.lang.Object ref = cellBlockCompressorClass_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
cellBlockCompressorClass_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -1984,7 +1984,7 @@ public final class RPCProtos {
* <code>optional string cell_block_compressor_class = 4;</code>
*/
public Builder setCellBlockCompressorClassBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1995,7 +1995,7 @@ public final class RPCProtos {
}
private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo versionInfo_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfoOrBuilder> versionInfoBuilder_;
/**
* <code>optional .hbase.pb.VersionInfo version_info = 5;</code>
@@ -2098,11 +2098,11 @@ public final class RPCProtos {
/**
* <code>optional .hbase.pb.VersionInfo version_info = 5;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfoOrBuilder>
getVersionInfoFieldBuilder() {
if (versionInfoBuilder_ == null) {
- versionInfoBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ versionInfoBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfoOrBuilder>(
getVersionInfo(),
getParentForChildren(),
@@ -2112,12 +2112,12 @@ public final class RPCProtos {
return versionInfoBuilder_;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -2135,22 +2135,22 @@ public final class RPCProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<ConnectionHeader>
- PARSER = new com.google.protobuf.AbstractParser<ConnectionHeader>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ConnectionHeader>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<ConnectionHeader>() {
public ConnectionHeader parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new ConnectionHeader(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<ConnectionHeader> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ConnectionHeader> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<ConnectionHeader> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ConnectionHeader> getParserForType() {
return PARSER;
}
@@ -2162,7 +2162,7 @@ public final class RPCProtos {
public interface CellBlockMetaOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.CellBlockMeta)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <pre>
@@ -2189,11 +2189,11 @@ public final class RPCProtos {
* Protobuf type {@code hbase.pb.CellBlockMeta}
*/
public static final class CellBlockMeta extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.CellBlockMeta)
CellBlockMetaOrBuilder {
// Use CellBlockMeta.newBuilder() to construct.
- private CellBlockMeta(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private CellBlockMeta(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private CellBlockMeta() {
@@ -2201,18 +2201,18 @@ public final class RPCProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private CellBlockMeta(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -2235,22 +2235,22 @@ public final class RPCProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_CellBlockMeta_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_CellBlockMeta_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2291,7 +2291,7 @@ public final class RPCProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeUInt32(1, length_);
@@ -2305,7 +2305,7 @@ public final class RPCProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt32Size(1, length_);
}
size += unknownFields.getSerializedSize();
@@ -2351,61 +2351,61 @@ public final class RPCProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -2423,7 +2423,7 @@ public final class RPCProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -2435,15 +2435,15 @@ public final class RPCProtos {
* Protobuf type {@code hbase.pb.CellBlockMeta}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.CellBlockMeta)
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_CellBlockMeta_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_CellBlockMeta_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2456,12 +2456,12 @@ public final class RPCProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -2472,7 +2472,7 @@ public final class RPCProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_CellBlockMeta_descriptor;
}
@@ -2506,29 +2506,29 @@ public final class RPCProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta)other);
} else {
@@ -2552,13 +2552,13 @@ public final class RPCProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -2618,12 +2618,12 @@ public final class RPCProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -2641,22 +2641,22 @@ public final class RPCProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<CellBlockMeta>
- PARSER = new com.google.protobuf.AbstractParser<CellBlockMeta>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<CellBlockMeta>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<CellBlockMeta>() {
public CellBlockMeta parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new CellBlockMeta(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<CellBlockMeta> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<CellBlockMeta> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<CellBlockMeta> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<CellBlockMeta> getParserForType() {
return PARSER;
}
@@ -2668,7 +2668,7 @@ public final class RPCProtos {
public interface ExceptionResponseOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.ExceptionResponse)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <pre>
@@ -2693,7 +2693,7 @@ public final class RPCProtos {
*
* <code>optional string exception_class_name = 1;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getExceptionClassNameBytes();
/**
@@ -2719,7 +2719,7 @@ public final class RPCProtos {
*
* <code>optional string stack_trace = 2;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getStackTraceBytes();
/**
@@ -2748,7 +2748,7 @@ public final class RPCProtos {
*
* <code>optional string hostname = 3;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getHostnameBytes();
/**
@@ -2786,11 +2786,11 @@ public final class RPCProtos {
* Protobuf type {@code hbase.pb.ExceptionResponse}
*/
public static final class ExceptionResponse extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.ExceptionResponse)
ExceptionResponseOrBuilder {
// Use ExceptionResponse.newBuilder() to construct.
- private ExceptionResponse(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private ExceptionResponse(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private ExceptionResponse() {
@@ -2802,18 +2802,18 @@ public final class RPCProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private ExceptionResponse(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -2830,19 +2830,19 @@ public final class RPCProtos {
break;
}
case 10: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000001;
exceptionClassName_ = bs;
break;
}
case 18: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000002;
stackTrace_ = bs;
break;
}
case 26: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000004;
hostname_ = bs;
break;
@@ -2859,22 +2859,22 @@ public final class RPCProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_ExceptionResponse_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_ExceptionResponse_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2906,8 +2906,8 @@ public final class RPCProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
exceptionClassName_ = s;
@@ -2922,17 +2922,17 @@ public final class RPCProtos {
*
* <code>optional string exception_class_name = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getExceptionClassNameBytes() {
java.lang.Object ref = exceptionClassName_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
exceptionClassName_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -2960,8 +2960,8 @@ public final class RPCProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
stackTrace_ = s;
@@ -2976,17 +2976,17 @@ public final class RPCProtos {
*
* <code>optional string stack_trace = 2;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getStackTraceBytes() {
java.lang.Object ref = stackTrace_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
stackTrace_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -3016,8 +3016,8 @@ public final class RPCProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
hostname_ = s;
@@ -3033,17 +3033,17 @@ public final class RPCProtos {
*
* <code>optional string hostname = 3;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getHostnameBytes() {
java.lang.Object ref = hostname_;
if (ref in
<TRUNCATED>
[09/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/QuotaProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/QuotaProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/QuotaProtos.java
index d31e338..d14336a 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/QuotaProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/QuotaProtos.java
@@ -6,19 +6,19 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class QuotaProtos {
private QuotaProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
/**
* Protobuf enum {@code hbase.pb.QuotaScope}
*/
public enum QuotaScope
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>CLUSTER = 1;</code>
*/
@@ -59,27 +59,27 @@ public final class QuotaProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<QuotaScope>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<QuotaScope>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
QuotaScope> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<QuotaScope>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<QuotaScope>() {
public QuotaScope findValueByNumber(int number) {
return QuotaScope.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.getDescriptor().getEnumTypes().get(0);
}
@@ -87,7 +87,7 @@ public final class QuotaProtos {
private static final QuotaScope[] VALUES = values();
public static QuotaScope valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -108,7 +108,7 @@ public final class QuotaProtos {
* Protobuf enum {@code hbase.pb.ThrottleType}
*/
public enum ThrottleType
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>REQUEST_NUMBER = 1;</code>
*/
@@ -185,27 +185,27 @@ public final class QuotaProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<ThrottleType>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<ThrottleType>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
ThrottleType> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<ThrottleType>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<ThrottleType>() {
public ThrottleType findValueByNumber(int number) {
return ThrottleType.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.getDescriptor().getEnumTypes().get(1);
}
@@ -213,7 +213,7 @@ public final class QuotaProtos {
private static final ThrottleType[] VALUES = values();
public static ThrottleType valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -234,7 +234,7 @@ public final class QuotaProtos {
* Protobuf enum {@code hbase.pb.QuotaType}
*/
public enum QuotaType
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>THROTTLE = 1;</code>
*/
@@ -266,27 +266,27 @@ public final class QuotaProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<QuotaType>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<QuotaType>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
QuotaType> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<QuotaType>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<QuotaType>() {
public QuotaType findValueByNumber(int number) {
return QuotaType.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.getDescriptor().getEnumTypes().get(2);
}
@@ -294,7 +294,7 @@ public final class QuotaProtos {
private static final QuotaType[] VALUES = values();
public static QuotaType valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -313,7 +313,7 @@ public final class QuotaProtos {
public interface TimedQuotaOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.TimedQuota)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required .hbase.pb.TimeUnit time_unit = 1;</code>
@@ -355,11 +355,11 @@ public final class QuotaProtos {
* Protobuf type {@code hbase.pb.TimedQuota}
*/
public static final class TimedQuota extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.TimedQuota)
TimedQuotaOrBuilder {
// Use TimedQuota.newBuilder() to construct.
- private TimedQuota(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private TimedQuota(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private TimedQuota() {
@@ -370,18 +370,18 @@ public final class QuotaProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private TimedQuota(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -431,22 +431,22 @@ public final class QuotaProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.internal_static_hbase_pb_TimedQuota_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.internal_static_hbase_pb_TimedQuota_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -530,7 +530,7 @@ public final class QuotaProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeEnum(1, timeUnit_);
@@ -553,19 +553,19 @@ public final class QuotaProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeEnumSize(1, timeUnit_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(2, softLimit_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeFloatSize(3, share_);
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeEnumSize(4, scope_);
}
size += unknownFields.getSerializedSize();
@@ -622,7 +622,7 @@ public final class QuotaProtos {
}
if (hasSoftLimit()) {
hash = (37 * hash) + SOFT_LIMIT_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getSoftLimit());
}
if (hasShare()) {
@@ -640,61 +640,61 @@ public final class QuotaProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -712,7 +712,7 @@ public final class QuotaProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -720,15 +720,15 @@ public final class QuotaProtos {
* Protobuf type {@code hbase.pb.TimedQuota}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.TimedQuota)
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.internal_static_hbase_pb_TimedQuota_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.internal_static_hbase_pb_TimedQuota_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -741,12 +741,12 @@ public final class QuotaProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -763,7 +763,7 @@ public final class QuotaProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.internal_static_hbase_pb_TimedQuota_descriptor;
}
@@ -809,29 +809,29 @@ public final class QuotaProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota)other);
} else {
@@ -867,13 +867,13 @@ public final class QuotaProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -1021,12 +1021,12 @@ public final class QuotaProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -1044,22 +1044,22 @@ public final class QuotaProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<TimedQuota>
- PARSER = new com.google.protobuf.AbstractParser<TimedQuota>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<TimedQuota>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<TimedQuota>() {
public TimedQuota parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new TimedQuota(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<TimedQuota> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<TimedQuota> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<TimedQuota> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<TimedQuota> getParserForType() {
return PARSER;
}
@@ -1071,7 +1071,7 @@ public final class QuotaProtos {
public interface ThrottleOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.Throttle)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>optional .hbase.pb.TimedQuota req_num = 1;</code>
@@ -1155,29 +1155,29 @@ public final class QuotaProtos {
* Protobuf type {@code hbase.pb.Throttle}
*/
public static final class Throttle extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.Throttle)
ThrottleOrBuilder {
// Use Throttle.newBuilder() to construct.
- private Throttle(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private Throttle(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private Throttle() {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private Throttle(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -1273,22 +1273,22 @@ public final class QuotaProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.internal_static_hbase_pb_Throttle_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.internal_static_hbase_pb_Throttle_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1468,7 +1468,7 @@ public final class QuotaProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeMessage(1, getReqNum());
@@ -1497,27 +1497,27 @@ public final class QuotaProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(1, getReqNum());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(2, getReqSize());
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(3, getWriteNum());
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(4, getWriteSize());
}
if (((bitField0_ & 0x00000010) == 0x00000010)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(5, getReadNum());
}
if (((bitField0_ & 0x00000020) == 0x00000020)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(6, getReadSize());
}
size += unknownFields.getSerializedSize();
@@ -1608,61 +1608,61 @@ public final class QuotaProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Throttle parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Throttle parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Throttle parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Throttle parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Throttle parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Throttle parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Throttle parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Throttle parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Throttle parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Throttle parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -1680,7 +1680,7 @@ public final class QuotaProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1688,15 +1688,15 @@ public final class QuotaProtos {
* Protobuf type {@code hbase.pb.Throttle}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.Throttle)
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.internal_static_hbase_pb_Throttle_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.internal_static_hbase_pb_Throttle_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1709,12 +1709,12 @@ public final class QuotaProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getReqNumFieldBuilder();
getReqSizeFieldBuilder();
@@ -1765,7 +1765,7 @@ public final class QuotaProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.internal_static_hbase_pb_Throttle_descriptor;
}
@@ -1843,29 +1843,29 @@ public final class QuotaProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Throttle) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Throttle)other);
} else {
@@ -1934,13 +1934,13 @@ public final class QuotaProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Throttle parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Throttle) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -1953,7 +1953,7 @@ public final class QuotaProtos {
private int bitField0_;
private org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota reqNum_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder> reqNumBuilder_;
/**
* <code>optional .hbase.pb.TimedQuota req_num = 1;</code>
@@ -2056,11 +2056,11 @@ public final class QuotaProtos {
/**
* <code>optional .hbase.pb.TimedQuota req_num = 1;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>
getReqNumFieldBuilder() {
if (reqNumBuilder_ == null) {
- reqNumBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ reqNumBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>(
getReqNum(),
getParentForChildren(),
@@ -2071,7 +2071,7 @@ public final class QuotaProtos {
}
private org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota reqSize_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder> reqSizeBuilder_;
/**
* <code>optional .hbase.pb.TimedQuota req_size = 2;</code>
@@ -2174,11 +2174,11 @@ public final class QuotaProtos {
/**
* <code>optional .hbase.pb.TimedQuota req_size = 2;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>
getReqSizeFieldBuilder() {
if (reqSizeBuilder_ == null) {
- reqSizeBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ reqSizeBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>(
getReqSize(),
getParentForChildren(),
@@ -2189,7 +2189,7 @@ public final class QuotaProtos {
}
private org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota writeNum_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder> writeNumBuilder_;
/**
* <code>optional .hbase.pb.TimedQuota write_num = 3;</code>
@@ -2292,11 +2292,11 @@ public final class QuotaProtos {
/**
* <code>optional .hbase.pb.TimedQuota write_num = 3;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>
getWriteNumFieldBuilder() {
if (writeNumBuilder_ == null) {
- writeNumBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ writeNumBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>(
getWriteNum(),
getParentForChildren(),
@@ -2307,7 +2307,7 @@ public final class QuotaProtos {
}
private org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota writeSize_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder> writeSizeBuilder_;
/**
* <code>optional .hbase.pb.TimedQuota write_size = 4;</code>
@@ -2410,11 +2410,11 @@ public final class QuotaProtos {
/**
* <code>optional .hbase.pb.TimedQuota write_size = 4;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>
getWriteSizeFieldBuilder() {
if (writeSizeBuilder_ == null) {
- writeSizeBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ writeSizeBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>(
getWriteSize(),
getParentForChildren(),
@@ -2425,7 +2425,7 @@ public final class QuotaProtos {
}
private org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota readNum_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder> readNumBuilder_;
/**
* <code>optional .hbase.pb.TimedQuota read_num = 5;</code>
@@ -2528,11 +2528,11 @@ public final class QuotaProtos {
/**
* <code>optional .hbase.pb.TimedQuota read_num = 5;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>
getReadNumFieldBuilder() {
if (readNumBuilder_ == null) {
- readNumBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ readNumBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>(
getReadNum(),
getParentForChildren(),
@@ -2543,7 +2543,7 @@ public final class QuotaProtos {
}
private org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota readSize_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder> readSizeBuilder_;
/**
* <code>optional .hbase.pb.TimedQuota read_size = 6;</code>
@@ -2646,11 +2646,11 @@ public final class QuotaProtos {
/**
* <code>optional .hbase.pb.TimedQuota read_size = 6;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>
getReadSizeFieldBuilder() {
if (readSizeBuilder_ == null) {
- readSizeBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ readSizeBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>(
getReadSize(),
getParentForChildren(),
@@ -2660,12 +2660,12 @@ public final class QuotaProtos {
return readSizeBuilder_;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -2683,22 +2683,22 @@ public final class QuotaProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<Throttle>
- PARSER = new com.google.protobuf.AbstractParser<Throttle>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Throttle>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Throttle>() {
public Throttle parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new Throttle(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<Throttle> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Throttle> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<Throttle> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Throttle> getParserForType() {
return PARSER;
}
@@ -2710,7 +2710,7 @@ public final class QuotaProtos {
public interface ThrottleRequestOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.ThrottleRequest)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>optional .hbase.pb.ThrottleType type = 1;</code>
@@ -2738,11 +2738,11 @@ public final class QuotaProtos {
* Protobuf type {@code hbase.pb.ThrottleRequest}
*/
public static final class ThrottleRequest extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.ThrottleRequest)
ThrottleRequestOrBuilder {
// Use ThrottleRequest.newBuilder() to construct.
- private ThrottleRequest(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private ThrottleRequest(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private ThrottleRequest() {
@@ -2750,18 +2750,18 @@ public final class QuotaProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private ThrottleRequest(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -2803,22 +2803,22 @@ public final class QuotaProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.internal_static_hbase_pb_ThrottleRequest_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.internal_static_hbase_pb_ThrottleRequest_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2879,7 +2879,7 @@ public final class QuotaProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeEnum(1, type_);
@@ -2896,11 +2896,11 @@ public final class QuotaProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeEnumSize(1, type_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(2, getTimedQuota());
}
size += unknownFields.getSerializedSize();
@@ -2954,61 +2954,61 @@ public final class QuotaProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -3026,7 +3026,7 @@ public final class QuotaProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -3034,15 +3034,15 @@ public final class QuotaProtos {
* Protobuf type {@code hbase.pb.ThrottleRequest}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.ThrottleRequest)
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequestOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.internal_static_hbase_pb_ThrottleRequest_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.internal_static_hbase_pb_ThrottleRequest_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -3055,12 +3055,12 @@ public final class QuotaProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getTimedQuotaFieldBuilder();
}
@@ -3078,7 +3078,7 @@ public final class QuotaProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.internal_static_hbase_pb_ThrottleRequest_descriptor;
}
@@ -3120,29 +3120,29 @@ public final class QuotaProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest)other);
} else {
@@ -3174,13 +3174,13 @@ public final class QuotaProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -3229,7 +3229,7 @@ public final class QuotaProtos {
}
private org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota timedQuota_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder> timedQuotaBuilder_;
/**
* <code>optional .hbase.pb.TimedQuota timed_quota = 2;</code>
@@ -3332,11 +3332,11 @@ public final class QuotaProtos {
/**
* <code>optional .hbase.pb.TimedQuota timed_quota = 2;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>
getTimedQuotaFieldBuilder() {
if (timedQuotaBuilder_ == null) {
- timedQuotaBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ timedQuotaBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>(
getTimedQuota(),
getParentForChildren(),
@@ -3346,12 +3346,12 @@ public final class QuotaProtos {
return timedQuotaBuilder_;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -3369,22 +3369,22 @@ public final class QuotaProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<ThrottleRequest>
- PARSER = new com.google.protobuf.AbstractParser<ThrottleRequest>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ThrottleRequest>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<ThrottleRequest>() {
public ThrottleRequest parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new ThrottleRequest(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<ThrottleRequest> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ThrottleRequest> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<ThrottleRequest> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ThrottleRequest> getParserForType() {
return PARSER;
}
@@ -3396,7 +3396,7 @@ public final class QuotaProtos {
public interface QuotasOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.Quotas)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>optional bool bypass_globals = 1 [default = false];</code>
@@ -3424,11 +3424,11 @@ public final class QuotaProtos {
* Protobuf type {@code hbase.pb.Quotas}
*/
public static final class Quotas extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.Quotas)
QuotasOrBuilder {
// Use Quotas.newBuilder() to construct.
- private Quotas(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private Quotas(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private Quotas() {
@@ -3436,18 +3436,18 @@ public final class QuotaProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private Quotas(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -3483,22 +3483,22 @@ public final class QuotaProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.internal_static_hbase_pb_Quotas_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.internal_static_hbase_pb_Quotas_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -3558,7 +3558,7 @@ public final class QuotaProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBool(1, bypassGlobals_);
@@ -3575,11 +3575,11 @@ public final class QuotaProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBoolSize(1, bypassGlobals_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(2, getThrottle());
}
size += unknownFields.getSerializedSize();
@@ -3622,7 +3622,7 @@ public final class QuotaProtos {
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasBypassGlobals()) {
hash = (37 * hash) + BYPASS_GLOBALS_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean(
getBypassGlobals());
}
if (hasThrottle()) {
@@ -3635,61 +3635,61 @@ public final class QuotaProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -3707,7 +3707,7 @@ public final class QuotaProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -3715,15 +3715,15 @@ public final class QuotaProtos {
* Protobuf type {@code hbase.pb.Quotas}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.Quotas)
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.QuotasOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.internal_static_hbase_pb_Quotas_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.internal_static_hbase_pb_Quotas_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -3736,12 +3736,12 @@ public final class QuotaProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getThrottleFieldBuilder();
}
@@ -3759,7 +3759,7 @@ public final class QuotaProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.internal_static_hbase_pb_Quotas_descriptor;
}
@@ -3801,29 +3801,29 @@ public final class QuotaProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field,
<TRUNCATED>
[26/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestRpcServiceProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestRpcServiceProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestRpcServiceProtos.java
index 86d12f3..200a7f4 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestRpcServiceProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestRpcServiceProtos.java
@@ -6,13 +6,13 @@ package org.apache.hadoop.hbase.shaded.ipc.protobuf.generated;
public final class TestRpcServiceProtos {
private TestRpcServiceProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
/**
* <pre>
@@ -23,7 +23,7 @@ public final class TestRpcServiceProtos {
* Protobuf service {@code TestProtobufRpcProto}
*/
public static abstract class TestProtobufRpcProto
- implements com.google.protobuf.Service {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.Service {
protected TestProtobufRpcProto() {}
public interface Interface {
@@ -31,103 +31,103 @@ public final class TestRpcServiceProtos {
* <code>rpc ping(.EmptyRequestProto) returns (.EmptyResponseProto);</code>
*/
public abstract void ping(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done);
/**
* <code>rpc echo(.EchoRequestProto) returns (.EchoResponseProto);</code>
*/
public abstract void echo(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto> done);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto> done);
/**
* <code>rpc error(.EmptyRequestProto) returns (.EmptyResponseProto);</code>
*/
public abstract void error(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done);
/**
* <code>rpc pause(.PauseRequestProto) returns (.EmptyResponseProto);</code>
*/
public abstract void pause(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done);
/**
* <code>rpc addr(.EmptyRequestProto) returns (.AddrResponseProto);</code>
*/
public abstract void addr(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto> done);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto> done);
}
- public static com.google.protobuf.Service newReflectiveService(
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Service newReflectiveService(
final Interface impl) {
return new TestProtobufRpcProto() {
@java.lang.Override
public void ping(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done) {
impl.ping(controller, request, done);
}
@java.lang.Override
public void echo(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto> done) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto> done) {
impl.echo(controller, request, done);
}
@java.lang.Override
public void error(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done) {
impl.error(controller, request, done);
}
@java.lang.Override
public void pause(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done) {
impl.pause(controller, request, done);
}
@java.lang.Override
public void addr(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto> done) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto> done) {
impl.addr(controller, request, done);
}
};
}
- public static com.google.protobuf.BlockingService
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService
newReflectiveBlockingService(final BlockingInterface impl) {
- return new com.google.protobuf.BlockingService() {
- public final com.google.protobuf.Descriptors.ServiceDescriptor
+ return new org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService() {
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public final com.google.protobuf.Message callBlockingMethod(
- com.google.protobuf.Descriptors.MethodDescriptor method,
- com.google.protobuf.RpcController controller,
- com.google.protobuf.Message request)
- throws com.google.protobuf.ServiceException {
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Message callBlockingMethod(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Message request)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException {
if (method.getService() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"Service.callBlockingMethod() given method descriptor for " +
@@ -149,9 +149,9 @@ public final class TestRpcServiceProtos {
}
}
- public final com.google.protobuf.Message
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Message
getRequestPrototype(
- com.google.protobuf.Descriptors.MethodDescriptor method) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) {
if (method.getService() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"Service.getRequestPrototype() given method " +
@@ -173,9 +173,9 @@ public final class TestRpcServiceProtos {
}
}
- public final com.google.protobuf.Message
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Message
getResponsePrototype(
- com.google.protobuf.Descriptors.MethodDescriptor method) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) {
if (method.getService() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"Service.getResponsePrototype() given method " +
@@ -204,58 +204,58 @@ public final class TestRpcServiceProtos {
* <code>rpc ping(.EmptyRequestProto) returns (.EmptyResponseProto);</code>
*/
public abstract void ping(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done);
/**
* <code>rpc echo(.EchoRequestProto) returns (.EchoResponseProto);</code>
*/
public abstract void echo(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto> done);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto> done);
/**
* <code>rpc error(.EmptyRequestProto) returns (.EmptyResponseProto);</code>
*/
public abstract void error(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done);
/**
* <code>rpc pause(.PauseRequestProto) returns (.EmptyResponseProto);</code>
*/
public abstract void pause(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done);
/**
* <code>rpc addr(.EmptyRequestProto) returns (.AddrResponseProto);</code>
*/
public abstract void addr(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto> done);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto> done);
public static final
- com.google.protobuf.Descriptors.ServiceDescriptor
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestRpcServiceProtos.getDescriptor().getServices().get(0);
}
- public final com.google.protobuf.Descriptors.ServiceDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor
getDescriptorForType() {
return getDescriptor();
}
public final void callMethod(
- com.google.protobuf.Descriptors.MethodDescriptor method,
- com.google.protobuf.RpcController controller,
- com.google.protobuf.Message request,
- com.google.protobuf.RpcCallback<
- com.google.protobuf.Message> done) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Message request,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Message> done) {
if (method.getService() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"Service.callMethod() given method descriptor for wrong " +
@@ -264,27 +264,27 @@ public final class TestRpcServiceProtos {
switch(method.getIndex()) {
case 0:
this.ping(controller, (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto)request,
- com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto>specializeCallback(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto>specializeCallback(
done));
return;
case 1:
this.echo(controller, (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto)request,
- com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto>specializeCallback(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto>specializeCallback(
done));
return;
case 2:
this.error(controller, (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto)request,
- com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto>specializeCallback(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto>specializeCallback(
done));
return;
case 3:
this.pause(controller, (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto)request,
- com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto>specializeCallback(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto>specializeCallback(
done));
return;
case 4:
this.addr(controller, (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto)request,
- com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto>specializeCallback(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto>specializeCallback(
done));
return;
default:
@@ -292,9 +292,9 @@ public final class TestRpcServiceProtos {
}
}
- public final com.google.protobuf.Message
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Message
getRequestPrototype(
- com.google.protobuf.Descriptors.MethodDescriptor method) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) {
if (method.getService() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"Service.getRequestPrototype() given method " +
@@ -316,9 +316,9 @@ public final class TestRpcServiceProtos {
}
}
- public final com.google.protobuf.Message
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Message
getResponsePrototype(
- com.google.protobuf.Descriptors.MethodDescriptor method) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) {
if (method.getService() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"Service.getResponsePrototype() given method " +
@@ -341,91 +341,91 @@ public final class TestRpcServiceProtos {
}
public static Stub newStub(
- com.google.protobuf.RpcChannel channel) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcChannel channel) {
return new Stub(channel);
}
public static final class Stub extends org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestRpcServiceProtos.TestProtobufRpcProto implements Interface {
- private Stub(com.google.protobuf.RpcChannel channel) {
+ private Stub(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcChannel channel) {
this.channel = channel;
}
- private final com.google.protobuf.RpcChannel channel;
+ private final org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcChannel channel;
- public com.google.protobuf.RpcChannel getChannel() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcChannel getChannel() {
return channel;
}
public void ping(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done) {
channel.callMethod(
getDescriptor().getMethods().get(0),
controller,
request,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance(),
- com.google.protobuf.RpcUtil.generalizeCallback(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback(
done,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.class,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance()));
}
public void echo(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto> done) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto> done) {
channel.callMethod(
getDescriptor().getMethods().get(1),
controller,
request,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto.getDefaultInstance(),
- com.google.protobuf.RpcUtil.generalizeCallback(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback(
done,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto.class,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto.getDefaultInstance()));
}
public void error(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done) {
channel.callMethod(
getDescriptor().getMethods().get(2),
controller,
request,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance(),
- com.google.protobuf.RpcUtil.generalizeCallback(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback(
done,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.class,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance()));
}
public void pause(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done) {
channel.callMethod(
getDescriptor().getMethods().get(3),
controller,
request,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance(),
- com.google.protobuf.RpcUtil.generalizeCallback(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback(
done,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.class,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance()));
}
public void addr(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto> done) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto> done) {
channel.callMethod(
getDescriptor().getMethods().get(4),
controller,
request,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto.getDefaultInstance(),
- com.google.protobuf.RpcUtil.generalizeCallback(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback(
done,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto.class,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto.getDefaultInstance()));
@@ -433,48 +433,48 @@ public final class TestRpcServiceProtos {
}
public static BlockingInterface newBlockingStub(
- com.google.protobuf.BlockingRpcChannel channel) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel) {
return new BlockingStub(channel);
}
public interface BlockingInterface {
public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto ping(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request)
- throws com.google.protobuf.ServiceException;
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto echo(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto request)
- throws com.google.protobuf.ServiceException;
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto error(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request)
- throws com.google.protobuf.ServiceException;
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto pause(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto request)
- throws com.google.protobuf.ServiceException;
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto addr(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request)
- throws com.google.protobuf.ServiceException;
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
}
private static final class BlockingStub implements BlockingInterface {
- private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) {
+ private BlockingStub(org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel) {
this.channel = channel;
}
- private final com.google.protobuf.BlockingRpcChannel channel;
+ private final org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel;
public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto ping(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request)
- throws com.google.protobuf.ServiceException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException {
return (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto) channel.callBlockingMethod(
getDescriptor().getMethods().get(0),
controller,
@@ -484,9 +484,9 @@ public final class TestRpcServiceProtos {
public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto echo(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto request)
- throws com.google.protobuf.ServiceException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException {
return (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto) channel.callBlockingMethod(
getDescriptor().getMethods().get(1),
controller,
@@ -496,9 +496,9 @@ public final class TestRpcServiceProtos {
public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto error(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request)
- throws com.google.protobuf.ServiceException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException {
return (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto) channel.callBlockingMethod(
getDescriptor().getMethods().get(2),
controller,
@@ -508,9 +508,9 @@ public final class TestRpcServiceProtos {
public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto pause(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto request)
- throws com.google.protobuf.ServiceException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException {
return (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto) channel.callBlockingMethod(
getDescriptor().getMethods().get(3),
controller,
@@ -520,9 +520,9 @@ public final class TestRpcServiceProtos {
public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto addr(
- com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request)
- throws com.google.protobuf.ServiceException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException {
return (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto) channel.callBlockingMethod(
getDescriptor().getMethods().get(4),
controller,
@@ -536,11 +536,11 @@ public final class TestRpcServiceProtos {
}
- public static com.google.protobuf.Descriptors.FileDescriptor
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
getDescriptor() {
return descriptor;
}
- private static com.google.protobuf.Descriptors.FileDescriptor
+ private static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
descriptor;
static {
java.lang.String[] descriptorData = {
@@ -555,17 +555,17 @@ public final class TestRpcServiceProtos {
"hadoop.hbase.shaded.ipc.protobuf.generat" +
"edB\024TestRpcServiceProtos\210\001\001\240\001\001"
};
- com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
- new com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
- public com.google.protobuf.ExtensionRegistry assignDescriptors(
- com.google.protobuf.Descriptors.FileDescriptor root) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
descriptor = root;
return null;
}
};
- com.google.protobuf.Descriptors.FileDescriptor
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
.internalBuildGeneratedFileFrom(descriptorData,
- new com.google.protobuf.Descriptors.FileDescriptor[] {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.getDescriptor(),
}, assigner);
org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.getDescriptor();
[21/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ClusterStatusProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ClusterStatusProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ClusterStatusProtos.java
index f945184..97c6d05 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ClusterStatusProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ClusterStatusProtos.java
@@ -6,17 +6,17 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class ClusterStatusProtos {
private ClusterStatusProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
public interface RegionStateOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.RegionState)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required .hbase.pb.RegionInfo region_info = 1;</code>
@@ -53,11 +53,11 @@ public final class ClusterStatusProtos {
* Protobuf type {@code hbase.pb.RegionState}
*/
public static final class RegionState extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.RegionState)
RegionStateOrBuilder {
// Use RegionState.newBuilder() to construct.
- private RegionState(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private RegionState(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private RegionState() {
@@ -66,18 +66,18 @@ public final class ClusterStatusProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private RegionState(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -124,22 +124,22 @@ public final class ClusterStatusProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.internal_static_hbase_pb_RegionState_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.internal_static_hbase_pb_RegionState_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -150,7 +150,7 @@ public final class ClusterStatusProtos {
* Protobuf enum {@code hbase.pb.RegionState.State}
*/
public enum State
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <pre>
* region is in an offline state
@@ -430,27 +430,27 @@ public final class ClusterStatusProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<State>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<State>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
State> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<State>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<State>() {
public State findValueByNumber(int number) {
return State.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState.getDescriptor().getEnumTypes().get(0);
}
@@ -458,7 +458,7 @@ public final class ClusterStatusProtos {
private static final State[] VALUES = values();
public static State valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -550,7 +550,7 @@ public final class ClusterStatusProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeMessage(1, getRegionInfo());
@@ -570,15 +570,15 @@ public final class ClusterStatusProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(1, getRegionInfo());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeEnumSize(2, state_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(3, stamp_);
}
size += unknownFields.getSerializedSize();
@@ -633,7 +633,7 @@ public final class ClusterStatusProtos {
}
if (hasStamp()) {
hash = (37 * hash) + STAMP_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getStamp());
}
hash = (29 * hash) + unknownFields.hashCode();
@@ -642,61 +642,61 @@ public final class ClusterStatusProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -714,7 +714,7 @@ public final class ClusterStatusProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -722,15 +722,15 @@ public final class ClusterStatusProtos {
* Protobuf type {@code hbase.pb.RegionState}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.RegionState)
org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStateOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.internal_static_hbase_pb_RegionState_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.internal_static_hbase_pb_RegionState_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -743,12 +743,12 @@ public final class ClusterStatusProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getRegionInfoFieldBuilder();
}
@@ -768,7 +768,7 @@ public final class ClusterStatusProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.internal_static_hbase_pb_RegionState_descriptor;
}
@@ -814,29 +814,29 @@ public final class ClusterStatusProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState)other);
} else {
@@ -875,13 +875,13 @@ public final class ClusterStatusProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -894,7 +894,7 @@ public final class ClusterStatusProtos {
private int bitField0_;
private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo regionInfo_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder> regionInfoBuilder_;
/**
* <code>required .hbase.pb.RegionInfo region_info = 1;</code>
@@ -997,11 +997,11 @@ public final class ClusterStatusProtos {
/**
* <code>required .hbase.pb.RegionInfo region_info = 1;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder>
getRegionInfoFieldBuilder() {
if (regionInfoBuilder_ == null) {
- regionInfoBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ regionInfoBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder>(
getRegionInfo(),
getParentForChildren(),
@@ -1079,12 +1079,12 @@ public final class ClusterStatusProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -1102,22 +1102,22 @@ public final class ClusterStatusProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<RegionState>
- PARSER = new com.google.protobuf.AbstractParser<RegionState>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RegionState>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<RegionState>() {
public RegionState parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new RegionState(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<RegionState> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RegionState> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<RegionState> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RegionState> getParserForType() {
return PARSER;
}
@@ -1129,7 +1129,7 @@ public final class ClusterStatusProtos {
public interface RegionInTransitionOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.RegionInTransition)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required .hbase.pb.RegionSpecifier spec = 1;</code>
@@ -1161,29 +1161,29 @@ public final class ClusterStatusProtos {
* Protobuf type {@code hbase.pb.RegionInTransition}
*/
public static final class RegionInTransition extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.RegionInTransition)
RegionInTransitionOrBuilder {
// Use RegionInTransition.newBuilder() to construct.
- private RegionInTransition(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private RegionInTransition(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private RegionInTransition() {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private RegionInTransition(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -1227,22 +1227,22 @@ public final class ClusterStatusProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.internal_static_hbase_pb_RegionInTransition_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.internal_static_hbase_pb_RegionInTransition_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1318,7 +1318,7 @@ public final class ClusterStatusProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeMessage(1, getSpec());
@@ -1335,11 +1335,11 @@ public final class ClusterStatusProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(1, getSpec());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(2, getRegionState());
}
size += unknownFields.getSerializedSize();
@@ -1394,61 +1394,61 @@ public final class ClusterStatusProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionInTransition parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionInTransition parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionInTransition parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionInTransition parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionInTransition parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionInTransition parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionInTransition parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionInTransition parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionInTransition parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionInTransition parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -1466,7 +1466,7 @@ public final class ClusterStatusProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1474,15 +1474,15 @@ public final class ClusterStatusProtos {
* Protobuf type {@code hbase.pb.RegionInTransition}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.RegionInTransition)
org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionInTransitionOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.internal_static_hbase_pb_RegionInTransition_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.internal_static_hbase_pb_RegionInTransition_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1495,12 +1495,12 @@ public final class ClusterStatusProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getSpecFieldBuilder();
getRegionStateFieldBuilder();
@@ -1523,7 +1523,7 @@ public final class ClusterStatusProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.internal_static_hbase_pb_RegionInTransition_descriptor;
}
@@ -1569,29 +1569,29 @@ public final class ClusterStatusProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionInTransition) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionInTransition)other);
} else {
@@ -1630,13 +1630,13 @@ public final class ClusterStatusProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionInTransition parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionInTransition) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -1649,7 +1649,7 @@ public final class ClusterStatusProtos {
private int bitField0_;
private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier spec_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> specBuilder_;
/**
* <code>required .hbase.pb.RegionSpecifier spec = 1;</code>
@@ -1752,11 +1752,11 @@ public final class ClusterStatusProtos {
/**
* <code>required .hbase.pb.RegionSpecifier spec = 1;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>
getSpecFieldBuilder() {
if (specBuilder_ == null) {
- specBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ specBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>(
getSpec(),
getParentForChildren(),
@@ -1767,7 +1767,7 @@ public final class ClusterStatusProtos {
}
private org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState regionState_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStateOrBuilder> regionStateBuilder_;
/**
* <code>required .hbase.pb.RegionState region_state = 2;</code>
@@ -1870,11 +1870,11 @@ public final class ClusterStatusProtos {
/**
* <code>required .hbase.pb.RegionState region_state = 2;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStateOrBuilder>
getRegionStateFieldBuilder() {
if (regionStateBuilder_ == null) {
- regionStateBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ regionStateBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStateOrBuilder>(
getRegionState(),
getParentForChildren(),
@@ -1884,12 +1884,12 @@ public final class ClusterStatusProtos {
return regionStateBuilder_;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -1907,22 +1907,22 @@ public final class ClusterStatusProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<RegionInTransition>
- PARSER = new com.google.protobuf.AbstractParser<RegionInTransition>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RegionInTransition>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<RegionInTransition>() {
public RegionInTransition parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new RegionInTransition(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<RegionInTransition> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RegionInTransition> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<RegionInTransition> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RegionInTransition> getParserForType() {
return PARSER;
}
@@ -1934,7 +1934,7 @@ public final class ClusterStatusProtos {
public interface StoreSequenceIdOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.StoreSequenceId)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required bytes family_name = 1;</code>
@@ -1943,7 +1943,7 @@ public final class ClusterStatusProtos {
/**
* <code>required bytes family_name = 1;</code>
*/
- com.google.protobuf.ByteString getFamilyName();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamilyName();
/**
* <code>required uint64 sequence_id = 2;</code>
@@ -1963,31 +1963,31 @@ public final class ClusterStatusProtos {
* Protobuf type {@code hbase.pb.StoreSequenceId}
*/
public static final class StoreSequenceId extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.StoreSequenceId)
StoreSequenceIdOrBuilder {
// Use StoreSequenceId.newBuilder() to construct.
- private StoreSequenceId(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private StoreSequenceId(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private StoreSequenceId() {
- familyName_ = com.google.protobuf.ByteString.EMPTY;
+ familyName_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
sequenceId_ = 0L;
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private StoreSequenceId(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -2015,22 +2015,22 @@ public final class ClusterStatusProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.internal_static_hbase_pb_StoreSequenceId_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.internal_static_hbase_pb_StoreSequenceId_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2039,7 +2039,7 @@ public final class ClusterStatusProtos {
private int bitField0_;
public static final int FAMILY_NAME_FIELD_NUMBER = 1;
- private com.google.protobuf.ByteString familyName_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString familyName_;
/**
* <code>required bytes family_name = 1;</code>
*/
@@ -2049,7 +2049,7 @@ public final class ClusterStatusProtos {
/**
* <code>required bytes family_name = 1;</code>
*/
- public com.google.protobuf.ByteString getFamilyName() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamilyName() {
return familyName_;
}
@@ -2086,7 +2086,7 @@ public final class ClusterStatusProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, familyName_);
@@ -2103,11 +2103,11 @@ public final class ClusterStatusProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(1, familyName_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(2, sequenceId_);
}
size += unknownFields.getSerializedSize();
@@ -2154,7 +2154,7 @@ public final class ClusterStatusProtos {
}
if (hasSequenceId()) {
hash = (37 * hash) + SEQUENCE_ID_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getSequenceId());
}
hash = (29 * hash) + unknownFields.hashCode();
@@ -2163,61 +2163,61 @@ public final class ClusterStatusProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -2235,7 +2235,7 @@ public final class ClusterStatusProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -2248,15 +2248,15 @@ public final class ClusterStatusProtos {
* Protobuf type {@code hbase.pb.StoreSequenceId}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.StoreSequenceId)
org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceIdOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.internal_static_hbase_pb_StoreSequenceId_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.internal_static_hbase_pb_StoreSequenceId_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2269,25 +2269,25 @@ public final class ClusterStatusProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
public Builder clear() {
super.clear();
- familyName_ = com.google.protobuf.ByteString.EMPTY;
+ familyName_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000001);
sequenceId_ = 0L;
bitField0_ = (bitField0_ & ~0x00000002);
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.internal_static_hbase_pb_StoreSequenceId_descriptor;
}
@@ -2325,29 +2325,29 @@ public final class ClusterStatusProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId)other);
} else {
@@ -2380,13 +2380,13 @@ public final class ClusterStatusProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -2398,7 +2398,7 @@ public final class ClusterStatusProtos {
}
private int bitField0_;
- private com.google.protobuf.ByteString familyName_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString familyName_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>required bytes family_name = 1;</code>
*/
@@ -2408,13 +2408,13 @@ public final class ClusterStatusProtos {
/**
* <code>required bytes family_name = 1;</code>
*/
- public com.google.protobuf.ByteString getFamilyName() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamilyName() {
return familyName_;
}
/**
* <code>required bytes family_name = 1;</code>
*/
- public Builder setFamilyName(com.google.protobuf.ByteString value) {
+ public Builder setFamilyName(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -2465,12 +2465,12 @@ public final class ClusterStatusProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -2488,22 +2488,22 @@ public final class ClusterStatusProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<StoreSequenceId>
- PARSER = new com.google.protobuf.AbstractParser<StoreSequenceId>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<StoreSequenceId>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<StoreSequenceId>() {
public StoreSequenceId parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new StoreSequenceId(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<StoreSequenceId> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<StoreSequenceId> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<StoreSequenceId> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<StoreSequenceId> getParserForType() {
return PARSER;
}
@@ -2515,7 +2515,7 @@ public final class ClusterStatusProtos {
public interface RegionStoreSequenceIdsOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.RegionStoreSequenceIds)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required uint64 last_flushed_sequence_id = 1;</code>
@@ -2560,11 +2560,11 @@ public final class ClusterStatusProtos {
* Protobuf type {@code hbase.pb.RegionStoreSequenceIds}
*/
public static final class RegionStoreSequenceIds extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.RegionStoreSequenceIds)
RegionStoreSequenceIdsOrBuilder {
// Use RegionStoreSequenceIds.newBuilder() to construct.
- private RegionStoreSequenceIds(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private RegionStoreSequenceIds(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private RegionStoreSequenceIds() {
@@ -2573,18 +2573,18 @@ public final class ClusterStatusProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private RegionStoreSequenceIds(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -2616,10 +2616,10 @@ public final class ClusterStatusProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
@@ -2629,12 +2629,12 @@ public final class ClusterStatusProtos {
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.internal_static_hbase_pb_RegionStoreSequenceIds_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.internal_static_hbase_pb_RegionStoreSequenceIds_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2712,7 +2712,7 @@ public final class ClusterStatusProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeUInt64(1, lastFlushedSequenceId_);
@@ -2729,11 +2729,11 @@ public final class ClusterStatusProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(1, lastFlushedSequenceId_);
}
for (int i = 0; i < storeSequenceId_.size(); i++) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(2, storeSequenceId_.get(i));
}
size += unknownFields.getSerializedSize();
@@ -2773,7 +2773,7 @@ public final class ClusterStatusProtos {
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasLastFlushedSequenceId()) {
hash = (37 * hash) + LAST_FLUSHED_SEQUENCE_ID_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getLastFlushedSequenceId());
}
if (getStoreSequenceIdCount() > 0) {
@@ -2786,61 +2786,61 @@ public final class ClusterStatusProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStoreSequenceIds parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStoreSequenceIds parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStoreSequenceIds parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStoreSequenceIds parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStoreSequenceIds parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStoreSequenceIds parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStoreSequenceIds parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStoreSequenceIds parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStoreSequenceIds parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStoreSequenceIds parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -2858,7 +2858,7 @@ public final class ClusterStatusProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -2872,15 +2872,15 @@ public final class ClusterStatusProtos {
* Protobuf type {@code hbase.pb.RegionStoreSequenceIds}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.RegionStoreSequenceIds)
org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStoreSequenceIdsOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.internal_static_hbase_pb_RegionStoreSequenceIds_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.internal_static_hbase_pb_RegionStoreSequenceIds_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2893,12 +2893,12 @@ public final class ClusterStatusProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getStoreSequenceIdFieldBuilder();
}
@@ -2916,7 +2916,7 @@ public final class ClusterStatusProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.internal_static_hbase_pb_RegionStoreSequenceIds_descriptor;
}
@@ -2959,29 +2959,29 @@ public final class ClusterStatusProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStoreSequenceIds) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStoreSequenceIds)other);
} else {
@@ -3014,7 +3014,7 @@ public final class ClusterStatusProtos {
storeSequenceId_ = other.storeSequenceId_;
bitField0_ = (bitField0_ & ~0x00000002);
storeSequenceIdBuilder_ =
- com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getStoreSequenceIdFieldBuilder() : null;
} else {
storeSequenceIdBuilder_.addAllMessages(other.storeSequenceId_);
@@ -3039,13 +3039,13 @@ public final class ClusterStatusProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStoreSequenceIds parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStoreSequenceIds) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -3098,7 +3098,7 @@ public final class ClusterStatusProtos {
}
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceIdOrBuilder> storeSequenceIdBuilder_;
/**
@@ -3230,7 +3230,7 @@ public final class ClusterStatusProtos {
java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId> values) {
if (storeSequenceIdBuilder_ == null) {
ensureStoreSequenceIdIsMutable();
- com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
values, storeSequenceId_);
onChanged();
} else {
@@ -3314,11 +3314,11 @@ public final class ClusterStatusProtos {
getStoreSequenceIdBuilderList() {
return getStoreSequenceIdFieldBuilder().getBuilderList();
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceIdOrBuilder>
getStoreSequenceIdFieldBuilder() {
if (storeSequenceIdBuilder_ == null) {
- storeSequenceIdBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3<
+ storeSequenceIdBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceIdOrBuilder>(
storeSequenceId_,
((bitField0_ & 0x00000002) == 0x00000002),
@@ -3329,12 +3329,12 @@ public final class ClusterStatusProtos {
return storeSequenceIdBuilder_;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -3352,22 +3352,22 @@ public final class ClusterStatusProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<RegionStoreSequenceIds>
- PARSER = new com.google.protobuf.AbstractParser<RegionStoreSequenceIds>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RegionStoreSequenceIds>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<RegionStoreSequenceIds>() {
public RegionStoreSequenceIds parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new RegionStoreSequenceIds(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<RegionStoreSequenceIds> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RegionStoreSequenceIds> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<RegionStoreSequenceIds> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RegionStoreSequenceIds> getParserForType() {
re
<TRUNCATED>
[10/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ProcedureProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ProcedureProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ProcedureProtos.java
index 40e4611..91989d7 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ProcedureProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ProcedureProtos.java
@@ -6,19 +6,19 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class ProcedureProtos {
private ProcedureProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
/**
* Protobuf enum {@code hbase.pb.ProcedureState}
*/
public enum ProcedureState
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <pre>
* Procedure in construction, not yet added to the executor
@@ -143,27 +143,27 @@ public final class ProcedureProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<ProcedureState>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<ProcedureState>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
ProcedureState> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<ProcedureState>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<ProcedureState>() {
public ProcedureState findValueByNumber(int number) {
return ProcedureState.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.getDescriptor().getEnumTypes().get(0);
}
@@ -171,7 +171,7 @@ public final class ProcedureProtos {
private static final ProcedureState[] VALUES = values();
public static ProcedureState valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -190,7 +190,7 @@ public final class ProcedureProtos {
public interface ProcedureOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.Procedure)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <pre>
@@ -215,7 +215,7 @@ public final class ProcedureProtos {
*
* <code>required string class_name = 1;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getClassNameBytes();
/**
@@ -264,7 +264,7 @@ public final class ProcedureProtos {
/**
* <code>optional string owner = 5;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getOwnerBytes();
/**
@@ -367,7 +367,7 @@ public final class ProcedureProtos {
*
* <code>optional bytes result = 11;</code>
*/
- com.google.protobuf.ByteString getResult();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getResult();
/**
* <pre>
@@ -384,7 +384,7 @@ public final class ProcedureProtos {
*
* <code>optional bytes state_data = 12;</code>
*/
- com.google.protobuf.ByteString getStateData();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getStateData();
/**
* <pre>
@@ -421,11 +421,11 @@ public final class ProcedureProtos {
* Protobuf type {@code hbase.pb.Procedure}
*/
public static final class Procedure extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.Procedure)
ProcedureOrBuilder {
// Use Procedure.newBuilder() to construct.
- private Procedure(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private Procedure(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private Procedure() {
@@ -438,25 +438,25 @@ public final class ProcedureProtos {
stackId_ = java.util.Collections.emptyList();
lastUpdate_ = 0L;
timeout_ = 0;
- result_ = com.google.protobuf.ByteString.EMPTY;
- stateData_ = com.google.protobuf.ByteString.EMPTY;
+ result_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
+ stateData_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
nonceGroup_ = 0L;
nonce_ = 0L;
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private Procedure(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -473,7 +473,7 @@ public final class ProcedureProtos {
break;
}
case 10: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000001;
className_ = bs;
break;
@@ -494,7 +494,7 @@ public final class ProcedureProtos {
break;
}
case 42: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000010;
owner_ = bs;
break;
@@ -576,10 +576,10 @@ public final class ProcedureProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000040) == 0x00000040)) {
@@ -589,12 +589,12 @@ public final class ProcedureProtos {
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.internal_static_hbase_pb_Procedure_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.internal_static_hbase_pb_Procedure_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -626,8 +626,8 @@ public final class ProcedureProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
className_ = s;
@@ -642,17 +642,17 @@ public final class ProcedureProtos {
*
* <code>required string class_name = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getClassNameBytes() {
java.lang.Object ref = className_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
className_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -725,8 +725,8 @@ public final class ProcedureProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
owner_ = s;
@@ -737,17 +737,17 @@ public final class ProcedureProtos {
/**
* <code>optional string owner = 5;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getOwnerBytes() {
java.lang.Object ref = owner_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
owner_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -873,7 +873,7 @@ public final class ProcedureProtos {
}
public static final int RESULT_FIELD_NUMBER = 11;
- private com.google.protobuf.ByteString result_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString result_;
/**
* <pre>
* opaque (user) result structure
@@ -891,12 +891,12 @@ public final class ProcedureProtos {
*
* <code>optional bytes result = 11;</code>
*/
- public com.google.protobuf.ByteString getResult() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getResult() {
return result_;
}
public static final int STATE_DATA_FIELD_NUMBER = 12;
- private com.google.protobuf.ByteString stateData_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString stateData_;
/**
* <pre>
* opaque (user) procedure internal-state
@@ -914,7 +914,7 @@ public final class ProcedureProtos {
*
* <code>optional bytes state_data = 12;</code>
*/
- public com.google.protobuf.ByteString getStateData() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getStateData() {
return stateData_;
}
@@ -986,10 +986,10 @@ public final class ProcedureProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 1, className_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, className_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
output.writeUInt64(2, parentId_);
@@ -1001,7 +1001,7 @@ public final class ProcedureProtos {
output.writeUInt64(4, startTime_);
}
if (((bitField0_ & 0x00000010) == 0x00000010)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 5, owner_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 5, owner_);
}
if (((bitField0_ & 0x00000020) == 0x00000020)) {
output.writeEnum(6, state_);
@@ -1039,62 +1039,62 @@ public final class ProcedureProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, className_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, className_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(2, parentId_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(3, procId_);
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(4, startTime_);
}
if (((bitField0_ & 0x00000010) == 0x00000010)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(5, owner_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(5, owner_);
}
if (((bitField0_ & 0x00000020) == 0x00000020)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeEnumSize(6, state_);
}
{
int dataSize = 0;
for (int i = 0; i < stackId_.size(); i++) {
- dataSize += com.google.protobuf.CodedOutputStream
+ dataSize += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt32SizeNoTag(stackId_.get(i));
}
size += dataSize;
size += 1 * getStackIdList().size();
}
if (((bitField0_ & 0x00000040) == 0x00000040)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(8, lastUpdate_);
}
if (((bitField0_ & 0x00000080) == 0x00000080)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt32Size(9, timeout_);
}
if (((bitField0_ & 0x00000100) == 0x00000100)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(10, getException());
}
if (((bitField0_ & 0x00000200) == 0x00000200)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(11, result_);
}
if (((bitField0_ & 0x00000400) == 0x00000400)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(12, stateData_);
}
if (((bitField0_ & 0x00000800) == 0x00000800)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(13, nonceGroup_);
}
if (((bitField0_ & 0x00001000) == 0x00001000)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(14, nonce_);
}
size += unknownFields.getSerializedSize();
@@ -1197,17 +1197,17 @@ public final class ProcedureProtos {
}
if (hasParentId()) {
hash = (37 * hash) + PARENT_ID_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getParentId());
}
if (hasProcId()) {
hash = (37 * hash) + PROC_ID_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getProcId());
}
if (hasStartTime()) {
hash = (37 * hash) + START_TIME_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getStartTime());
}
if (hasOwner()) {
@@ -1224,7 +1224,7 @@ public final class ProcedureProtos {
}
if (hasLastUpdate()) {
hash = (37 * hash) + LAST_UPDATE_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getLastUpdate());
}
if (hasTimeout()) {
@@ -1245,12 +1245,12 @@ public final class ProcedureProtos {
}
if (hasNonceGroup()) {
hash = (37 * hash) + NONCE_GROUP_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getNonceGroup());
}
if (hasNonce()) {
hash = (37 * hash) + NONCE_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getNonce());
}
hash = (29 * hash) + unknownFields.hashCode();
@@ -1259,61 +1259,61 @@ public final class ProcedureProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -1331,7 +1331,7 @@ public final class ProcedureProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1344,15 +1344,15 @@ public final class ProcedureProtos {
* Protobuf type {@code hbase.pb.Procedure}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.Procedure)
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.internal_static_hbase_pb_Procedure_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.internal_static_hbase_pb_Procedure_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1365,12 +1365,12 @@ public final class ProcedureProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getExceptionFieldBuilder();
}
@@ -1401,9 +1401,9 @@ public final class ProcedureProtos {
exceptionBuilder_.clear();
}
bitField0_ = (bitField0_ & ~0x00000200);
- result_ = com.google.protobuf.ByteString.EMPTY;
+ result_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000400);
- stateData_ = com.google.protobuf.ByteString.EMPTY;
+ stateData_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000800);
nonceGroup_ = 0L;
bitField0_ = (bitField0_ & ~0x00001000);
@@ -1412,7 +1412,7 @@ public final class ProcedureProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.internal_static_hbase_pb_Procedure_descriptor;
}
@@ -1503,29 +1503,29 @@ public final class ProcedureProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure)other);
} else {
@@ -1614,13 +1614,13 @@ public final class ProcedureProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -1653,8 +1653,8 @@ public final class ProcedureProtos {
public java.lang.String getClassName() {
java.lang.Object ref = className_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
className_ = s;
@@ -1671,17 +1671,17 @@ public final class ProcedureProtos {
*
* <code>required string class_name = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getClassNameBytes() {
java.lang.Object ref = className_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
className_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -1722,7 +1722,7 @@ public final class ProcedureProtos {
* <code>required string class_name = 1;</code>
*/
public Builder setClassNameBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1857,8 +1857,8 @@ public final class ProcedureProtos {
public java.lang.String getOwner() {
java.lang.Object ref = owner_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
owner_ = s;
@@ -1871,17 +1871,17 @@ public final class ProcedureProtos {
/**
* <code>optional string owner = 5;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getOwnerBytes() {
java.lang.Object ref = owner_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
owner_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -1910,7 +1910,7 @@ public final class ProcedureProtos {
* <code>optional string owner = 5;</code>
*/
public Builder setOwnerBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -2047,7 +2047,7 @@ public final class ProcedureProtos {
public Builder addAllStackId(
java.lang.Iterable<? extends java.lang.Integer> values) {
ensureStackIdIsMutable();
- com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
values, stackId_);
onChanged();
return this;
@@ -2131,7 +2131,7 @@ public final class ProcedureProtos {
}
private org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage exception_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage, org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessageOrBuilder> exceptionBuilder_;
/**
* <pre>
@@ -2270,11 +2270,11 @@ public final class ProcedureProtos {
*
* <code>optional .hbase.pb.ForeignExceptionMessage exception = 10;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage, org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessageOrBuilder>
getExceptionFieldBuilder() {
if (exceptionBuilder_ == null) {
- exceptionBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ exceptionBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage, org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessageOrBuilder>(
getException(),
getParentForChildren(),
@@ -2284,7 +2284,7 @@ public final class ProcedureProtos {
return exceptionBuilder_;
}
- private com.google.protobuf.ByteString result_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString result_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <pre>
* opaque (user) result structure
@@ -2302,7 +2302,7 @@ public final class ProcedureProtos {
*
* <code>optional bytes result = 11;</code>
*/
- public com.google.protobuf.ByteString getResult() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getResult() {
return result_;
}
/**
@@ -2312,7 +2312,7 @@ public final class ProcedureProtos {
*
* <code>optional bytes result = 11;</code>
*/
- public Builder setResult(com.google.protobuf.ByteString value) {
+ public Builder setResult(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -2335,7 +2335,7 @@ public final class ProcedureProtos {
return this;
}
- private com.google.protobuf.ByteString stateData_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString stateData_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <pre>
* opaque (user) procedure internal-state
@@ -2353,7 +2353,7 @@ public final class ProcedureProtos {
*
* <code>optional bytes state_data = 12;</code>
*/
- public com.google.protobuf.ByteString getStateData() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getStateData() {
return stateData_;
}
/**
@@ -2363,7 +2363,7 @@ public final class ProcedureProtos {
*
* <code>optional bytes state_data = 12;</code>
*/
- public Builder setStateData(com.google.protobuf.ByteString value) {
+ public Builder setStateData(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -2466,12 +2466,12 @@ public final class ProcedureProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -2489,22 +2489,22 @@ public final class ProcedureProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<Procedure>
- PARSER = new com.google.protobuf.AbstractParser<Procedure>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Procedure>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Procedure>() {
public Procedure parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new Procedure(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<Procedure> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Procedure> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<Procedure> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Procedure> getParserForType() {
return PARSER;
}
@@ -2516,7 +2516,7 @@ public final class ProcedureProtos {
public interface SequentialProcedureDataOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.SequentialProcedureData)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required bool executed = 1;</code>
@@ -2536,11 +2536,11 @@ public final class ProcedureProtos {
* Protobuf type {@code hbase.pb.SequentialProcedureData}
*/
public static final class SequentialProcedureData extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.SequentialProcedureData)
SequentialProcedureDataOrBuilder {
// Use SequentialProcedureData.newBuilder() to construct.
- private SequentialProcedureData(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private SequentialProcedureData(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private SequentialProcedureData() {
@@ -2548,18 +2548,18 @@ public final class ProcedureProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private SequentialProcedureData(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -2582,22 +2582,22 @@ public final class ProcedureProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.internal_static_hbase_pb_SequentialProcedureData_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.internal_static_hbase_pb_SequentialProcedureData_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2634,7 +2634,7 @@ public final class ProcedureProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBool(1, executed_);
@@ -2648,7 +2648,7 @@ public final class ProcedureProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBoolSize(1, executed_);
}
size += unknownFields.getSerializedSize();
@@ -2686,7 +2686,7 @@ public final class ProcedureProtos {
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasExecuted()) {
hash = (37 * hash) + EXECUTED_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean(
getExecuted());
}
hash = (29 * hash) + unknownFields.hashCode();
@@ -2695,61 +2695,61 @@ public final class ProcedureProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.SequentialProcedureData parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.SequentialProcedureData parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.SequentialProcedureData parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.SequentialProcedureData parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.SequentialProcedureData parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.SequentialProcedureData parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.SequentialProcedureData parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.SequentialProcedureData parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.SequentialProcedureData parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.SequentialProcedureData parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -2767,7 +2767,7 @@ public final class ProcedureProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -2780,15 +2780,15 @@ public final class ProcedureProtos {
* Protobuf type {@code hbase.pb.SequentialProcedureData}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.SequentialProcedureData)
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.SequentialProcedureDataOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.internal_static_hbase_pb_SequentialProcedureData_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.internal_static_hbase_pb_SequentialProcedureData_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2801,12 +2801,12 @@ public final class ProcedureProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -2817,7 +2817,7 @@ public final class ProcedureProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.internal_static_hbase_pb_SequentialProcedureData_descriptor;
}
@@ -2851,29 +2851,29 @@ public final class ProcedureProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.SequentialProcedureData) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.SequentialProcedureData)other);
} else {
@@ -2900,13 +2900,13 @@ public final class ProcedureProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.SequentialProcedureData parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.SequentialProcedureData) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -2950,12 +2950,12 @@ public final class ProcedureProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -2973,22 +2973,22 @@ public final class ProcedureProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<SequentialProcedureData>
- PARSER = new com.google.protobuf.AbstractParser<SequentialProcedureData>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<SequentialProcedureData>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<SequentialProcedureData>() {
public SequentialProcedureData parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new SequentialProcedureData(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<SequentialProcedureData> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<SequentialProcedureData> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<SequentialProcedureData> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<SequentialProcedureData> getParserForType() {
return PARSER;
}
@@ -3000,7 +3000,7 @@ public final class ProcedureProtos {
public interface StateMachineProcedureDataOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.StateMachineProcedureData)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>repeated uint32 state = 1;</code>
@@ -3024,11 +3024,11 @@ public final class ProcedureProtos {
* Protobuf type {@code hbase.pb.StateMachineProcedureData}
*/
public static final class StateMachineProcedureData extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.StateMachineProcedureData)
StateMachineProcedureDataOrBuilder {
// Use StateMachineProcedureData.newBuilder() to construct.
- private StateMachineProcedureData(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private StateMachineProcedureData(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private StateMachineProcedureData() {
@@ -3036,18 +3036,18 @@ public final class ProcedureProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private StateMachineProcedureData(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -3086,10 +3086,10 @@ public final class ProcedureProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
@@ -3099,12 +3099,12 @@ public final class ProcedureProtos {
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.internal_static_hbase_pb_StateMachineProcedureData_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.internal_static_hbase_pb_StateMachineProcedureData_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -3143,7 +3143,7 @@ public final class ProcedureProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
for (int i = 0; i < state_.size(); i++) {
output.writeUInt32(1, state_.get(i));
@@ -3159,7 +3159,7 @@ public final class ProcedureProtos {
{
int dataSize = 0;
for (int i = 0; i < state_.size(); i++) {
- dataSize += com.google.protobuf.CodedOutputStream
+ dataSize += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt32SizeNoTag(state_.get(i));
}
size += dataSize;
@@ -3205,61 +3205,61 @@ public final class ProcedureProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.StateMachineProcedureData parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.StateMachineProcedureData parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.StateMachineProcedureData parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.StateMachineProcedureData parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.StateMachineProcedureData parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.StateMachineProcedureData parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.StateMachineProcedureData parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.StateMachineProcedureData parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.StateMachineProcedureData parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.StateMachineProcedureData parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -3277,7 +3277,7 @@ public final class ProcedureProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -3290,15 +3290,15 @@ public final class ProcedureProtos {
* Protobuf type {@code hbase.pb.StateMachineProcedureData}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.StateMachineProcedureData)
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.StateMachineProcedureDataOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.internal_static_hbase_pb_StateMachineProcedureData_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.internal_static_hbase_pb_StateMachineProcedureData_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -3311,12 +3311,12 @@ public final class ProcedureProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -3327,7 +3327,7 @@ public final class ProcedureProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.internal_static_hbase_pb_StateMachineProcedureData_descriptor;
}
@@ -3360,29 +3360,29 @@ public final class ProcedureProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.StateMachineProcedureData) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.StateMachineProcedureData)other);
} else {
@@ -3413,13 +3413,13 @@ public final class ProcedureProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.StateMachineProcedureData parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.StateMachineProcedureData) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -3482,7 +3482,7 @@ public final class ProcedureProtos {
public Builder addAllState(
java.lang.Iterable<? extends java.lang.Integer> values) {
ensureStateIsMutable();
- com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
values, state_);
onChanged();
return this;
@@ -3497,12 +3497,12 @@ public final class ProcedureProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -3520,22 +3520,22 @@ public final class ProcedureProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<StateMachineProcedureData>
- PARSER = new com.google.protobuf.AbstractParser<StateMachineProcedureData>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<StateMachineProcedureData>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<StateMachineProcedureData>() {
public StateMachineProcedureData parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new StateMachineProcedureData(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<StateMachineProcedureData> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<StateMachineProcedureData> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<StateMachineProcedureData> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<StateMachineProcedureData> getParserForType() {
return PARSER;
}
@@ -3547,7 +3547,7 @@ public final class ProcedureProtos {
public interface ProcedureWALHeaderOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.ProcedureWALHeader)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required uint32 version = 1;</code>
@@ -3594,11 +3594,11 @@ public final class ProcedureProtos {
* Protobuf type {@code hbase.pb.ProcedureWALHeader}
*/
public static final class ProcedureWALHeader extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.ProcedureWALHeader)
ProcedureWALHeaderOrBuilder {
// Use ProcedureWALHeader.newBuilder() to construct.
- private ProcedureWALHeader(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private ProcedureWALHeader(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private ProcedureWALHeader() {
@@ -3609,18 +3609,18 @@ public final class ProcedureProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private ProcedureWALHeader(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -3658,22 +3658,22 @@ public final class ProcedureProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.internal_static_hbase_pb_ProcedureWALHeader_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.internal_static_hbase_pb_ProcedureWALHeader_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -3767,7 +3767,7 @@ public final class ProcedureProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeUInt32(1, version_);
@@ -3790,19 +3790,19 @@ public final class ProcedureProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt32Size(1, version_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt32Size(2, type_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(3, logId_);
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(4, minProcId_);
}
size += unknownFields.getSerializedSize();
@@ -3863,12 +3863,12 @@ public final class ProcedureProtos {
}
if (hasLogId()) {
hash = (37 * hash) + LOG_ID_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getLogId());
}
if (hasMinProcId()) {
hash = (37 * hash) + MIN_PROC_ID_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getMinProcId());
}
hash = (29 * hash) + unknownFields.hashCode();
@@ -3877,61 +3877,61 @@ public final class ProcedureProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureWALHeader parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureWALHeader parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureWALHeader parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureWALHeader parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureWALHeader parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureWALHeader parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureWALHeader parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+
<TRUNCATED>
[20/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ComparatorProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ComparatorProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ComparatorProtos.java
index e1c605d..1c6c373 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ComparatorProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ComparatorProtos.java
@@ -6,17 +6,17 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class ComparatorProtos {
private ComparatorProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
public interface ComparatorOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.Comparator)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required string name = 1;</code>
@@ -29,7 +29,7 @@ public final class ComparatorProtos {
/**
* <code>required string name = 1;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getNameBytes();
/**
@@ -39,37 +39,37 @@ public final class ComparatorProtos {
/**
* <code>optional bytes serialized_comparator = 2;</code>
*/
- com.google.protobuf.ByteString getSerializedComparator();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getSerializedComparator();
}
/**
* Protobuf type {@code hbase.pb.Comparator}
*/
public static final class Comparator extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.Comparator)
ComparatorOrBuilder {
// Use Comparator.newBuilder() to construct.
- private Comparator(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private Comparator(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private Comparator() {
name_ = "";
- serializedComparator_ = com.google.protobuf.ByteString.EMPTY;
+ serializedComparator_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private Comparator(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -86,7 +86,7 @@ public final class ComparatorProtos {
break;
}
case 10: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000001;
name_ = bs;
break;
@@ -98,22 +98,22 @@ public final class ComparatorProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.internal_static_hbase_pb_Comparator_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.internal_static_hbase_pb_Comparator_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -137,8 +137,8 @@ public final class ComparatorProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
name_ = s;
@@ -149,22 +149,22 @@ public final class ComparatorProtos {
/**
* <code>required string name = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getNameBytes() {
java.lang.Object ref = name_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
name_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
public static final int SERIALIZED_COMPARATOR_FIELD_NUMBER = 2;
- private com.google.protobuf.ByteString serializedComparator_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString serializedComparator_;
/**
* <code>optional bytes serialized_comparator = 2;</code>
*/
@@ -174,7 +174,7 @@ public final class ComparatorProtos {
/**
* <code>optional bytes serialized_comparator = 2;</code>
*/
- public com.google.protobuf.ByteString getSerializedComparator() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getSerializedComparator() {
return serializedComparator_;
}
@@ -192,10 +192,10 @@ public final class ComparatorProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 1, name_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, name_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
output.writeBytes(2, serializedComparator_);
@@ -209,10 +209,10 @@ public final class ComparatorProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, name_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, name_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(2, serializedComparator_);
}
size += unknownFields.getSerializedSize();
@@ -267,61 +267,61 @@ public final class ComparatorProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.Comparator parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.Comparator parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.Comparator parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.Comparator parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.Comparator parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.Comparator parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.Comparator parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.Comparator parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.Comparator parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.Comparator parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -339,7 +339,7 @@ public final class ComparatorProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -347,15 +347,15 @@ public final class ComparatorProtos {
* Protobuf type {@code hbase.pb.Comparator}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.Comparator)
org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ComparatorOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.internal_static_hbase_pb_Comparator_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.internal_static_hbase_pb_Comparator_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -368,12 +368,12 @@ public final class ComparatorProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -381,12 +381,12 @@ public final class ComparatorProtos {
super.clear();
name_ = "";
bitField0_ = (bitField0_ & ~0x00000001);
- serializedComparator_ = com.google.protobuf.ByteString.EMPTY;
+ serializedComparator_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000002);
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.internal_static_hbase_pb_Comparator_descriptor;
}
@@ -424,29 +424,29 @@ public final class ComparatorProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.Comparator) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.Comparator)other);
} else {
@@ -478,13 +478,13 @@ public final class ComparatorProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.Comparator parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.Comparator) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -509,8 +509,8 @@ public final class ComparatorProtos {
public java.lang.String getName() {
java.lang.Object ref = name_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
name_ = s;
@@ -523,17 +523,17 @@ public final class ComparatorProtos {
/**
* <code>required string name = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getNameBytes() {
java.lang.Object ref = name_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
name_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -562,7 +562,7 @@ public final class ComparatorProtos {
* <code>required string name = 1;</code>
*/
public Builder setNameBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -572,7 +572,7 @@ public final class ComparatorProtos {
return this;
}
- private com.google.protobuf.ByteString serializedComparator_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString serializedComparator_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>optional bytes serialized_comparator = 2;</code>
*/
@@ -582,13 +582,13 @@ public final class ComparatorProtos {
/**
* <code>optional bytes serialized_comparator = 2;</code>
*/
- public com.google.protobuf.ByteString getSerializedComparator() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getSerializedComparator() {
return serializedComparator_;
}
/**
* <code>optional bytes serialized_comparator = 2;</code>
*/
- public Builder setSerializedComparator(com.google.protobuf.ByteString value) {
+ public Builder setSerializedComparator(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -607,12 +607,12 @@ public final class ComparatorProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -630,22 +630,22 @@ public final class ComparatorProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<Comparator>
- PARSER = new com.google.protobuf.AbstractParser<Comparator>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Comparator>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Comparator>() {
public Comparator parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new Comparator(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<Comparator> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Comparator> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<Comparator> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Comparator> getParserForType() {
return PARSER;
}
@@ -657,7 +657,7 @@ public final class ComparatorProtos {
public interface ByteArrayComparableOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.ByteArrayComparable)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>optional bytes value = 1;</code>
@@ -666,36 +666,36 @@ public final class ComparatorProtos {
/**
* <code>optional bytes value = 1;</code>
*/
- com.google.protobuf.ByteString getValue();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getValue();
}
/**
* Protobuf type {@code hbase.pb.ByteArrayComparable}
*/
public static final class ByteArrayComparable extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.ByteArrayComparable)
ByteArrayComparableOrBuilder {
// Use ByteArrayComparable.newBuilder() to construct.
- private ByteArrayComparable(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private ByteArrayComparable(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private ByteArrayComparable() {
- value_ = com.google.protobuf.ByteString.EMPTY;
+ value_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private ByteArrayComparable(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -718,22 +718,22 @@ public final class ComparatorProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.internal_static_hbase_pb_ByteArrayComparable_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.internal_static_hbase_pb_ByteArrayComparable_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -742,7 +742,7 @@ public final class ComparatorProtos {
private int bitField0_;
public static final int VALUE_FIELD_NUMBER = 1;
- private com.google.protobuf.ByteString value_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value_;
/**
* <code>optional bytes value = 1;</code>
*/
@@ -752,7 +752,7 @@ public final class ComparatorProtos {
/**
* <code>optional bytes value = 1;</code>
*/
- public com.google.protobuf.ByteString getValue() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getValue() {
return value_;
}
@@ -766,7 +766,7 @@ public final class ComparatorProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, value_);
@@ -780,7 +780,7 @@ public final class ComparatorProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(1, value_);
}
size += unknownFields.getSerializedSize();
@@ -826,61 +826,61 @@ public final class ComparatorProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -898,7 +898,7 @@ public final class ComparatorProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -906,15 +906,15 @@ public final class ComparatorProtos {
* Protobuf type {@code hbase.pb.ByteArrayComparable}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.ByteArrayComparable)
org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparableOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.internal_static_hbase_pb_ByteArrayComparable_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.internal_static_hbase_pb_ByteArrayComparable_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -927,23 +927,23 @@ public final class ComparatorProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
public Builder clear() {
super.clear();
- value_ = com.google.protobuf.ByteString.EMPTY;
+ value_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000001);
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.internal_static_hbase_pb_ByteArrayComparable_descriptor;
}
@@ -977,29 +977,29 @@ public final class ComparatorProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable)other);
} else {
@@ -1023,13 +1023,13 @@ public final class ComparatorProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -1041,7 +1041,7 @@ public final class ComparatorProtos {
}
private int bitField0_;
- private com.google.protobuf.ByteString value_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>optional bytes value = 1;</code>
*/
@@ -1051,13 +1051,13 @@ public final class ComparatorProtos {
/**
* <code>optional bytes value = 1;</code>
*/
- public com.google.protobuf.ByteString getValue() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getValue() {
return value_;
}
/**
* <code>optional bytes value = 1;</code>
*/
- public Builder setValue(com.google.protobuf.ByteString value) {
+ public Builder setValue(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1076,12 +1076,12 @@ public final class ComparatorProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -1099,22 +1099,22 @@ public final class ComparatorProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<ByteArrayComparable>
- PARSER = new com.google.protobuf.AbstractParser<ByteArrayComparable>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ByteArrayComparable>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<ByteArrayComparable>() {
public ByteArrayComparable parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new ByteArrayComparable(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<ByteArrayComparable> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ByteArrayComparable> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<ByteArrayComparable> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ByteArrayComparable> getParserForType() {
return PARSER;
}
@@ -1126,7 +1126,7 @@ public final class ComparatorProtos {
public interface BinaryComparatorOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.BinaryComparator)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required .hbase.pb.ByteArrayComparable comparable = 1;</code>
@@ -1145,29 +1145,29 @@ public final class ComparatorProtos {
* Protobuf type {@code hbase.pb.BinaryComparator}
*/
public static final class BinaryComparator extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.BinaryComparator)
BinaryComparatorOrBuilder {
// Use BinaryComparator.newBuilder() to construct.
- private BinaryComparator(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private BinaryComparator(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private BinaryComparator() {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private BinaryComparator(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -1198,22 +1198,22 @@ public final class ComparatorProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.internal_static_hbase_pb_BinaryComparator_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.internal_static_hbase_pb_BinaryComparator_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1256,7 +1256,7 @@ public final class ComparatorProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeMessage(1, getComparable());
@@ -1270,7 +1270,7 @@ public final class ComparatorProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(1, getComparable());
}
size += unknownFields.getSerializedSize();
@@ -1316,61 +1316,61 @@ public final class ComparatorProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.BinaryComparator parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.BinaryComparator parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.BinaryComparator parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.BinaryComparator parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.BinaryComparator parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.BinaryComparator parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.BinaryComparator parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.BinaryComparator parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.BinaryComparator parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.BinaryComparator parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -1388,7 +1388,7 @@ public final class ComparatorProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1396,15 +1396,15 @@ public final class ComparatorProtos {
* Protobuf type {@code hbase.pb.BinaryComparator}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.BinaryComparator)
org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.BinaryComparatorOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.internal_static_hbase_pb_BinaryComparator_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.internal_static_hbase_pb_BinaryComparator_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1417,12 +1417,12 @@ public final class ComparatorProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getComparableFieldBuilder();
}
@@ -1438,7 +1438,7 @@ public final class ComparatorProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.internal_static_hbase_pb_BinaryComparator_descriptor;
}
@@ -1476,29 +1476,29 @@ public final class ComparatorProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.BinaryComparator) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.BinaryComparator)other);
} else {
@@ -1525,13 +1525,13 @@ public final class ComparatorProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.BinaryComparator parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.BinaryComparator) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -1544,7 +1544,7 @@ public final class ComparatorProtos {
private int bitField0_;
private org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable comparable_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable, org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparableOrBuilder> comparableBuilder_;
/**
* <code>required .hbase.pb.ByteArrayComparable comparable = 1;</code>
@@ -1647,11 +1647,11 @@ public final class ComparatorProtos {
/**
* <code>required .hbase.pb.ByteArrayComparable comparable = 1;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable, org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparableOrBuilder>
getComparableFieldBuilder() {
if (comparableBuilder_ == null) {
- comparableBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ comparableBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable, org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparableOrBuilder>(
getComparable(),
getParentForChildren(),
@@ -1661,12 +1661,12 @@ public final class ComparatorProtos {
return comparableBuilder_;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -1684,22 +1684,22 @@ public final class ComparatorProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<BinaryComparator>
- PARSER = new com.google.protobuf.AbstractParser<BinaryComparator>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<BinaryComparator>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<BinaryComparator>() {
public BinaryComparator parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new BinaryComparator(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<BinaryComparator> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<BinaryComparator> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<BinaryComparator> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<BinaryComparator> getParserForType() {
return PARSER;
}
@@ -1711,7 +1711,7 @@ public final class ComparatorProtos {
public interface LongComparatorOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.LongComparator)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required .hbase.pb.ByteArrayComparable comparable = 1;</code>
@@ -1730,29 +1730,29 @@ public final class ComparatorProtos {
* Protobuf type {@code hbase.pb.LongComparator}
*/
public static final class LongComparator extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.LongComparator)
LongComparatorOrBuilder {
// Use LongComparator.newBuilder() to construct.
- private LongComparator(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private LongComparator(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private LongComparator() {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private LongComparator(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -1783,22 +1783,22 @@ public final class ComparatorProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.internal_static_hbase_pb_LongComparator_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.internal_static_hbase_pb_LongComparator_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1841,7 +1841,7 @@ public final class ComparatorProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeMessage(1, getComparable());
@@ -1855,7 +1855,7 @@ public final class ComparatorProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(1, getComparable());
}
size += unknownFields.getSerializedSize();
@@ -1901,61 +1901,61 @@ public final class ComparatorProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.LongComparator parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.LongComparator parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.LongComparator parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.LongComparator parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.LongComparator parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.LongComparator parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.LongComparator parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.LongComparator parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.LongComparator parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.LongComparator parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -1973,7 +1973,7 @@ public final class ComparatorProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1981,15 +1981,15 @@ public final class ComparatorProtos {
* Protobuf type {@code hbase.pb.LongComparator}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.LongComparator)
org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.LongComparatorOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.internal_static_hbase_pb_LongComparator_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.internal_static_hbase_pb_LongComparator_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2002,12 +2002,12 @@ public final class ComparatorProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getComparableFieldBuilder();
}
@@ -2023,7 +2023,7 @@ public final class ComparatorProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.internal_static_hbase_pb_LongComparator_descriptor;
}
@@ -2061,29 +2061,29 @@ public final class ComparatorProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.LongComparator) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.LongComparator)other);
} else {
@@ -2110,13 +2110,13 @@ public final class ComparatorProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.LongComparator parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.LongComparator) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -2129,7 +2129,7 @@ public final class ComparatorProtos {
private int bitField0_;
private org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable comparable_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable, org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparableOrBuilder> comparableBuilder_;
/**
* <code>required .hbase.pb.ByteArrayComparable comparable = 1;</code>
@@ -2232,11 +2232,11 @@ public final class ComparatorProtos {
/**
* <code>required .hbase.pb.ByteArrayComparable comparable = 1;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable, org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparableOrBuilder>
getComparableFieldBuilder() {
if (comparableBuilder_ == null) {
- comparableBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ comparableBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable, org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparable.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.ByteArrayComparableOrBuilder>(
getComparable(),
getParentForChildren(),
@@ -2246,12 +2246,12 @@ public final class ComparatorProtos {
return comparableBuilder_;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -2269,22 +2269,22 @@ public final class ComparatorProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<LongComparator>
- PARSER = new com.google.protobuf.AbstractParser<LongComparator>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<LongComparator>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<LongComparator>() {
public LongComparator parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new LongComparator(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<LongComparator> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<LongComparator> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<LongComparator> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<LongComparator> getParserForType() {
return PARSER;
}
@@ -2296,7 +2296,7 @@ public final class ComparatorProtos {
public interface BinaryPrefixComparatorOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.BinaryPrefixComparator)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required .hbase.pb.ByteArrayComparable comparable = 1;</code>
@@ -2315,29 +2315,29 @@ public final class ComparatorProtos {
* Protobuf type {@code hbase.pb.BinaryPrefixComparator}
*/
public static final class BinaryPrefixComparator extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.BinaryPrefixComparator)
BinaryPrefixComparatorOrBuilder {
// Use BinaryPrefixComparator.newBuilder() to construct.
- private BinaryPrefixComparator(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private BinaryPrefixComparator(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private BinaryPrefixComparator() {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private BinaryPrefixComparator(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -2368,22 +2368,22 @@ public final class ComparatorProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.internal_static_hbase_pb_BinaryPrefixComparator_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ComparatorProtos.internal_static_hbase_pb_BinaryPrefixComparator_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2426,7 +2426,7 @@ public final class ComparatorProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.i
<TRUNCATED>
[25/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AdminProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AdminProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AdminProtos.java
index e15dbc4..20020d4 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AdminProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AdminProtos.java
@@ -6,17 +6,17 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class AdminProtos {
private AdminProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
public interface GetRegionInfoRequestOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.GetRegionInfoRequest)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required .hbase.pb.RegionSpecifier region = 1;</code>
@@ -44,11 +44,11 @@ public final class AdminProtos {
* Protobuf type {@code hbase.pb.GetRegionInfoRequest}
*/
public static final class GetRegionInfoRequest extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.GetRegionInfoRequest)
GetRegionInfoRequestOrBuilder {
// Use GetRegionInfoRequest.newBuilder() to construct.
- private GetRegionInfoRequest(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private GetRegionInfoRequest(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private GetRegionInfoRequest() {
@@ -56,18 +56,18 @@ public final class AdminProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private GetRegionInfoRequest(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -103,22 +103,22 @@ public final class AdminProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoRequest_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoRequest_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -180,7 +180,7 @@ public final class AdminProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeMessage(1, getRegion());
@@ -197,11 +197,11 @@ public final class AdminProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(1, getRegion());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBoolSize(2, compactionState_);
}
size += unknownFields.getSerializedSize();
@@ -248,7 +248,7 @@ public final class AdminProtos {
}
if (hasCompactionState()) {
hash = (37 * hash) + COMPACTION_STATE_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean(
getCompactionState());
}
hash = (29 * hash) + unknownFields.hashCode();
@@ -257,61 +257,61 @@ public final class AdminProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -329,7 +329,7 @@ public final class AdminProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -337,15 +337,15 @@ public final class AdminProtos {
* Protobuf type {@code hbase.pb.GetRegionInfoRequest}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.GetRegionInfoRequest)
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequestOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoRequest_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoRequest_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -358,12 +358,12 @@ public final class AdminProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getRegionFieldBuilder();
}
@@ -381,7 +381,7 @@ public final class AdminProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoRequest_descriptor;
}
@@ -423,29 +423,29 @@ public final class AdminProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest)other);
} else {
@@ -478,13 +478,13 @@ public final class AdminProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -497,7 +497,7 @@ public final class AdminProtos {
private int bitField0_;
private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier region_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBuilder_;
/**
* <code>required .hbase.pb.RegionSpecifier region = 1;</code>
@@ -600,11 +600,11 @@ public final class AdminProtos {
/**
* <code>required .hbase.pb.RegionSpecifier region = 1;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>
getRegionFieldBuilder() {
if (regionBuilder_ == null) {
- regionBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ regionBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>(
getRegion(),
getParentForChildren(),
@@ -646,12 +646,12 @@ public final class AdminProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -669,22 +669,22 @@ public final class AdminProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<GetRegionInfoRequest>
- PARSER = new com.google.protobuf.AbstractParser<GetRegionInfoRequest>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<GetRegionInfoRequest>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<GetRegionInfoRequest>() {
public GetRegionInfoRequest parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new GetRegionInfoRequest(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<GetRegionInfoRequest> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<GetRegionInfoRequest> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<GetRegionInfoRequest> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<GetRegionInfoRequest> getParserForType() {
return PARSER;
}
@@ -696,7 +696,7 @@ public final class AdminProtos {
public interface GetRegionInfoResponseOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.GetRegionInfoResponse)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required .hbase.pb.RegionInfo region_info = 1;</code>
@@ -733,11 +733,11 @@ public final class AdminProtos {
* Protobuf type {@code hbase.pb.GetRegionInfoResponse}
*/
public static final class GetRegionInfoResponse extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.GetRegionInfoResponse)
GetRegionInfoResponseOrBuilder {
// Use GetRegionInfoResponse.newBuilder() to construct.
- private GetRegionInfoResponse(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private GetRegionInfoResponse(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private GetRegionInfoResponse() {
@@ -746,18 +746,18 @@ public final class AdminProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private GetRegionInfoResponse(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -804,22 +804,22 @@ public final class AdminProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoResponse_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoResponse_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -830,7 +830,7 @@ public final class AdminProtos {
* Protobuf enum {@code hbase.pb.GetRegionInfoResponse.CompactionState}
*/
public enum CompactionState
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>NONE = 0;</code>
*/
@@ -889,27 +889,27 @@ public final class AdminProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<CompactionState>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<CompactionState>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
CompactionState> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<CompactionState>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<CompactionState>() {
public CompactionState findValueByNumber(int number) {
return CompactionState.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.getDescriptor().getEnumTypes().get(0);
}
@@ -917,7 +917,7 @@ public final class AdminProtos {
private static final CompactionState[] VALUES = values();
public static CompactionState valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -1005,7 +1005,7 @@ public final class AdminProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeMessage(1, getRegionInfo());
@@ -1025,15 +1025,15 @@ public final class AdminProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(1, getRegionInfo());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeEnumSize(2, compactionState_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBoolSize(3, isRecovering_);
}
size += unknownFields.getSerializedSize();
@@ -1088,7 +1088,7 @@ public final class AdminProtos {
}
if (hasIsRecovering()) {
hash = (37 * hash) + ISRECOVERING_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean(
getIsRecovering());
}
hash = (29 * hash) + unknownFields.hashCode();
@@ -1097,61 +1097,61 @@ public final class AdminProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -1169,7 +1169,7 @@ public final class AdminProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1177,15 +1177,15 @@ public final class AdminProtos {
* Protobuf type {@code hbase.pb.GetRegionInfoResponse}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.GetRegionInfoResponse)
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponseOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoResponse_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoResponse_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1198,12 +1198,12 @@ public final class AdminProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getRegionInfoFieldBuilder();
}
@@ -1223,7 +1223,7 @@ public final class AdminProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoResponse_descriptor;
}
@@ -1269,29 +1269,29 @@ public final class AdminProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse)other);
} else {
@@ -1327,13 +1327,13 @@ public final class AdminProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -1346,7 +1346,7 @@ public final class AdminProtos {
private int bitField0_;
private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo regionInfo_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder> regionInfoBuilder_;
/**
* <code>required .hbase.pb.RegionInfo region_info = 1;</code>
@@ -1449,11 +1449,11 @@ public final class AdminProtos {
/**
* <code>required .hbase.pb.RegionInfo region_info = 1;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder>
getRegionInfoFieldBuilder() {
if (regionInfoBuilder_ == null) {
- regionInfoBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ regionInfoBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder>(
getRegionInfo(),
getParentForChildren(),
@@ -1531,12 +1531,12 @@ public final class AdminProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -1554,22 +1554,22 @@ public final class AdminProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<GetRegionInfoResponse>
- PARSER = new com.google.protobuf.AbstractParser<GetRegionInfoResponse>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<GetRegionInfoResponse>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<GetRegionInfoResponse>() {
public GetRegionInfoResponse parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new GetRegionInfoResponse(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<GetRegionInfoResponse> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<GetRegionInfoResponse> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<GetRegionInfoResponse> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<GetRegionInfoResponse> getParserForType() {
return PARSER;
}
@@ -1581,7 +1581,7 @@ public final class AdminProtos {
public interface GetStoreFileRequestOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.GetStoreFileRequest)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required .hbase.pb.RegionSpecifier region = 1;</code>
@@ -1599,7 +1599,7 @@ public final class AdminProtos {
/**
* <code>repeated bytes family = 2;</code>
*/
- java.util.List<com.google.protobuf.ByteString> getFamilyList();
+ java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString> getFamilyList();
/**
* <code>repeated bytes family = 2;</code>
*/
@@ -1607,7 +1607,7 @@ public final class AdminProtos {
/**
* <code>repeated bytes family = 2;</code>
*/
- com.google.protobuf.ByteString getFamily(int index);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamily(int index);
}
/**
* <pre>
@@ -1619,11 +1619,11 @@ public final class AdminProtos {
* Protobuf type {@code hbase.pb.GetStoreFileRequest}
*/
public static final class GetStoreFileRequest extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.GetStoreFileRequest)
GetStoreFileRequestOrBuilder {
// Use GetStoreFileRequest.newBuilder() to construct.
- private GetStoreFileRequest(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private GetStoreFileRequest(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private GetStoreFileRequest() {
@@ -1631,18 +1631,18 @@ public final class AdminProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private GetStoreFileRequest(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -1673,7 +1673,7 @@ public final class AdminProtos {
}
case 18: {
if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
- family_ = new java.util.ArrayList<com.google.protobuf.ByteString>();
+ family_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString>();
mutable_bitField0_ |= 0x00000002;
}
family_.add(input.readBytes());
@@ -1681,10 +1681,10 @@ public final class AdminProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
@@ -1694,12 +1694,12 @@ public final class AdminProtos {
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetStoreFileRequest_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetStoreFileRequest_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1729,11 +1729,11 @@ public final class AdminProtos {
}
public static final int FAMILY_FIELD_NUMBER = 2;
- private java.util.List<com.google.protobuf.ByteString> family_;
+ private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString> family_;
/**
* <code>repeated bytes family = 2;</code>
*/
- public java.util.List<com.google.protobuf.ByteString>
+ public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString>
getFamilyList() {
return family_;
}
@@ -1746,7 +1746,7 @@ public final class AdminProtos {
/**
* <code>repeated bytes family = 2;</code>
*/
- public com.google.protobuf.ByteString getFamily(int index) {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamily(int index) {
return family_.get(index);
}
@@ -1768,7 +1768,7 @@ public final class AdminProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeMessage(1, getRegion());
@@ -1785,13 +1785,13 @@ public final class AdminProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(1, getRegion());
}
{
int dataSize = 0;
for (int i = 0; i < family_.size(); i++) {
- dataSize += com.google.protobuf.CodedOutputStream
+ dataSize += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSizeNoTag(family_.get(i));
}
size += dataSize;
@@ -1846,61 +1846,61 @@ public final class AdminProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -1918,7 +1918,7 @@ public final class AdminProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1932,15 +1932,15 @@ public final class AdminProtos {
* Protobuf type {@code hbase.pb.GetStoreFileRequest}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.GetStoreFileRequest)
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequestOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetStoreFileRequest_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetStoreFileRequest_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1953,12 +1953,12 @@ public final class AdminProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getRegionFieldBuilder();
}
@@ -1976,7 +1976,7 @@ public final class AdminProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetStoreFileRequest_descriptor;
}
@@ -2019,29 +2019,29 @@ public final class AdminProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest)other);
} else {
@@ -2081,13 +2081,13 @@ public final class AdminProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -2100,7 +2100,7 @@ public final class AdminProtos {
private int bitField0_;
private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier region_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBuilder_;
/**
* <code>required .hbase.pb.RegionSpecifier region = 1;</code>
@@ -2203,11 +2203,11 @@ public final class AdminProtos {
/**
* <code>required .hbase.pb.RegionSpecifier region = 1;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>
getRegionFieldBuilder() {
if (regionBuilder_ == null) {
- regionBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ regionBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>(
getRegion(),
getParentForChildren(),
@@ -2217,17 +2217,17 @@ public final class AdminProtos {
return regionBuilder_;
}
- private java.util.List<com.google.protobuf.ByteString> family_ = java.util.Collections.emptyList();
+ private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString> family_ = java.util.Collections.emptyList();
private void ensureFamilyIsMutable() {
if (!((bitField0_ & 0x00000002) == 0x00000002)) {
- family_ = new java.util.ArrayList<com.google.protobuf.ByteString>(family_);
+ family_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString>(family_);
bitField0_ |= 0x00000002;
}
}
/**
* <code>repeated bytes family = 2;</code>
*/
- public java.util.List<com.google.protobuf.ByteString>
+ public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString>
getFamilyList() {
return java.util.Collections.unmodifiableList(family_);
}
@@ -2240,14 +2240,14 @@ public final class AdminProtos {
/**
* <code>repeated bytes family = 2;</code>
*/
- public com.google.protobuf.ByteString getFamily(int index) {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamily(int index) {
return family_.get(index);
}
/**
* <code>repeated bytes family = 2;</code>
*/
public Builder setFamily(
- int index, com.google.protobuf.ByteString value) {
+ int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -2259,7 +2259,7 @@ public final class AdminProtos {
/**
* <code>repeated bytes family = 2;</code>
*/
- public Builder addFamily(com.google.protobuf.ByteString value) {
+ public Builder addFamily(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -2272,9 +2272,9 @@ public final class AdminProtos {
* <code>repeated bytes family = 2;</code>
*/
public Builder addAllFamily(
- java.lang.Iterable<? extends com.google.protobuf.ByteString> values) {
+ java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString> values) {
ensureFamilyIsMutable();
- com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
values, family_);
onChanged();
return this;
@@ -2289,12 +2289,12 @@ public final class AdminProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -2312,22 +2312,22 @@ public final class AdminProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<GetStoreFileRequest>
- PARSER = new com.google.protobuf.AbstractParser<GetStoreFileRequest>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<GetStoreFileRequest>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<GetStoreFileRequest>() {
public GetStoreFileRequest parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new GetStoreFileRequest(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<GetStoreFileRequest> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<GetStoreFileRequest> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<GetStoreFileRequest> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<GetStoreFileRequest> getParserForType() {
return PARSER;
}
@@ -2339,7 +2339,7 @@ public final class AdminProtos {
public interface GetStoreFileResponseOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.GetStoreFileResponse)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>repeated string store_file = 1;</code>
@@ -2357,37 +2357,37 @@ public final class AdminProtos {
/**
* <code>repeated string store_file = 1;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getStoreFileBytes(int index);
}
/**
* Protobuf type {@code hbase.pb.GetStoreFileResponse}
*/
public static final class GetStoreFileResponse extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.GetStoreFileResponse)
GetStoreFileResponseOrBuilder {
// Use GetStoreFileResponse.newBuilder() to construct.
- private GetStoreFileResponse(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private GetStoreFileResponse(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private GetStoreFileResponse() {
- storeFile_ = com.google.protobuf.LazyStringArrayList.EMPTY;
+ storeFile_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private GetStoreFileResponse(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -2404,9 +2404,9 @@ public final class AdminProtos {
break;
}
case 10: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
- storeFile_ = new com.google.protobuf.LazyStringArrayList();
+ storeFile_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList();
mutable_bitField0_ |= 0x00000001;
}
storeFile_.add(bs);
@@ -2414,10 +2414,10 @@ public final class AdminProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
@@ -2427,12 +2427,12 @@ public final class AdminProtos {
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetStoreFileResponse_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetStoreFileResponse_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2440,11 +2440,11 @@ public final class AdminProtos {
}
public static final int STORE_FILE_FIELD_NUMBER = 1;
- private com.google.protobuf.LazyStringList storeFile_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringList storeFile_;
/**
* <code>repeated string store_file = 1;</code>
*/
- public com.google.protobuf.ProtocolStringList
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolStringList
getStoreFileList() {
return storeFile_;
}
@@ -2463,7 +2463,7 @@ public final class AdminProtos {
/**
* <code>repeated string store_file = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getStoreFileBytes(int index) {
return storeFile_.getByteString(index);
}
@@ -2478,10 +2478,10 @@ public final class AdminProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
for (int i = 0; i < storeFile_.size(); i++) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 1, storeFile_.getRaw(i));
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, storeFile_.getRaw(i));
}
unknownFields.writeTo(output);
}
@@ -2539,61 +2539,61 @@ public final class AdminProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileResponse parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileResponse parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileResponse parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileResponse parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileResponse parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileResponse parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileResponse parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileResponse parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileResponse parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileResponse parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -2611,7 +2611,7 @@ public final class AdminProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -2619,15 +2619,15 @@ public final class AdminProtos {
* Protobuf type {@code hbase.pb.GetStoreFileResponse}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.GetStoreFileResponse)
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileResponseOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetStoreFileResponse_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetStoreFileResponse_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2640,23 +2640,23 @@ public final class AdminProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
public Builder clear() {
super.clear();
- storeFile_ = com.google.protobuf.LazyStringArrayList.EMPTY;
+ storeFile_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
bitField0_ = (bitField0_ & ~0x00000001);
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetStoreFileResponse_descriptor;
}
@@ -2689,29 +2689,29 @@ public final class AdminProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileResponse) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileResponse)other);
} else {
@@ -2742,13 +2742,13 @@ public final class AdminProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileResponse parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileResponse) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -2760,17 +2760,17 @@ public final class AdminProtos {
}
private int bitField0_;
- private com.google.protobuf.LazyStringList storeFile_ = com.google.protobuf.LazyStringArrayList.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringList storeFile_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
private void ensureStoreFileIsMutable() {
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
- storeFile_ = new com.google.protobuf.LazyStringArrayList(storeFile_);
+ storeFile_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList(storeFile_);
bitField0_ |= 0x00000001;
}
}
/**
* <code>repeated string store_file = 1;</code>
*/
- public com.google.protobuf.ProtocolStringList
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolStringList
getStoreFileList() {
return storeFile_.getUnmodifiableView();
}
@@ -2789,7 +2789,7 @@ public final class AdminProtos {
/**
* <code>repeated string store_file = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getStoreFileBytes(int index) {
return storeFile_.getByteString(index);
}
@@ -2825,7 +2825,7 @@ public final class AdminProtos {
public Builder addAllStoreFile(
java.lang.Iterable<java.lang.String> values) {
ensureStoreFileIsMutable();
- com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
values, storeFile_);
onChanged();
return this;
@@ -2834,7 +2834,7 @@ public final class AdminProtos {
* <code>repeated string store_file = 1;</code>
*/
public Builder clearStoreFile() {
- storeFile_ = com.google.protobuf.LazyStringArrayList.EMPTY;
+ storeFile_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
return this;
@@ -2843,7 +2843,7 @@ public final class AdminProtos {
* <code>repeated string sto
<TRUNCATED>
[23/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ClientProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ClientProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ClientProtos.java
index c4ec758..80b858f 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ClientProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ClientProtos.java
@@ -6,13 +6,13 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class ClientProtos {
private ClientProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
/**
* <pre>
@@ -23,7 +23,7 @@ public final class ClientProtos {
* Protobuf enum {@code hbase.pb.Consistency}
*/
public enum Consistency
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>STRONG = 0;</code>
*/
@@ -64,27 +64,27 @@ public final class ClientProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<Consistency>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<Consistency>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
Consistency> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<Consistency>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<Consistency>() {
public Consistency findValueByNumber(int number) {
return Consistency.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.getDescriptor().getEnumTypes().get(0);
}
@@ -92,7 +92,7 @@ public final class ClientProtos {
private static final Consistency[] VALUES = values();
public static Consistency valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -111,7 +111,7 @@ public final class ClientProtos {
public interface AuthorizationsOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.Authorizations)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>repeated string label = 1;</code>
@@ -129,7 +129,7 @@ public final class ClientProtos {
/**
* <code>repeated string label = 1;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getLabelBytes(int index);
}
/**
@@ -141,30 +141,30 @@ public final class ClientProtos {
* Protobuf type {@code hbase.pb.Authorizations}
*/
public static final class Authorizations extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.Authorizations)
AuthorizationsOrBuilder {
// Use Authorizations.newBuilder() to construct.
- private Authorizations(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private Authorizations(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private Authorizations() {
- label_ = com.google.protobuf.LazyStringArrayList.EMPTY;
+ label_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private Authorizations(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -181,9 +181,9 @@ public final class ClientProtos {
break;
}
case 10: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
- label_ = new com.google.protobuf.LazyStringArrayList();
+ label_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList();
mutable_bitField0_ |= 0x00000001;
}
label_.add(bs);
@@ -191,10 +191,10 @@ public final class ClientProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
@@ -204,12 +204,12 @@ public final class ClientProtos {
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Authorizations_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Authorizations_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -217,11 +217,11 @@ public final class ClientProtos {
}
public static final int LABEL_FIELD_NUMBER = 1;
- private com.google.protobuf.LazyStringList label_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringList label_;
/**
* <code>repeated string label = 1;</code>
*/
- public com.google.protobuf.ProtocolStringList
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolStringList
getLabelList() {
return label_;
}
@@ -240,7 +240,7 @@ public final class ClientProtos {
/**
* <code>repeated string label = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getLabelBytes(int index) {
return label_.getByteString(index);
}
@@ -255,10 +255,10 @@ public final class ClientProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
for (int i = 0; i < label_.size(); i++) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 1, label_.getRaw(i));
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, label_.getRaw(i));
}
unknownFields.writeTo(output);
}
@@ -316,61 +316,61 @@ public final class ClientProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -388,7 +388,7 @@ public final class ClientProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -401,15 +401,15 @@ public final class ClientProtos {
* Protobuf type {@code hbase.pb.Authorizations}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.Authorizations)
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.AuthorizationsOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Authorizations_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Authorizations_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -422,23 +422,23 @@ public final class ClientProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
public Builder clear() {
super.clear();
- label_ = com.google.protobuf.LazyStringArrayList.EMPTY;
+ label_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
bitField0_ = (bitField0_ & ~0x00000001);
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Authorizations_descriptor;
}
@@ -471,29 +471,29 @@ public final class ClientProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations)other);
} else {
@@ -524,13 +524,13 @@ public final class ClientProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -542,17 +542,17 @@ public final class ClientProtos {
}
private int bitField0_;
- private com.google.protobuf.LazyStringList label_ = com.google.protobuf.LazyStringArrayList.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringList label_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
private void ensureLabelIsMutable() {
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
- label_ = new com.google.protobuf.LazyStringArrayList(label_);
+ label_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList(label_);
bitField0_ |= 0x00000001;
}
}
/**
* <code>repeated string label = 1;</code>
*/
- public com.google.protobuf.ProtocolStringList
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolStringList
getLabelList() {
return label_.getUnmodifiableView();
}
@@ -571,7 +571,7 @@ public final class ClientProtos {
/**
* <code>repeated string label = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getLabelBytes(int index) {
return label_.getByteString(index);
}
@@ -607,7 +607,7 @@ public final class ClientProtos {
public Builder addAllLabel(
java.lang.Iterable<java.lang.String> values) {
ensureLabelIsMutable();
- com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
values, label_);
onChanged();
return this;
@@ -616,7 +616,7 @@ public final class ClientProtos {
* <code>repeated string label = 1;</code>
*/
public Builder clearLabel() {
- label_ = com.google.protobuf.LazyStringArrayList.EMPTY;
+ label_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
return this;
@@ -625,7 +625,7 @@ public final class ClientProtos {
* <code>repeated string label = 1;</code>
*/
public Builder addLabelBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -635,12 +635,12 @@ public final class ClientProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -658,22 +658,22 @@ public final class ClientProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<Authorizations>
- PARSER = new com.google.protobuf.AbstractParser<Authorizations>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Authorizations>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Authorizations>() {
public Authorizations parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new Authorizations(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<Authorizations> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Authorizations> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<Authorizations> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Authorizations> getParserForType() {
return PARSER;
}
@@ -685,7 +685,7 @@ public final class ClientProtos {
public interface CellVisibilityOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.CellVisibility)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required string expression = 1;</code>
@@ -698,7 +698,7 @@ public final class ClientProtos {
/**
* <code>required string expression = 1;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getExpressionBytes();
}
/**
@@ -710,11 +710,11 @@ public final class ClientProtos {
* Protobuf type {@code hbase.pb.CellVisibility}
*/
public static final class CellVisibility extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.CellVisibility)
CellVisibilityOrBuilder {
// Use CellVisibility.newBuilder() to construct.
- private CellVisibility(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private CellVisibility(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private CellVisibility() {
@@ -722,18 +722,18 @@ public final class ClientProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private CellVisibility(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -750,29 +750,29 @@ public final class ClientProtos {
break;
}
case 10: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000001;
expression_ = bs;
break;
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_CellVisibility_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_CellVisibility_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -796,8 +796,8 @@ public final class ClientProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
expression_ = s;
@@ -808,17 +808,17 @@ public final class ClientProtos {
/**
* <code>required string expression = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getExpressionBytes() {
java.lang.Object ref = expression_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
expression_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -836,10 +836,10 @@ public final class ClientProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 1, expression_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, expression_);
}
unknownFields.writeTo(output);
}
@@ -850,7 +850,7 @@ public final class ClientProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, expression_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, expression_);
}
size += unknownFields.getSerializedSize();
memoizedSize = size;
@@ -895,61 +895,61 @@ public final class ClientProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -967,7 +967,7 @@ public final class ClientProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -980,15 +980,15 @@ public final class ClientProtos {
* Protobuf type {@code hbase.pb.CellVisibility}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.CellVisibility)
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibilityOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_CellVisibility_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_CellVisibility_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1001,12 +1001,12 @@ public final class ClientProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -1017,7 +1017,7 @@ public final class ClientProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_CellVisibility_descriptor;
}
@@ -1051,29 +1051,29 @@ public final class ClientProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility)other);
} else {
@@ -1102,13 +1102,13 @@ public final class ClientProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -1133,8 +1133,8 @@ public final class ClientProtos {
public java.lang.String getExpression() {
java.lang.Object ref = expression_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
expression_ = s;
@@ -1147,17 +1147,17 @@ public final class ClientProtos {
/**
* <code>required string expression = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getExpressionBytes() {
java.lang.Object ref = expression_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
expression_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -1186,7 +1186,7 @@ public final class ClientProtos {
* <code>required string expression = 1;</code>
*/
public Builder setExpressionBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1196,12 +1196,12 @@ public final class ClientProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -1219,22 +1219,22 @@ public final class ClientProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<CellVisibility>
- PARSER = new com.google.protobuf.AbstractParser<CellVisibility>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<CellVisibility>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<CellVisibility>() {
public CellVisibility parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new CellVisibility(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<CellVisibility> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<CellVisibility> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<CellVisibility> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<CellVisibility> getParserForType() {
return PARSER;
}
@@ -1246,7 +1246,7 @@ public final class ClientProtos {
public interface ColumnOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.Column)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required bytes family = 1;</code>
@@ -1255,12 +1255,12 @@ public final class ClientProtos {
/**
* <code>required bytes family = 1;</code>
*/
- com.google.protobuf.ByteString getFamily();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamily();
/**
* <code>repeated bytes qualifier = 2;</code>
*/
- java.util.List<com.google.protobuf.ByteString> getQualifierList();
+ java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString> getQualifierList();
/**
* <code>repeated bytes qualifier = 2;</code>
*/
@@ -1268,7 +1268,7 @@ public final class ClientProtos {
/**
* <code>repeated bytes qualifier = 2;</code>
*/
- com.google.protobuf.ByteString getQualifier(int index);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getQualifier(int index);
}
/**
* <pre>
@@ -1279,31 +1279,31 @@ public final class ClientProtos {
* Protobuf type {@code hbase.pb.Column}
*/
public static final class Column extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.Column)
ColumnOrBuilder {
// Use Column.newBuilder() to construct.
- private Column(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private Column(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private Column() {
- family_ = com.google.protobuf.ByteString.EMPTY;
+ family_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
qualifier_ = java.util.Collections.emptyList();
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private Column(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -1326,7 +1326,7 @@ public final class ClientProtos {
}
case 18: {
if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
- qualifier_ = new java.util.ArrayList<com.google.protobuf.ByteString>();
+ qualifier_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString>();
mutable_bitField0_ |= 0x00000002;
}
qualifier_.add(input.readBytes());
@@ -1334,10 +1334,10 @@ public final class ClientProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
@@ -1347,12 +1347,12 @@ public final class ClientProtos {
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Column_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Column_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1361,7 +1361,7 @@ public final class ClientProtos {
private int bitField0_;
public static final int FAMILY_FIELD_NUMBER = 1;
- private com.google.protobuf.ByteString family_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString family_;
/**
* <code>required bytes family = 1;</code>
*/
@@ -1371,16 +1371,16 @@ public final class ClientProtos {
/**
* <code>required bytes family = 1;</code>
*/
- public com.google.protobuf.ByteString getFamily() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamily() {
return family_;
}
public static final int QUALIFIER_FIELD_NUMBER = 2;
- private java.util.List<com.google.protobuf.ByteString> qualifier_;
+ private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString> qualifier_;
/**
* <code>repeated bytes qualifier = 2;</code>
*/
- public java.util.List<com.google.protobuf.ByteString>
+ public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString>
getQualifierList() {
return qualifier_;
}
@@ -1393,7 +1393,7 @@ public final class ClientProtos {
/**
* <code>repeated bytes qualifier = 2;</code>
*/
- public com.google.protobuf.ByteString getQualifier(int index) {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getQualifier(int index) {
return qualifier_.get(index);
}
@@ -1411,7 +1411,7 @@ public final class ClientProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, family_);
@@ -1428,13 +1428,13 @@ public final class ClientProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(1, family_);
}
{
int dataSize = 0;
for (int i = 0; i < qualifier_.size(); i++) {
- dataSize += com.google.protobuf.CodedOutputStream
+ dataSize += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSizeNoTag(qualifier_.get(i));
}
size += dataSize;
@@ -1489,61 +1489,61 @@ public final class ClientProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -1561,7 +1561,7 @@ public final class ClientProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1574,15 +1574,15 @@ public final class ClientProtos {
* Protobuf type {@code hbase.pb.Column}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.Column)
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ColumnOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Column_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Column_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1595,25 +1595,25 @@ public final class ClientProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
public Builder clear() {
super.clear();
- family_ = com.google.protobuf.ByteString.EMPTY;
+ family_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000001);
qualifier_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000002);
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Column_descriptor;
}
@@ -1652,29 +1652,29 @@ public final class ClientProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column)other);
} else {
@@ -1711,13 +1711,13 @@ public final class ClientProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -1729,7 +1729,7 @@ public final class ClientProtos {
}
private int bitField0_;
- private com.google.protobuf.ByteString family_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString family_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>required bytes family = 1;</code>
*/
@@ -1739,13 +1739,13 @@ public final class ClientProtos {
/**
* <code>required bytes family = 1;</code>
*/
- public com.google.protobuf.ByteString getFamily() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamily() {
return family_;
}
/**
* <code>required bytes family = 1;</code>
*/
- public Builder setFamily(com.google.protobuf.ByteString value) {
+ public Builder setFamily(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1764,17 +1764,17 @@ public final class ClientProtos {
return this;
}
- private java.util.List<com.google.protobuf.ByteString> qualifier_ = java.util.Collections.emptyList();
+ private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString> qualifier_ = java.util.Collections.emptyList();
private void ensureQualifierIsMutable() {
if (!((bitField0_ & 0x00000002) == 0x00000002)) {
- qualifier_ = new java.util.ArrayList<com.google.protobuf.ByteString>(qualifier_);
+ qualifier_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString>(qualifier_);
bitField0_ |= 0x00000002;
}
}
/**
* <code>repeated bytes qualifier = 2;</code>
*/
- public java.util.List<com.google.protobuf.ByteString>
+ public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString>
getQualifierList() {
return java.util.Collections.unmodifiableList(qualifier_);
}
@@ -1787,14 +1787,14 @@ public final class ClientProtos {
/**
* <code>repeated bytes qualifier = 2;</code>
*/
- public com.google.protobuf.ByteString getQualifier(int index) {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getQualifier(int index) {
return qualifier_.get(index);
}
/**
* <code>repeated bytes qualifier = 2;</code>
*/
public Builder setQualifier(
- int index, com.google.protobuf.ByteString value) {
+ int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1806,7 +1806,7 @@ public final class ClientProtos {
/**
* <code>repeated bytes qualifier = 2;</code>
*/
- public Builder addQualifier(com.google.protobuf.ByteString value) {
+ public Builder addQualifier(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1819,9 +1819,9 @@ public final class ClientProtos {
* <code>repeated bytes qualifier = 2;</code>
*/
public Builder addAllQualifier(
- java.lang.Iterable<? extends com.google.protobuf.ByteString> values) {
+ java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString> values) {
ensureQualifierIsMutable();
- com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
values, qualifier_);
onChanged();
return this;
@@ -1836,12 +1836,12 @@ public final class ClientProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -1859,22 +1859,22 @@ public final class ClientProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<Column>
- PARSER = new com.google.protobuf.AbstractParser<Column>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Column>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Column>() {
public Column parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new Column(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<Column> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Column> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<Column> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Column> getParserForType() {
return PARSER;
}
@@ -1886,7 +1886,7 @@ public final class ClientProtos {
public interface GetOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.Get)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required bytes row = 1;</code>
@@ -1895,7 +1895,7 @@ public final class ClientProtos {
/**
* <code>required bytes row = 1;</code>
*/
- com.google.protobuf.ByteString getRow();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getRow();
/**
* <code>repeated .hbase.pb.Column column = 2;</code>
@@ -2070,15 +2070,15 @@ public final class ClientProtos {
* Protobuf type {@code hbase.pb.Get}
*/
public static final class Get extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.Get)
GetOrBuilder {
// Use Get.newBuilder() to construct.
- private Get(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private Get(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private Get() {
- row_ = com.google.protobuf.ByteString.EMPTY;
+ row_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
column_ = java.util.Collections.emptyList();
attribute_ = java.util.Collections.emptyList();
maxVersions_ = 1;
@@ -2091,18 +2091,18 @@ public final class ClientProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private Get(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -2214,10 +2214,10 @@ public final class ClientProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
@@ -2233,12 +2233,12 @@ public final class ClientProtos {
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Get_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Get_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2247,7 +2247,7 @@ public final class ClientProtos {
private int bitField0_;
public static final int ROW_FIELD_NUMBER = 1;
- private com.google.protobuf.ByteString row_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString row_;
/**
* <code>required bytes row = 1;</code>
*/
@@ -2257,7 +2257,7 @@ public final class ClientProtos {
/**
* <code>required bytes row = 1;</code>
*/
- public com.google.protobuf.ByteString getRow() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getRow() {
return row_;
}
@@ -2547,7 +2547,7 @@ public final class ClientProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, row_);
@@ -2594,51 +2594,51 @@ public final class ClientProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(1, row_);
}
for (int i = 0; i < column_.size(); i++) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(2, column_.get(i));
}
for (int i = 0; i < attribute_.size(); i++) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(3, attribute_.get(i));
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(4, getFilter());
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(5, getTimeRange());
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt32Size(6, maxVersions_);
}
if (((bitField0_ & 0x00000010) == 0x00000010)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBoolSize(7, cacheBlocks_);
}
if (((bitField0_ & 0x00000020) == 0x00000020)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt32Size(8, storeLimit_);
}
if (((bitField0_ & 0x00000040) == 0x00000040)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt32Size(9, storeOffset_);
}
if (((bitField0_ & 0x00000080) == 0x00000080)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBoolSize(10, existenceOnly_);
}
if (((bitField0_ & 0x00000100) == 0x00000100)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeEnumSize(12, consistency_);
}
for (int i = 0; i < cfTimeRange_.size(); i++) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(13, cfTimeRange_.get(i));
}
size += unknownFields.getSerializedSize();
@@ -2745,7 +2745,7 @@ public final class ClientProtos {
}
if (hasCacheBlocks()) {
hash = (37 * hash) + CACHE_BLOCKS_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean(
getCacheBlocks());
}
if (hasStoreLimit()) {
@@ -2758,7 +2758,7 @@ public final class ClientProtos {
}
if (hasExistenceOnly()) {
hash = (37 * hash) + EXISTENCE_ONLY_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean(
getExistenceOnly());
}
if (hasConsistency()) {
@@ -2775,61 +2775,61 @@ public final class ClientProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Get parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Get parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Get parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Get parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Get parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Get parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Get parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Get parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Get parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Get parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -2847,7 +2847,7 @@ public final class ClientProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -2862,15 +2862,15 @@ public final class ClientProtos {
* Protobuf type {@code hbase.pb.Get}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.Get)
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.GetOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Get_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Get_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2883,12 +2883,12 @@ public final class ClientProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getColumnFieldBuilder();
getAttributeFieldBuilder();
@@ -2899,7 +2899,7 @@ public final class ClientProtos {
}
public Builder clear() {
super.clear();
- row_ = com.google.protobuf.ByteString.EMPTY;
+ row_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000001);
if (columnBuilder_ == null) {
column_ = java.util.Collections.emptyList();
@@ -2946,7 +2946,7 @@ public final class ClientProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.sha
<TRUNCATED>
[04/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/TracingProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/TracingProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/TracingProtos.java
index 940a498..755bf65 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/TracingProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/TracingProtos.java
@@ -6,17 +6,17 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class TracingProtos {
private TracingProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
public interface RPCTInfoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.RPCTInfo)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>optional int64 trace_id = 1;</code>
@@ -48,11 +48,11 @@ public final class TracingProtos {
* Protobuf type {@code hbase.pb.RPCTInfo}
*/
public static final class RPCTInfo extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.RPCTInfo)
RPCTInfoOrBuilder {
// Use RPCTInfo.newBuilder() to construct.
- private RPCTInfo(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private RPCTInfo(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private RPCTInfo() {
@@ -61,18 +61,18 @@ public final class TracingProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private RPCTInfo(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -100,22 +100,22 @@ public final class TracingProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.TracingProtos.internal_static_hbase_pb_RPCTInfo_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.TracingProtos.internal_static_hbase_pb_RPCTInfo_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -163,7 +163,7 @@ public final class TracingProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeInt64(1, traceId_);
@@ -180,11 +180,11 @@ public final class TracingProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeInt64Size(1, traceId_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeInt64Size(2, parentId_);
}
size += unknownFields.getSerializedSize();
@@ -227,12 +227,12 @@ public final class TracingProtos {
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTraceId()) {
hash = (37 * hash) + TRACE_ID_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getTraceId());
}
if (hasParentId()) {
hash = (37 * hash) + PARENT_ID_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getParentId());
}
hash = (29 * hash) + unknownFields.hashCode();
@@ -241,61 +241,61 @@ public final class TracingProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.TracingProtos.RPCTInfo parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.TracingProtos.RPCTInfo parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.TracingProtos.RPCTInfo parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.TracingProtos.RPCTInfo parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.TracingProtos.RPCTInfo parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.TracingProtos.RPCTInfo parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.TracingProtos.RPCTInfo parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.TracingProtos.RPCTInfo parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.TracingProtos.RPCTInfo parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.TracingProtos.RPCTInfo parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -313,7 +313,7 @@ public final class TracingProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -329,15 +329,15 @@ public final class TracingProtos {
* Protobuf type {@code hbase.pb.RPCTInfo}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.RPCTInfo)
org.apache.hadoop.hbase.shaded.protobuf.generated.TracingProtos.RPCTInfoOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.TracingProtos.internal_static_hbase_pb_RPCTInfo_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.TracingProtos.internal_static_hbase_pb_RPCTInfo_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -350,12 +350,12 @@ public final class TracingProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -368,7 +368,7 @@ public final class TracingProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.TracingProtos.internal_static_hbase_pb_RPCTInfo_descriptor;
}
@@ -406,29 +406,29 @@ public final class TracingProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.TracingProtos.RPCTInfo) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.TracingProtos.RPCTInfo)other);
} else {
@@ -455,13 +455,13 @@ public final class TracingProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.TracingProtos.RPCTInfo parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.TracingProtos.RPCTInfo) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -537,12 +537,12 @@ public final class TracingProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -560,22 +560,22 @@ public final class TracingProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<RPCTInfo>
- PARSER = new com.google.protobuf.AbstractParser<RPCTInfo>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RPCTInfo>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<RPCTInfo>() {
public RPCTInfo parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new RPCTInfo(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<RPCTInfo> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RPCTInfo> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<RPCTInfo> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<RPCTInfo> getParserForType() {
return PARSER;
}
@@ -585,17 +585,17 @@ public final class TracingProtos {
}
- private static final com.google.protobuf.Descriptors.Descriptor
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
internal_static_hbase_pb_RPCTInfo_descriptor;
private static final
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hbase_pb_RPCTInfo_fieldAccessorTable;
- public static com.google.protobuf.Descriptors.FileDescriptor
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
getDescriptor() {
return descriptor;
}
- private static com.google.protobuf.Descriptors.FileDescriptor
+ private static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
descriptor;
static {
java.lang.String[] descriptorData = {
@@ -604,22 +604,22 @@ public final class TracingProtos {
"g.apache.hadoop.hbase.shaded.protobuf.ge" +
"neratedB\rTracingProtosH\001\240\001\001"
};
- com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
- new com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
- public com.google.protobuf.ExtensionRegistry assignDescriptors(
- com.google.protobuf.Descriptors.FileDescriptor root) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
descriptor = root;
return null;
}
};
- com.google.protobuf.Descriptors.FileDescriptor
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
.internalBuildGeneratedFileFrom(descriptorData,
- new com.google.protobuf.Descriptors.FileDescriptor[] {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
}, assigner);
internal_static_hbase_pb_RPCTInfo_descriptor =
getDescriptor().getMessageTypes().get(0);
internal_static_hbase_pb_RPCTInfo_fieldAccessorTable = new
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hbase_pb_RPCTInfo_descriptor,
new java.lang.String[] { "TraceId", "ParentId", });
}
[17/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FSProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FSProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FSProtos.java
index 5b456c8..cbe92ca 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FSProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FSProtos.java
@@ -6,17 +6,17 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class FSProtos {
private FSProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
public interface HBaseVersionFileContentOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.HBaseVersionFileContent)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required string version = 1;</code>
@@ -29,7 +29,7 @@ public final class FSProtos {
/**
* <code>required string version = 1;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getVersionBytes();
}
/**
@@ -41,11 +41,11 @@ public final class FSProtos {
* Protobuf type {@code hbase.pb.HBaseVersionFileContent}
*/
public static final class HBaseVersionFileContent extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.HBaseVersionFileContent)
HBaseVersionFileContentOrBuilder {
// Use HBaseVersionFileContent.newBuilder() to construct.
- private HBaseVersionFileContent(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private HBaseVersionFileContent(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private HBaseVersionFileContent() {
@@ -53,18 +53,18 @@ public final class FSProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private HBaseVersionFileContent(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -81,29 +81,29 @@ public final class FSProtos {
break;
}
case 10: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000001;
version_ = bs;
break;
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_HBaseVersionFileContent_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_HBaseVersionFileContent_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -127,8 +127,8 @@ public final class FSProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
version_ = s;
@@ -139,17 +139,17 @@ public final class FSProtos {
/**
* <code>required string version = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getVersionBytes() {
java.lang.Object ref = version_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
version_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -167,10 +167,10 @@ public final class FSProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 1, version_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, version_);
}
unknownFields.writeTo(output);
}
@@ -181,7 +181,7 @@ public final class FSProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, version_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, version_);
}
size += unknownFields.getSerializedSize();
memoizedSize = size;
@@ -226,61 +226,61 @@ public final class FSProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -298,7 +298,7 @@ public final class FSProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -311,15 +311,15 @@ public final class FSProtos {
* Protobuf type {@code hbase.pb.HBaseVersionFileContent}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.HBaseVersionFileContent)
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContentOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_HBaseVersionFileContent_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_HBaseVersionFileContent_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -332,12 +332,12 @@ public final class FSProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -348,7 +348,7 @@ public final class FSProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_HBaseVersionFileContent_descriptor;
}
@@ -382,29 +382,29 @@ public final class FSProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent)other);
} else {
@@ -433,13 +433,13 @@ public final class FSProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -464,8 +464,8 @@ public final class FSProtos {
public java.lang.String getVersion() {
java.lang.Object ref = version_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
version_ = s;
@@ -478,17 +478,17 @@ public final class FSProtos {
/**
* <code>required string version = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getVersionBytes() {
java.lang.Object ref = version_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
version_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -517,7 +517,7 @@ public final class FSProtos {
* <code>required string version = 1;</code>
*/
public Builder setVersionBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -527,12 +527,12 @@ public final class FSProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -550,22 +550,22 @@ public final class FSProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<HBaseVersionFileContent>
- PARSER = new com.google.protobuf.AbstractParser<HBaseVersionFileContent>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<HBaseVersionFileContent>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<HBaseVersionFileContent>() {
public HBaseVersionFileContent parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new HBaseVersionFileContent(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<HBaseVersionFileContent> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<HBaseVersionFileContent> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<HBaseVersionFileContent> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<HBaseVersionFileContent> getParserForType() {
return PARSER;
}
@@ -577,7 +577,7 @@ public final class FSProtos {
public interface ReferenceOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.Reference)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required bytes splitkey = 1;</code>
@@ -586,7 +586,7 @@ public final class FSProtos {
/**
* <code>required bytes splitkey = 1;</code>
*/
- com.google.protobuf.ByteString getSplitkey();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getSplitkey();
/**
* <code>required .hbase.pb.Reference.Range range = 2;</code>
@@ -606,31 +606,31 @@ public final class FSProtos {
* Protobuf type {@code hbase.pb.Reference}
*/
public static final class Reference extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.Reference)
ReferenceOrBuilder {
// Use Reference.newBuilder() to construct.
- private Reference(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private Reference(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private Reference() {
- splitkey_ = com.google.protobuf.ByteString.EMPTY;
+ splitkey_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
range_ = 0;
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private Reference(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -664,22 +664,22 @@ public final class FSProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_Reference_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_Reference_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -690,7 +690,7 @@ public final class FSProtos {
* Protobuf enum {@code hbase.pb.Reference.Range}
*/
public enum Range
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>TOP = 0;</code>
*/
@@ -731,27 +731,27 @@ public final class FSProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<Range>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<Range>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
Range> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<Range>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<Range>() {
public Range findValueByNumber(int number) {
return Range.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.getDescriptor().getEnumTypes().get(0);
}
@@ -759,7 +759,7 @@ public final class FSProtos {
private static final Range[] VALUES = values();
public static Range valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -778,7 +778,7 @@ public final class FSProtos {
private int bitField0_;
public static final int SPLITKEY_FIELD_NUMBER = 1;
- private com.google.protobuf.ByteString splitkey_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString splitkey_;
/**
* <code>required bytes splitkey = 1;</code>
*/
@@ -788,7 +788,7 @@ public final class FSProtos {
/**
* <code>required bytes splitkey = 1;</code>
*/
- public com.google.protobuf.ByteString getSplitkey() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getSplitkey() {
return splitkey_;
}
@@ -826,7 +826,7 @@ public final class FSProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, splitkey_);
@@ -843,11 +843,11 @@ public final class FSProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(1, splitkey_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeEnumSize(2, range_);
}
size += unknownFields.getSerializedSize();
@@ -901,61 +901,61 @@ public final class FSProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -973,7 +973,7 @@ public final class FSProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -986,15 +986,15 @@ public final class FSProtos {
* Protobuf type {@code hbase.pb.Reference}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.Reference)
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.ReferenceOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_Reference_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_Reference_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1007,25 +1007,25 @@ public final class FSProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
public Builder clear() {
super.clear();
- splitkey_ = com.google.protobuf.ByteString.EMPTY;
+ splitkey_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000001);
range_ = 0;
bitField0_ = (bitField0_ & ~0x00000002);
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_Reference_descriptor;
}
@@ -1063,29 +1063,29 @@ public final class FSProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference)other);
} else {
@@ -1118,13 +1118,13 @@ public final class FSProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -1136,7 +1136,7 @@ public final class FSProtos {
}
private int bitField0_;
- private com.google.protobuf.ByteString splitkey_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString splitkey_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>required bytes splitkey = 1;</code>
*/
@@ -1146,13 +1146,13 @@ public final class FSProtos {
/**
* <code>required bytes splitkey = 1;</code>
*/
- public com.google.protobuf.ByteString getSplitkey() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getSplitkey() {
return splitkey_;
}
/**
* <code>required bytes splitkey = 1;</code>
*/
- public Builder setSplitkey(com.google.protobuf.ByteString value) {
+ public Builder setSplitkey(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1207,12 +1207,12 @@ public final class FSProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -1230,22 +1230,22 @@ public final class FSProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<Reference>
- PARSER = new com.google.protobuf.AbstractParser<Reference>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Reference>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Reference>() {
public Reference parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new Reference(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<Reference> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Reference> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<Reference> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Reference> getParserForType() {
return PARSER;
}
@@ -1255,22 +1255,22 @@ public final class FSProtos {
}
- private static final com.google.protobuf.Descriptors.Descriptor
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
internal_static_hbase_pb_HBaseVersionFileContent_descriptor;
private static final
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hbase_pb_HBaseVersionFileContent_fieldAccessorTable;
- private static final com.google.protobuf.Descriptors.Descriptor
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
internal_static_hbase_pb_Reference_descriptor;
private static final
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hbase_pb_Reference_fieldAccessorTable;
- public static com.google.protobuf.Descriptors.FileDescriptor
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
getDescriptor() {
return descriptor;
}
- private static com.google.protobuf.Descriptors.FileDescriptor
+ private static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
descriptor;
static {
java.lang.String[] descriptorData = {
@@ -1281,28 +1281,28 @@ public final class FSProtos {
"OTTOM\020\001BB\n1org.apache.hadoop.hbase.shade" +
"d.protobuf.generatedB\010FSProtosH\001\240\001\001"
};
- com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
- new com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
- public com.google.protobuf.ExtensionRegistry assignDescriptors(
- com.google.protobuf.Descriptors.FileDescriptor root) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
descriptor = root;
return null;
}
};
- com.google.protobuf.Descriptors.FileDescriptor
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
.internalBuildGeneratedFileFrom(descriptorData,
- new com.google.protobuf.Descriptors.FileDescriptor[] {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
}, assigner);
internal_static_hbase_pb_HBaseVersionFileContent_descriptor =
getDescriptor().getMessageTypes().get(0);
internal_static_hbase_pb_HBaseVersionFileContent_fieldAccessorTable = new
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hbase_pb_HBaseVersionFileContent_descriptor,
new java.lang.String[] { "Version", });
internal_static_hbase_pb_Reference_descriptor =
getDescriptor().getMessageTypes().get(1);
internal_static_hbase_pb_Reference_fieldAccessorTable = new
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hbase_pb_Reference_descriptor,
new java.lang.String[] { "Splitkey", "Range", });
}
[11/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProtos.java
index 4426013..03ef208 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProtos.java
@@ -6,19 +6,19 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class MasterProtos {
private MasterProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
/**
* Protobuf enum {@code hbase.pb.MasterSwitchType}
*/
public enum MasterSwitchType
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>SPLIT = 0;</code>
*/
@@ -59,27 +59,27 @@ public final class MasterProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<MasterSwitchType>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<MasterSwitchType>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
MasterSwitchType> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<MasterSwitchType>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<MasterSwitchType>() {
public MasterSwitchType findValueByNumber(int number) {
return MasterSwitchType.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.getDescriptor().getEnumTypes().get(0);
}
@@ -87,7 +87,7 @@ public final class MasterProtos {
private static final MasterSwitchType[] VALUES = values();
public static MasterSwitchType valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -106,7 +106,7 @@ public final class MasterProtos {
public interface AddColumnRequestOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.AddColumnRequest)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
@@ -156,11 +156,11 @@ public final class MasterProtos {
* Protobuf type {@code hbase.pb.AddColumnRequest}
*/
public static final class AddColumnRequest extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.AddColumnRequest)
AddColumnRequestOrBuilder {
// Use AddColumnRequest.newBuilder() to construct.
- private AddColumnRequest(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private AddColumnRequest(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private AddColumnRequest() {
@@ -169,18 +169,18 @@ public final class MasterProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private AddColumnRequest(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -234,22 +234,22 @@ public final class MasterProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AddColumnRequest_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AddColumnRequest_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -355,7 +355,7 @@ public final class MasterProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeMessage(1, getTableName());
@@ -378,19 +378,19 @@ public final class MasterProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(1, getTableName());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(2, getColumnFamilies());
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(3, nonceGroup_);
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(4, nonce_);
}
size += unknownFields.getSerializedSize();
@@ -451,12 +451,12 @@ public final class MasterProtos {
}
if (hasNonceGroup()) {
hash = (37 * hash) + NONCE_GROUP_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getNonceGroup());
}
if (hasNonce()) {
hash = (37 * hash) + NONCE_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getNonce());
}
hash = (29 * hash) + unknownFields.hashCode();
@@ -465,61 +465,61 @@ public final class MasterProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -537,7 +537,7 @@ public final class MasterProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -545,15 +545,15 @@ public final class MasterProtos {
* Protobuf type {@code hbase.pb.AddColumnRequest}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.AddColumnRequest)
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequestOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AddColumnRequest_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AddColumnRequest_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -566,12 +566,12 @@ public final class MasterProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getTableNameFieldBuilder();
getColumnFamiliesFieldBuilder();
@@ -598,7 +598,7 @@ public final class MasterProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AddColumnRequest_descriptor;
}
@@ -652,29 +652,29 @@ public final class MasterProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest)other);
} else {
@@ -719,13 +719,13 @@ public final class MasterProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -738,7 +738,7 @@ public final class MasterProtos {
private int bitField0_;
private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
@@ -841,11 +841,11 @@ public final class MasterProtos {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>
getTableNameFieldBuilder() {
if (tableNameBuilder_ == null) {
- tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ tableNameBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
getTableName(),
getParentForChildren(),
@@ -856,7 +856,7 @@ public final class MasterProtos {
}
private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> columnFamiliesBuilder_;
/**
* <code>required .hbase.pb.ColumnFamilySchema column_families = 2;</code>
@@ -959,11 +959,11 @@ public final class MasterProtos {
/**
* <code>required .hbase.pb.ColumnFamilySchema column_families = 2;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder>
getColumnFamiliesFieldBuilder() {
if (columnFamiliesBuilder_ == null) {
- columnFamiliesBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ columnFamiliesBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder>(
getColumnFamilies(),
getParentForChildren(),
@@ -1037,12 +1037,12 @@ public final class MasterProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -1060,22 +1060,22 @@ public final class MasterProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<AddColumnRequest>
- PARSER = new com.google.protobuf.AbstractParser<AddColumnRequest>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<AddColumnRequest>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<AddColumnRequest>() {
public AddColumnRequest parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new AddColumnRequest(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<AddColumnRequest> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<AddColumnRequest> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<AddColumnRequest> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<AddColumnRequest> getParserForType() {
return PARSER;
}
@@ -1087,7 +1087,7 @@ public final class MasterProtos {
public interface AddColumnResponseOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.AddColumnResponse)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>optional uint64 proc_id = 1;</code>
@@ -1102,11 +1102,11 @@ public final class MasterProtos {
* Protobuf type {@code hbase.pb.AddColumnResponse}
*/
public static final class AddColumnResponse extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.AddColumnResponse)
AddColumnResponseOrBuilder {
// Use AddColumnResponse.newBuilder() to construct.
- private AddColumnResponse(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private AddColumnResponse(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private AddColumnResponse() {
@@ -1114,18 +1114,18 @@ public final class MasterProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private AddColumnResponse(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -1148,22 +1148,22 @@ public final class MasterProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AddColumnResponse_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AddColumnResponse_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1196,7 +1196,7 @@ public final class MasterProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeUInt64(1, procId_);
@@ -1210,7 +1210,7 @@ public final class MasterProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(1, procId_);
}
size += unknownFields.getSerializedSize();
@@ -1248,7 +1248,7 @@ public final class MasterProtos {
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasProcId()) {
hash = (37 * hash) + PROC_ID_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getProcId());
}
hash = (29 * hash) + unknownFields.hashCode();
@@ -1257,61 +1257,61 @@ public final class MasterProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -1329,7 +1329,7 @@ public final class MasterProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1337,15 +1337,15 @@ public final class MasterProtos {
* Protobuf type {@code hbase.pb.AddColumnResponse}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.AddColumnResponse)
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponseOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AddColumnResponse_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AddColumnResponse_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1358,12 +1358,12 @@ public final class MasterProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -1374,7 +1374,7 @@ public final class MasterProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AddColumnResponse_descriptor;
}
@@ -1408,29 +1408,29 @@ public final class MasterProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse)other);
} else {
@@ -1454,13 +1454,13 @@ public final class MasterProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -1504,12 +1504,12 @@ public final class MasterProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -1527,22 +1527,22 @@ public final class MasterProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<AddColumnResponse>
- PARSER = new com.google.protobuf.AbstractParser<AddColumnResponse>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<AddColumnResponse>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<AddColumnResponse>() {
public AddColumnResponse parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new AddColumnResponse(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<AddColumnResponse> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<AddColumnResponse> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<AddColumnResponse> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<AddColumnResponse> getParserForType() {
return PARSER;
}
@@ -1554,7 +1554,7 @@ public final class MasterProtos {
public interface DeleteColumnRequestOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.DeleteColumnRequest)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
@@ -1576,7 +1576,7 @@ public final class MasterProtos {
/**
* <code>required bytes column_name = 2;</code>
*/
- com.google.protobuf.ByteString getColumnName();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getColumnName();
/**
* <code>optional uint64 nonce_group = 3 [default = 0];</code>
@@ -1600,32 +1600,32 @@ public final class MasterProtos {
* Protobuf type {@code hbase.pb.DeleteColumnRequest}
*/
public static final class DeleteColumnRequest extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.DeleteColumnRequest)
DeleteColumnRequestOrBuilder {
// Use DeleteColumnRequest.newBuilder() to construct.
- private DeleteColumnRequest(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private DeleteColumnRequest(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private DeleteColumnRequest() {
- columnName_ = com.google.protobuf.ByteString.EMPTY;
+ columnName_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
nonceGroup_ = 0L;
nonce_ = 0L;
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private DeleteColumnRequest(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -1671,22 +1671,22 @@ public final class MasterProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteColumnRequest_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteColumnRequest_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1716,7 +1716,7 @@ public final class MasterProtos {
}
public static final int COLUMN_NAME_FIELD_NUMBER = 2;
- private com.google.protobuf.ByteString columnName_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString columnName_;
/**
* <code>required bytes column_name = 2;</code>
*/
@@ -1726,7 +1726,7 @@ public final class MasterProtos {
/**
* <code>required bytes column_name = 2;</code>
*/
- public com.google.protobuf.ByteString getColumnName() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getColumnName() {
return columnName_;
}
@@ -1782,7 +1782,7 @@ public final class MasterProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeMessage(1, getTableName());
@@ -1805,19 +1805,19 @@ public final class MasterProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(1, getTableName());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(2, columnName_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(3, nonceGroup_);
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(4, nonce_);
}
size += unknownFields.getSerializedSize();
@@ -1878,12 +1878,12 @@ public final class MasterProtos {
}
if (hasNonceGroup()) {
hash = (37 * hash) + NONCE_GROUP_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getNonceGroup());
}
if (hasNonce()) {
hash = (37 * hash) + NONCE_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getNonce());
}
hash = (29 * hash) + unknownFields.hashCode();
@@ -1892,61 +1892,61 @@ public final class MasterProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -1964,7 +1964,7 @@ public final class MasterProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1972,15 +1972,15 @@ public final class MasterProtos {
* Protobuf type {@code hbase.pb.DeleteColumnRequest}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.DeleteColumnRequest)
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequestOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteColumnRequest_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteColumnRequest_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1993,12 +1993,12 @@ public final class MasterProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getTableNameFieldBuilder();
}
@@ -2011,7 +2011,7 @@ public final class MasterProtos {
tableNameBuilder_.clear();
}
bitField0_ = (bitField0_ & ~0x00000001);
- columnName_ = com.google.protobuf.ByteString.EMPTY;
+ columnName_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000002);
nonceGroup_ = 0L;
bitField0_ = (bitField0_ & ~0x00000004);
@@ -2020,7 +2020,7 @@ public final class MasterProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteColumnRequest_descriptor;
}
@@ -2070,29 +2070,29 @@ public final class MasterProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest)other);
} else {
@@ -2134,13 +2134,13 @@ public final class MasterProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -2153,7 +2153,7 @@ public final class MasterProtos {
private int bitField0_;
private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
@@ -2256,11 +2256,11 @@ public final class MasterProtos {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>
getTableNameFieldBuilder() {
if (tableNameBuilder_ == null) {
- tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ tableNameBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
getTableName(),
getParentForChildren(),
@@ -2270,7 +2270,7 @@ public final class MasterProtos {
return tableNameBuilder_;
}
- private com.google.protobuf.ByteString columnName_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString columnName_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>required bytes column_name = 2;</code>
*/
@@ -2280,13 +2280,13 @@ public final class MasterProtos {
/**
* <code>required bytes column_name = 2;</code>
*/
- public com.google.protobuf.ByteString getColumnName() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getColumnName() {
return columnName_;
}
/**
* <code>required bytes column_name = 2;</code>
*/
- public Builder setColumnName(com.google.protobuf.ByteString value) {
+ public Builder setColumnName(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -2369,12 +2369,12 @@ public final class MasterProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -2392,22 +2392,22 @@ public final class MasterProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<DeleteColumnRequest>
- PARSER = new com.google.protobuf.AbstractParser<DeleteColumnRequest>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<DeleteColumnRequest>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<DeleteColumnRequest>() {
public DeleteColumnRequest parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new DeleteColumnRequest(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<DeleteColumnRequest> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<DeleteColumnRequest> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<DeleteColumnRequest> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<DeleteColumnRequest> getParserForType() {
return PARSER;
}
@@ -2419,7 +2419,7 @@ public final class MasterProtos {
public interface DeleteColumnResponseOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.DeleteColumnResponse)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>optional uint64 proc_id = 1;</code>
@@ -2434,11 +2434,11 @@ public final class MasterProtos {
* Protobuf type {@code hbase.pb.DeleteColumnResponse}
*/
public static final class DeleteColumnResponse extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.DeleteColumnResponse)
DeleteColumnResponseOrBuilder {
// Use DeleteColumnResponse.newBuilder() to construct.
- private DeleteColumnResponse(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private DeleteColumnResponse(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private DeleteColumnResponse() {
@@ -2446,18 +2446,18 @@ public final class MasterProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private DeleteColumnResponse(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -2480,22 +2480,22 @@ public final class MasterProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteColumnResponse_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteColumnResponse_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2528,7 +2528,7 @@ public final class MasterProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeUInt64(1, procId_);
@@ -2542,7 +2542,7 @@ public final class MasterProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(1, procId_);
}
size += unknownFields.getSerializedSize();
@@ -2580,7 +2580,7 @@ public final class MasterProtos {
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasProcId()) {
hash = (37 * hash) + PROC_ID_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getProcId());
}
hash = (29 * hash) + unknownFields.hashCode();
@@ -2589,61 +2589,61 @@ public final class MasterProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -2661,7 +2661,7 @@ public final class MasterProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -2669,15 +2669,15 @@ public final class MasterProtos {
* Protobuf type {@code hbase.pb.DeleteColumnResponse}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.DeleteColumnResponse)
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponseOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteColumnResponse_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteColumnResponse_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2690,12 +2690,12 @@ public final class MasterProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -2706,7 +2706,7 @@ public final class MasterProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteColumnResponse_descriptor;
}
@@ -2740,29 +2740,29 @@ public final class MasterProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse)other);
} else {
@@ -2786,13 +2786,13 @@ public final class MasterProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -2836,12 +2836,12 @@ public final class MasterProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -2859,22 +2859,22 @@ public final class MasterProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<DeleteColumnResponse>
- PARSER = new com.google.protobuf.AbstractParser<DeleteColumnResponse>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<DeleteColumnResponse>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<DeleteColumnResponse>() {
public DeleteColumnResponse parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new DeleteColumnResponse(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<DeleteColumnResponse> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<DeleteColumnResponse> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<DeleteColumnResponse> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<DeleteColumnResponse> getParserForType() {
return PARSER;
}
@@ -2886,7 +2886,7 @@ public final class MasterProtos {
public interface ModifyColumnRequestOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.ModifyColumnRequest)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
@@ -2936,11 +2936,11 @@ public final class MasterProtos {
* Protobuf type {@code hbase.pb.ModifyColumnRequest}
*/
public static final class ModifyColumnRequest extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.ModifyColumnRequest)
ModifyColumnRequestOrBuilder {
// Use ModifyColumnRequest.newBuilder() to construct.
- private ModifyColumnRequest(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private ModifyColumnRequest(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private ModifyColumnRequest() {
@@ -2949,18 +2949,18 @@ public final class MasterProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private ModifyColumnRequest(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -3014,22 +3014,22 @@ public final class MasterProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtoco
<TRUNCATED>
[14/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/HFileProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/HFileProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/HFileProtos.java
index 2a3d649..00e7080 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/HFileProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/HFileProtos.java
@@ -6,17 +6,17 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class HFileProtos {
private HFileProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
public interface FileInfoProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.FileInfoProto)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>repeated .hbase.pb.BytesBytesPair map_entry = 1;</code>
@@ -50,11 +50,11 @@ public final class HFileProtos {
* Protobuf type {@code hbase.pb.FileInfoProto}
*/
public static final class FileInfoProto extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.FileInfoProto)
FileInfoProtoOrBuilder {
// Use FileInfoProto.newBuilder() to construct.
- private FileInfoProto(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private FileInfoProto(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private FileInfoProto() {
@@ -62,18 +62,18 @@ public final class HFileProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private FileInfoProto(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -100,10 +100,10 @@ public final class HFileProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
@@ -113,12 +113,12 @@ public final class HFileProtos {
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.internal_static_hbase_pb_FileInfoProto_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.internal_static_hbase_pb_FileInfoProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -176,7 +176,7 @@ public final class HFileProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
for (int i = 0; i < mapEntry_.size(); i++) {
output.writeMessage(1, mapEntry_.get(i));
@@ -190,7 +190,7 @@ public final class HFileProtos {
size = 0;
for (int i = 0; i < mapEntry_.size(); i++) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(1, mapEntry_.get(i));
}
size += unknownFields.getSerializedSize();
@@ -233,61 +233,61 @@ public final class HFileProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileInfoProto parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileInfoProto parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileInfoProto parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileInfoProto parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileInfoProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileInfoProto parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileInfoProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileInfoProto parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileInfoProto parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileInfoProto parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -305,7 +305,7 @@ public final class HFileProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -317,15 +317,15 @@ public final class HFileProtos {
* Protobuf type {@code hbase.pb.FileInfoProto}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.FileInfoProto)
org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileInfoProtoOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.internal_static_hbase_pb_FileInfoProto_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.internal_static_hbase_pb_FileInfoProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -338,12 +338,12 @@ public final class HFileProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getMapEntryFieldBuilder();
}
@@ -359,7 +359,7 @@ public final class HFileProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.internal_static_hbase_pb_FileInfoProto_descriptor;
}
@@ -396,29 +396,29 @@ public final class HFileProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileInfoProto) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileInfoProto)other);
} else {
@@ -448,7 +448,7 @@ public final class HFileProtos {
mapEntry_ = other.mapEntry_;
bitField0_ = (bitField0_ & ~0x00000001);
mapEntryBuilder_ =
- com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getMapEntryFieldBuilder() : null;
} else {
mapEntryBuilder_.addAllMessages(other.mapEntry_);
@@ -470,13 +470,13 @@ public final class HFileProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileInfoProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileInfoProto) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -497,7 +497,7 @@ public final class HFileProtos {
}
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder> mapEntryBuilder_;
/**
@@ -629,7 +629,7 @@ public final class HFileProtos {
java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair> values) {
if (mapEntryBuilder_ == null) {
ensureMapEntryIsMutable();
- com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
values, mapEntry_);
onChanged();
} else {
@@ -713,11 +713,11 @@ public final class HFileProtos {
getMapEntryBuilderList() {
return getMapEntryFieldBuilder().getBuilderList();
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder>
getMapEntryFieldBuilder() {
if (mapEntryBuilder_ == null) {
- mapEntryBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3<
+ mapEntryBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder>(
mapEntry_,
((bitField0_ & 0x00000001) == 0x00000001),
@@ -728,12 +728,12 @@ public final class HFileProtos {
return mapEntryBuilder_;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -751,22 +751,22 @@ public final class HFileProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<FileInfoProto>
- PARSER = new com.google.protobuf.AbstractParser<FileInfoProto>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<FileInfoProto>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<FileInfoProto>() {
public FileInfoProto parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new FileInfoProto(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<FileInfoProto> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<FileInfoProto> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<FileInfoProto> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<FileInfoProto> getParserForType() {
return PARSER;
}
@@ -778,7 +778,7 @@ public final class HFileProtos {
public interface FileTrailerProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.FileTrailerProto)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>optional uint64 file_info_offset = 1;</code>
@@ -881,7 +881,7 @@ public final class HFileProtos {
/**
* <code>optional string comparator_class_name = 11;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getComparatorClassNameBytes();
/**
@@ -900,7 +900,7 @@ public final class HFileProtos {
/**
* <code>optional bytes encryption_key = 13;</code>
*/
- com.google.protobuf.ByteString getEncryptionKey();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getEncryptionKey();
}
/**
* <pre>
@@ -910,11 +910,11 @@ public final class HFileProtos {
* Protobuf type {@code hbase.pb.FileTrailerProto}
*/
public static final class FileTrailerProto extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.FileTrailerProto)
FileTrailerProtoOrBuilder {
// Use FileTrailerProto.newBuilder() to construct.
- private FileTrailerProto(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private FileTrailerProto(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private FileTrailerProto() {
@@ -930,22 +930,22 @@ public final class HFileProtos {
lastDataBlockOffset_ = 0L;
comparatorClassName_ = "";
compressionCodec_ = 0;
- encryptionKey_ = com.google.protobuf.ByteString.EMPTY;
+ encryptionKey_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private FileTrailerProto(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -1012,7 +1012,7 @@ public final class HFileProtos {
break;
}
case 90: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000400;
comparatorClassName_ = bs;
break;
@@ -1029,22 +1029,22 @@ public final class HFileProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.internal_static_hbase_pb_FileTrailerProto_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.internal_static_hbase_pb_FileTrailerProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1218,8 +1218,8 @@ public final class HFileProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
comparatorClassName_ = s;
@@ -1230,17 +1230,17 @@ public final class HFileProtos {
/**
* <code>optional string comparator_class_name = 11;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getComparatorClassNameBytes() {
java.lang.Object ref = comparatorClassName_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
comparatorClassName_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -1260,7 +1260,7 @@ public final class HFileProtos {
}
public static final int ENCRYPTION_KEY_FIELD_NUMBER = 13;
- private com.google.protobuf.ByteString encryptionKey_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString encryptionKey_;
/**
* <code>optional bytes encryption_key = 13;</code>
*/
@@ -1270,7 +1270,7 @@ public final class HFileProtos {
/**
* <code>optional bytes encryption_key = 13;</code>
*/
- public com.google.protobuf.ByteString getEncryptionKey() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getEncryptionKey() {
return encryptionKey_;
}
@@ -1284,7 +1284,7 @@ public final class HFileProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeUInt64(1, fileInfoOffset_);
@@ -1317,7 +1317,7 @@ public final class HFileProtos {
output.writeUInt64(10, lastDataBlockOffset_);
}
if (((bitField0_ & 0x00000400) == 0x00000400)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 11, comparatorClassName_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 11, comparatorClassName_);
}
if (((bitField0_ & 0x00000800) == 0x00000800)) {
output.writeUInt32(12, compressionCodec_);
@@ -1334,54 +1334,54 @@ public final class HFileProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(1, fileInfoOffset_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(2, loadOnOpenDataOffset_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(3, uncompressedDataIndexSize_);
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(4, totalUncompressedBytes_);
}
if (((bitField0_ & 0x00000010) == 0x00000010)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt32Size(5, dataIndexCount_);
}
if (((bitField0_ & 0x00000020) == 0x00000020)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt32Size(6, metaIndexCount_);
}
if (((bitField0_ & 0x00000040) == 0x00000040)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(7, entryCount_);
}
if (((bitField0_ & 0x00000080) == 0x00000080)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt32Size(8, numDataIndexLevels_);
}
if (((bitField0_ & 0x00000100) == 0x00000100)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(9, firstDataBlockOffset_);
}
if (((bitField0_ & 0x00000200) == 0x00000200)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(10, lastDataBlockOffset_);
}
if (((bitField0_ & 0x00000400) == 0x00000400)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(11, comparatorClassName_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(11, comparatorClassName_);
}
if (((bitField0_ & 0x00000800) == 0x00000800)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt32Size(12, compressionCodec_);
}
if (((bitField0_ & 0x00001000) == 0x00001000)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(13, encryptionKey_);
}
size += unknownFields.getSerializedSize();
@@ -1479,22 +1479,22 @@ public final class HFileProtos {
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasFileInfoOffset()) {
hash = (37 * hash) + FILE_INFO_OFFSET_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getFileInfoOffset());
}
if (hasLoadOnOpenDataOffset()) {
hash = (37 * hash) + LOAD_ON_OPEN_DATA_OFFSET_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getLoadOnOpenDataOffset());
}
if (hasUncompressedDataIndexSize()) {
hash = (37 * hash) + UNCOMPRESSED_DATA_INDEX_SIZE_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getUncompressedDataIndexSize());
}
if (hasTotalUncompressedBytes()) {
hash = (37 * hash) + TOTAL_UNCOMPRESSED_BYTES_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getTotalUncompressedBytes());
}
if (hasDataIndexCount()) {
@@ -1507,7 +1507,7 @@ public final class HFileProtos {
}
if (hasEntryCount()) {
hash = (37 * hash) + ENTRY_COUNT_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getEntryCount());
}
if (hasNumDataIndexLevels()) {
@@ -1516,12 +1516,12 @@ public final class HFileProtos {
}
if (hasFirstDataBlockOffset()) {
hash = (37 * hash) + FIRST_DATA_BLOCK_OFFSET_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getFirstDataBlockOffset());
}
if (hasLastDataBlockOffset()) {
hash = (37 * hash) + LAST_DATA_BLOCK_OFFSET_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getLastDataBlockOffset());
}
if (hasComparatorClassName()) {
@@ -1542,61 +1542,61 @@ public final class HFileProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileTrailerProto parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileTrailerProto parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileTrailerProto parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileTrailerProto parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileTrailerProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileTrailerProto parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileTrailerProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileTrailerProto parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileTrailerProto parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileTrailerProto parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -1614,7 +1614,7 @@ public final class HFileProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1626,15 +1626,15 @@ public final class HFileProtos {
* Protobuf type {@code hbase.pb.FileTrailerProto}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.FileTrailerProto)
org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileTrailerProtoOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.internal_static_hbase_pb_FileTrailerProto_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.internal_static_hbase_pb_FileTrailerProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1647,12 +1647,12 @@ public final class HFileProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -1682,12 +1682,12 @@ public final class HFileProtos {
bitField0_ = (bitField0_ & ~0x00000400);
compressionCodec_ = 0;
bitField0_ = (bitField0_ & ~0x00000800);
- encryptionKey_ = com.google.protobuf.ByteString.EMPTY;
+ encryptionKey_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00001000);
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.internal_static_hbase_pb_FileTrailerProto_descriptor;
}
@@ -1769,29 +1769,29 @@ public final class HFileProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileTrailerProto) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileTrailerProto)other);
} else {
@@ -1853,13 +1853,13 @@ public final class HFileProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileTrailerProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.HFileProtos.FileTrailerProto) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -2204,8 +2204,8 @@ public final class HFileProtos {
public java.lang.String getComparatorClassName() {
java.lang.Object ref = comparatorClassName_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
comparatorClassName_ = s;
@@ -2218,17 +2218,17 @@ public final class HFileProtos {
/**
* <code>optional string comparator_class_name = 11;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getComparatorClassNameBytes() {
java.lang.Object ref = comparatorClassName_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
comparatorClassName_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -2257,7 +2257,7 @@ public final class HFileProtos {
* <code>optional string comparator_class_name = 11;</code>
*/
public Builder setComparatorClassNameBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -2299,7 +2299,7 @@ public final class HFileProtos {
return this;
}
- private com.google.protobuf.ByteString encryptionKey_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString encryptionKey_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>optional bytes encryption_key = 13;</code>
*/
@@ -2309,13 +2309,13 @@ public final class HFileProtos {
/**
* <code>optional bytes encryption_key = 13;</code>
*/
- public com.google.protobuf.ByteString getEncryptionKey() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getEncryptionKey() {
return encryptionKey_;
}
/**
* <code>optional bytes encryption_key = 13;</code>
*/
- public Builder setEncryptionKey(com.google.protobuf.ByteString value) {
+ public Builder setEncryptionKey(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -2334,12 +2334,12 @@ public final class HFileProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -2357,22 +2357,22 @@ public final class HFileProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<FileTrailerProto>
- PARSER = new com.google.protobuf.AbstractParser<FileTrailerProto>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<FileTrailerProto>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<FileTrailerProto>() {
public FileTrailerProto parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new FileTrailerProto(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<FileTrailerProto> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<FileTrailerProto> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<FileTrailerProto> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<FileTrailerProto> getParserForType() {
return PARSER;
}
@@ -2382,22 +2382,22 @@ public final class HFileProtos {
}
- private static final com.google.protobuf.Descriptors.Descriptor
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
internal_static_hbase_pb_FileInfoProto_descriptor;
private static final
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hbase_pb_FileInfoProto_fieldAccessorTable;
- private static final com.google.protobuf.Descriptors.Descriptor
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
internal_static_hbase_pb_FileTrailerProto_descriptor;
private static final
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hbase_pb_FileTrailerProto_fieldAccessorTable;
- public static com.google.protobuf.Descriptors.FileDescriptor
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
getDescriptor() {
return descriptor;
}
- private static com.google.protobuf.Descriptors.FileDescriptor
+ private static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
descriptor;
static {
java.lang.String[] descriptorData = {
@@ -2417,29 +2417,29 @@ public final class HFileProtos {
"oop.hbase.shaded.protobuf.generatedB\013HFi" +
"leProtosH\001\210\001\001\240\001\001"
};
- com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
- new com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
- public com.google.protobuf.ExtensionRegistry assignDescriptors(
- com.google.protobuf.Descriptors.FileDescriptor root) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
descriptor = root;
return null;
}
};
- com.google.protobuf.Descriptors.FileDescriptor
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
.internalBuildGeneratedFileFrom(descriptorData,
- new com.google.protobuf.Descriptors.FileDescriptor[] {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.getDescriptor(),
}, assigner);
internal_static_hbase_pb_FileInfoProto_descriptor =
getDescriptor().getMessageTypes().get(0);
internal_static_hbase_pb_FileInfoProto_fieldAccessorTable = new
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hbase_pb_FileInfoProto_descriptor,
new java.lang.String[] { "MapEntry", });
internal_static_hbase_pb_FileTrailerProto_descriptor =
getDescriptor().getMessageTypes().get(1);
internal_static_hbase_pb_FileTrailerProto_fieldAccessorTable = new
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hbase_pb_FileTrailerProto_descriptor,
new java.lang.String[] { "FileInfoOffset", "LoadOnOpenDataOffset", "UncompressedDataIndexSize", "TotalUncompressedBytes", "DataIndexCount", "MetaIndexCount", "EntryCount", "NumDataIndexLevels", "FirstDataBlockOffset", "LastDataBlockOffset", "ComparatorClassName", "CompressionCodec", "EncryptionKey", });
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.getDescriptor();
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/LoadBalancerProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/LoadBalancerProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/LoadBalancerProtos.java
index 2226f76..0b941d8 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/LoadBalancerProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/LoadBalancerProtos.java
@@ -6,17 +6,17 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class LoadBalancerProtos {
private LoadBalancerProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
public interface LoadBalancerStateOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.LoadBalancerState)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>optional bool balancer_on = 1;</code>
@@ -31,11 +31,11 @@ public final class LoadBalancerProtos {
* Protobuf type {@code hbase.pb.LoadBalancerState}
*/
public static final class LoadBalancerState extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.LoadBalancerState)
LoadBalancerStateOrBuilder {
// Use LoadBalancerState.newBuilder() to construct.
- private LoadBalancerState(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private LoadBalancerState(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private LoadBalancerState() {
@@ -43,18 +43,18 @@ public final class LoadBalancerProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private LoadBalancerState(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -77,22 +77,22 @@ public final class LoadBalancerProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.LoadBalancerProtos.internal_static_hbase_pb_LoadBalancerState_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.LoadBalancerProtos.internal_static_hbase_pb_LoadBalancerState_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -125,7 +125,7 @@ public final class LoadBalancerProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBool(1, balancerOn_);
@@ -139,7 +139,7 @@ public final class LoadBalancerProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBoolSize(1, balancerOn_);
}
size += unknownFields.getSerializedSize();
@@ -177,7 +177,7 @@ public final class LoadBalancerProtos {
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasBalancerOn()) {
hash = (37 * hash) + BALANCER_ON_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean(
getBalancerOn());
}
hash = (29 * hash) + unknownFields.hashCode();
@@ -186,61 +186,61 @@ public final class LoadBalancerProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.LoadBalancerProtos.LoadBalancerState parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.LoadBalancerProtos.LoadBalancerState parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.LoadBalancerProtos.LoadBalancerState parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.LoadBalancerProtos.LoadBalancerState parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.LoadBalancerProtos.LoadBalancerState parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.LoadBalancerProtos.LoadBalancerState parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.LoadBalancerProtos.LoadBalancerState parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.LoadBalancerProtos.LoadBalancerState parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.LoadBalancerProtos.LoadBalancerState parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.LoadBalancerProtos.LoadBalancerState parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -258,7 +258,7 @@ public final class LoadBalancerProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -266,15 +266,15 @@ public final class LoadBalancerProtos {
* Protobuf type {@code hbase.pb.LoadBalancerState}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.LoadBalancerState)
org.apache.hadoop.hbase.shaded.protobuf.generated.LoadBalancerProtos.LoadBalancerStateOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.LoadBalancerProtos.internal_static_hbase_pb_LoadBalancerState_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.LoadBalancerProtos.internal_static_hbase_pb_LoadBalancerState_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -287,12 +287,12 @@ public final class LoadBalancerProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -303,7 +303,7 @@ public final class LoadBalancerProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.LoadBalancerProtos.internal_static_hbase_pb_LoadBalancerState_descriptor;
}
@@ -337,29 +337,29 @@ public final class LoadBalancerProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.LoadBalancerProtos.LoadBalancerState) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.LoadBalancerProtos.LoadBalancerState)other);
} else {
@@ -383,13 +383,13 @@ public final class LoadBalancerProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.LoadBalancerProtos.LoadBalancerState parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.LoadBalancerProtos.LoadBalancerState) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -433,12 +433,12 @@ public final class LoadBalancerProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -456,22 +456,22 @@ public final class LoadBalancerProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<LoadBalancerState>
- PARSER = new com.google.protobuf.AbstractParser<LoadBalancerState>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<LoadBalancerState>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<LoadBalancerState>() {
public LoadBalancerState parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new LoadBalancerState(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<LoadBalancerState> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<LoadBalancerState> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<LoadBalancerState> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<LoadBalancerState> getParserForType() {
return PARSER;
}
@@ -481,17 +481,17 @@ public final class LoadBalancerProtos {
}
- private static final com.google.protobuf.Descriptors.Descriptor
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
internal_static_hbase_pb_LoadBalancerState_descriptor;
private static final
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hbase_pb_LoadBalancerState_fieldAccessorTable;
- public static com.google.protobuf.Descriptors.FileDescriptor
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
getDescriptor() {
return descriptor;
}
- private static com.google.protobuf.Descriptors.FileDescriptor
+ private static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
descriptor;
static {
java.lang.String[] descriptorData = {
@@ -500,22 +500,22 @@ public final class LoadBalancerProtos {
"apache.hadoop.hbase.shaded.protobuf.gene" +
"ratedB\022LoadBalancerProtosH\001\240\001\001"
};
- com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
- new com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
- public com.google.protobuf.ExtensionRegistry assignDescriptors(
- com.google.protobuf.Descriptors.FileDescriptor root) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
descriptor = root;
return null;
}
};
- com.google.protobuf.Descriptors.FileDescriptor
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
.internalBuildGeneratedFileFrom(descriptorData,
- new com.google.protobuf.Descriptors.FileDescriptor[] {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
}, assigner);
internal_static_hbase_pb_LoadBalancerState_descriptor =
getDescriptor().getMessageTypes().get(0);
internal_static_hbase_pb_LoadBalancerState_fieldAccessorTable = new
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hbase_pb_LoadBalancerState_descriptor,
new java.lang.String[] { "BalancerOn", });
}
[22/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ClusterIdProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ClusterIdProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ClusterIdProtos.java
index bd132af..326f217 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ClusterIdProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ClusterIdProtos.java
@@ -6,17 +6,17 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class ClusterIdProtos {
private ClusterIdProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
public interface ClusterIdOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.ClusterId)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <pre>
@@ -41,7 +41,7 @@ public final class ClusterIdProtos {
*
* <code>required string cluster_id = 1;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getClusterIdBytes();
}
/**
@@ -54,11 +54,11 @@ public final class ClusterIdProtos {
* Protobuf type {@code hbase.pb.ClusterId}
*/
public static final class ClusterId extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.ClusterId)
ClusterIdOrBuilder {
// Use ClusterId.newBuilder() to construct.
- private ClusterId(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private ClusterId(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private ClusterId() {
@@ -66,18 +66,18 @@ public final class ClusterIdProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private ClusterId(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -94,29 +94,29 @@ public final class ClusterIdProtos {
break;
}
case 10: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000001;
clusterId_ = bs;
break;
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterIdProtos.internal_static_hbase_pb_ClusterId_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterIdProtos.internal_static_hbase_pb_ClusterId_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -148,8 +148,8 @@ public final class ClusterIdProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
clusterId_ = s;
@@ -164,17 +164,17 @@ public final class ClusterIdProtos {
*
* <code>required string cluster_id = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getClusterIdBytes() {
java.lang.Object ref = clusterId_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
clusterId_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -192,10 +192,10 @@ public final class ClusterIdProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 1, clusterId_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, clusterId_);
}
unknownFields.writeTo(output);
}
@@ -206,7 +206,7 @@ public final class ClusterIdProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, clusterId_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, clusterId_);
}
size += unknownFields.getSerializedSize();
memoizedSize = size;
@@ -251,61 +251,61 @@ public final class ClusterIdProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterIdProtos.ClusterId parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterIdProtos.ClusterId parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterIdProtos.ClusterId parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterIdProtos.ClusterId parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterIdProtos.ClusterId parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterIdProtos.ClusterId parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterIdProtos.ClusterId parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterIdProtos.ClusterId parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterIdProtos.ClusterId parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterIdProtos.ClusterId parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -323,7 +323,7 @@ public final class ClusterIdProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -337,15 +337,15 @@ public final class ClusterIdProtos {
* Protobuf type {@code hbase.pb.ClusterId}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.ClusterId)
org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterIdProtos.ClusterIdOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterIdProtos.internal_static_hbase_pb_ClusterId_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterIdProtos.internal_static_hbase_pb_ClusterId_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -358,12 +358,12 @@ public final class ClusterIdProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -374,7 +374,7 @@ public final class ClusterIdProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterIdProtos.internal_static_hbase_pb_ClusterId_descriptor;
}
@@ -408,29 +408,29 @@ public final class ClusterIdProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterIdProtos.ClusterId) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterIdProtos.ClusterId)other);
} else {
@@ -459,13 +459,13 @@ public final class ClusterIdProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterIdProtos.ClusterId parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterIdProtos.ClusterId) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -498,8 +498,8 @@ public final class ClusterIdProtos {
public java.lang.String getClusterId() {
java.lang.Object ref = clusterId_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
clusterId_ = s;
@@ -516,17 +516,17 @@ public final class ClusterIdProtos {
*
* <code>required string cluster_id = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getClusterIdBytes() {
java.lang.Object ref = clusterId_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
clusterId_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -567,7 +567,7 @@ public final class ClusterIdProtos {
* <code>required string cluster_id = 1;</code>
*/
public Builder setClusterIdBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -577,12 +577,12 @@ public final class ClusterIdProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -600,22 +600,22 @@ public final class ClusterIdProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<ClusterId>
- PARSER = new com.google.protobuf.AbstractParser<ClusterId>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ClusterId>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<ClusterId>() {
public ClusterId parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new ClusterId(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<ClusterId> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ClusterId> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<ClusterId> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ClusterId> getParserForType() {
return PARSER;
}
@@ -625,17 +625,17 @@ public final class ClusterIdProtos {
}
- private static final com.google.protobuf.Descriptors.Descriptor
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
internal_static_hbase_pb_ClusterId_descriptor;
private static final
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hbase_pb_ClusterId_fieldAccessorTable;
- public static com.google.protobuf.Descriptors.FileDescriptor
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
getDescriptor() {
return descriptor;
}
- private static com.google.protobuf.Descriptors.FileDescriptor
+ private static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
descriptor;
static {
java.lang.String[] descriptorData = {
@@ -644,22 +644,22 @@ public final class ClusterIdProtos {
"p.hbase.shaded.protobuf.generatedB\017Clust" +
"erIdProtosH\001\240\001\001"
};
- com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
- new com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
- public com.google.protobuf.ExtensionRegistry assignDescriptors(
- com.google.protobuf.Descriptors.FileDescriptor root) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
descriptor = root;
return null;
}
};
- com.google.protobuf.Descriptors.FileDescriptor
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
.internalBuildGeneratedFileFrom(descriptorData,
- new com.google.protobuf.Descriptors.FileDescriptor[] {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
}, assigner);
internal_static_hbase_pb_ClusterId_descriptor =
getDescriptor().getMessageTypes().get(0);
internal_static_hbase_pb_ClusterId_fieldAccessorTable = new
- com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hbase_pb_ClusterId_descriptor,
new java.lang.String[] { "ClusterId", });
}
[29/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
Posted by st...@apache.org.
HBASE-16741 Amend the generate protobufs out-of-band build step to include shade, pulling in protobuf source and a hook for patching protobuf
Removed ByteStringer from hbase-protocol-shaded. Use the protobuf-3.1.0
trick directly instead. Makes stuff cleaner. All under 'shaded' dir is
now generated.
Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/32be831c
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/32be831c
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/32be831c
Branch: refs/heads/HBASE-16264
Commit: 32be831ce56beab404d463cd7ada54a98f9e99f8
Parents: 35cf550
Author: stack <st...@apache.org>
Authored: Fri Sep 30 14:54:58 2016 -0700
Committer: stack <st...@apache.org>
Committed: Fri Sep 30 14:54:58 2016 -0700
----------------------------------------------------------------------
.../org/apache/hadoop/hbase/HRegionInfo.java | 6 +-
.../hbase/filter/ColumnPaginationFilter.java | 4 +-
.../hadoop/hbase/filter/ColumnPrefixFilter.java | 4 +-
.../hadoop/hbase/filter/ColumnRangeFilter.java | 10 +-
.../hbase/filter/DependentColumnFilter.java | 6 +-
.../FirstKeyValueMatchingQualifiersFilter.java | 4 +-
.../hadoop/hbase/filter/FuzzyRowFilter.java | 8 +-
.../hbase/filter/InclusiveStopFilter.java | 7 +-
.../hbase/filter/MultiRowRangeFilter.java | 9 +-
.../filter/MultipleColumnPrefixFilter.java | 7 +-
.../hadoop/hbase/filter/PrefixFilter.java | 4 +-
.../hbase/filter/SingleColumnValueFilter.java | 8 +-
.../hadoop/hbase/security/EncryptionUtil.java | 8 +-
.../hbase/shaded/protobuf/ProtobufUtil.java | 106 +-
.../hbase/shaded/protobuf/RequestConverter.java | 36 +-
.../apache/hadoop/hbase/zookeeper/ZKUtil.java | 11 +-
.../hbase/client/TestClientNoCluster.java | 17 +-
.../hadoop/hbase/procedure2/ProcedureUtil.java | 8 +-
hbase-protocol-shaded/README.txt | 77 +-
hbase-protocol-shaded/pom.xml | 558 +-
.../protobuf/generated/TestProcedureProtos.java | 178 +-
.../ipc/protobuf/generated/TestProtos.java | 858 +-
.../generated/TestRpcServiceProtos.java | 226 +-
.../shaded/protobuf/generated/AdminProtos.java | 5566 +++---
.../shaded/protobuf/generated/CellProtos.java | 474 +-
.../shaded/protobuf/generated/ClientProtos.java | 6168 +++---
.../protobuf/generated/ClusterIdProtos.java | 178 +-
.../protobuf/generated/ClusterStatusProtos.java | 1620 +-
.../protobuf/generated/ComparatorProtos.java | 1378 +-
.../protobuf/generated/EncryptionProtos.java | 234 +-
.../protobuf/generated/ErrorHandlingProtos.java | 632 +-
.../shaded/protobuf/generated/FSProtos.java | 336 +-
.../shaded/protobuf/generated/FilterProtos.java | 4210 ++---
.../shaded/protobuf/generated/HBaseProtos.java | 4144 ++---
.../shaded/protobuf/generated/HFileProtos.java | 366 +-
.../protobuf/generated/LoadBalancerProtos.java | 148 +-
.../protobuf/generated/MapReduceProtos.java | 326 +-
.../generated/MasterProcedureProtos.java | 2986 +--
.../shaded/protobuf/generated/MasterProtos.java | 16494 ++++++++---------
.../protobuf/generated/ProcedureProtos.java | 1246 +-
.../shaded/protobuf/generated/QuotaProtos.java | 760 +-
.../shaded/protobuf/generated/RPCProtos.java | 1128 +-
.../generated/RegionNormalizerProtos.java | 148 +-
.../generated/RegionServerStatusProtos.java | 1804 +-
.../protobuf/generated/SnapshotProtos.java | 870 +-
.../protobuf/generated/TracingProtos.java | 152 +-
.../shaded/protobuf/generated/WALProtos.java | 2030 +-
.../protobuf/generated/ZooKeeperProtos.java | 1744 +-
.../hadoop/hbase/shaded/util/ByteStringer.java | 48 -
.../src/main/protobuf/RSGroup.proto | 34 -
.../src/main/protobuf/RSGroupAdmin.proto | 136 -
.../apache/hadoop/hbase/codec/MessageCodec.java | 11 +-
.../org/apache/hadoop/hbase/io/Reference.java | 4 +-
.../hadoop/hbase/io/hfile/FixedFileTrailer.java | 8 +-
.../org/apache/hadoop/hbase/io/hfile/HFile.java | 12 +-
.../hadoop/hbase/master/MasterRpcServices.java | 4 +-
.../procedure/DeleteColumnFamilyProcedure.java | 4 +-
.../hbase/protobuf/ReplicationProtbufUtil.java | 10 +-
.../hadoop/hbase/regionserver/HRegion.java | 35 +-
.../hbase/regionserver/HRegionServer.java | 8 +-
.../wal/AbstractProtobufLogWriter.java | 8 +-
.../hbase/snapshot/SnapshotManifestV1.java | 12 +-
.../hbase/snapshot/SnapshotManifestV2.java | 5 +-
.../hbase/regionserver/TestHRegionInfo.java | 10 +-
.../regionserver/TestHRegionReplayEvents.java | 41 +-
.../hbase/regionserver/TestPriorityRpc.java | 7 +-
.../regionserver/TestReplicationSink.java | 9 +-
.../TestReplicationSourceManager.java | 4 +-
.../hbase/snapshot/TestSnapshotManifest.java | 4 +-
pom.xml | 3 +-
70 files changed, 28784 insertions(+), 28935 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java
index 09e53ce..da0d941 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java
@@ -33,11 +33,11 @@ import org.apache.hadoop.hbase.client.RegionReplicaUtil;
import org.apache.hadoop.hbase.KeyValue.KVComparator;
import org.apache.hadoop.hbase.exceptions.DeserializationException;
import org.apache.hadoop.hbase.master.RegionState;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo;
import org.apache.hadoop.hbase.util.ByteArrayHashKey;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.hbase.util.HashKey;
import org.apache.hadoop.hbase.util.JenkinsHash;
@@ -873,10 +873,10 @@ public class HRegionInfo implements Comparable<HRegionInfo> {
builder.setTableName(ProtobufUtil.toProtoTableName(info.getTable()));
builder.setRegionId(info.getRegionId());
if (info.getStartKey() != null) {
- builder.setStartKey(ByteStringer.wrap(info.getStartKey()));
+ builder.setStartKey(UnsafeByteOperations.unsafeWrap(info.getStartKey()));
}
if (info.getEndKey() != null) {
- builder.setEndKey(ByteStringer.wrap(info.getEndKey()));
+ builder.setEndKey(UnsafeByteOperations.unsafeWrap(info.getEndKey()));
}
builder.setOffline(info.isOffline());
builder.setSplit(info.isSplit());
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java
index d2255d1..7d4571e 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java
@@ -28,11 +28,11 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience;
import org.apache.hadoop.hbase.classification.InterfaceStability;
import org.apache.hadoop.hbase.exceptions.DeserializationException;
import org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.util.Bytes;
import com.google.common.base.Preconditions;
import org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
/**
* A filter, based on the ColumnCountGetFilter, takes two arguments: limit and offset.
@@ -172,7 +172,7 @@ public class ColumnPaginationFilter extends FilterBase {
builder.setOffset(this.offset);
}
if (this.columnOffset != null) {
- builder.setColumnOffset(ByteStringer.wrap(this.columnOffset));
+ builder.setColumnOffset(UnsafeByteOperations.unsafeWrap(this.columnOffset));
}
return builder.build().toByteArray();
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java
index 46fb82d..806841f 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java
@@ -30,11 +30,11 @@ import org.apache.hadoop.hbase.classification.InterfaceStability;
import org.apache.hadoop.hbase.exceptions.DeserializationException;
import org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
import org.apache.hadoop.hbase.util.ByteBufferUtils;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.util.Bytes;
import com.google.common.base.Preconditions;
import org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
/**
* This filter is used for selecting only those keys with columns that matches
@@ -112,7 +112,7 @@ public class ColumnPrefixFilter extends FilterBase {
public byte [] toByteArray() {
FilterProtos.ColumnPrefixFilter.Builder builder =
FilterProtos.ColumnPrefixFilter.newBuilder();
- if (this.prefix != null) builder.setPrefix(ByteStringer.wrap(this.prefix));
+ if (this.prefix != null) builder.setPrefix(UnsafeByteOperations.unsafeWrap(this.prefix));
return builder.build().toByteArray();
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java
index 78d7fb5..99f9926 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java
@@ -30,12 +30,12 @@ import org.apache.hadoop.hbase.CellUtil;
import org.apache.hadoop.hbase.classification.InterfaceAudience;
import org.apache.hadoop.hbase.classification.InterfaceStability;
import org.apache.hadoop.hbase.exceptions.DeserializationException;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
import org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.util.Bytes;
import com.google.common.base.Preconditions;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
/**
* This filter is used for selecting only those keys with columns that are
@@ -175,9 +175,11 @@ public class ColumnRangeFilter extends FilterBase {
public byte [] toByteArray() {
FilterProtos.ColumnRangeFilter.Builder builder =
FilterProtos.ColumnRangeFilter.newBuilder();
- if (this.minColumn != null) builder.setMinColumn(ByteStringer.wrap(this.minColumn));
+ if (this.minColumn != null) builder.setMinColumn(
+ UnsafeByteOperations.unsafeWrap(this.minColumn));
builder.setMinColumnInclusive(this.minColumnInclusive);
- if (this.maxColumn != null) builder.setMaxColumn(ByteStringer.wrap(this.maxColumn));
+ if (this.maxColumn != null) builder.setMaxColumn(
+ UnsafeByteOperations.unsafeWrap(this.maxColumn));
builder.setMaxColumnInclusive(this.maxColumnInclusive);
return builder.build().toByteArray();
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java
index f273e63..287a090 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java
@@ -32,12 +32,12 @@ import org.apache.hadoop.hbase.classification.InterfaceStability;
import org.apache.hadoop.hbase.exceptions.DeserializationException;
import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
import org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.util.Bytes;
import com.google.common.base.Preconditions;
import org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
/**
* A filter for adding inter-column timestamp matching
@@ -222,10 +222,10 @@ public class DependentColumnFilter extends CompareFilter {
FilterProtos.DependentColumnFilter.newBuilder();
builder.setCompareFilter(super.convert());
if (this.columnFamily != null) {
- builder.setColumnFamily(ByteStringer.wrap(this.columnFamily));
+ builder.setColumnFamily(UnsafeByteOperations.unsafeWrap(this.columnFamily));
}
if (this.columnQualifier != null) {
- builder.setColumnQualifier(ByteStringer.wrap(this.columnQualifier));
+ builder.setColumnQualifier(UnsafeByteOperations.unsafeWrap(this.columnQualifier));
}
builder.setDropDependentColumn(this.dropDependentColumn);
return builder.build().toByteArray();
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyValueMatchingQualifiersFilter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyValueMatchingQualifiersFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyValueMatchingQualifiersFilter.java
index 98513f5..82d6c57 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyValueMatchingQualifiersFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyValueMatchingQualifiersFilter.java
@@ -27,11 +27,11 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience;
import org.apache.hadoop.hbase.classification.InterfaceStability;
import org.apache.hadoop.hbase.exceptions.DeserializationException;
import org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString;
import org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
/**
* The filter looks for the given columns in KeyValue. Once there is a match for
@@ -88,7 +88,7 @@ public class FirstKeyValueMatchingQualifiersFilter extends FirstKeyOnlyFilter {
FilterProtos.FirstKeyValueMatchingQualifiersFilter.Builder builder =
FilterProtos.FirstKeyValueMatchingQualifiersFilter.newBuilder();
for (byte[] qualifier : qualifiers) {
- if (qualifier != null) builder.addQualifiers(ByteStringer.wrap(qualifier));
+ if (qualifier != null) builder.addQualifiers(UnsafeByteOperations.unsafeWrap(qualifier));
}
return builder.build().toByteArray();
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java
index e8a75db..5fc12b9 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java
@@ -29,16 +29,16 @@ import org.apache.hadoop.hbase.CellUtil;
import org.apache.hadoop.hbase.classification.InterfaceAudience;
import org.apache.hadoop.hbase.classification.InterfaceStability;
import org.apache.hadoop.hbase.exceptions.DeserializationException;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
import org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.hbase.util.Pair;
import org.apache.hadoop.hbase.util.UnsafeAccess;
import org.apache.hadoop.hbase.util.UnsafeAvailChecker;
import com.google.common.annotations.VisibleForTesting;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
/**
* This is optimized version of a standard FuzzyRowFilter Filters data based on fuzzy row key.
@@ -257,8 +257,8 @@ public class FuzzyRowFilter extends FilterBase {
FilterProtos.FuzzyRowFilter.Builder builder = FilterProtos.FuzzyRowFilter.newBuilder();
for (Pair<byte[], byte[]> fuzzyData : fuzzyKeysData) {
BytesBytesPair.Builder bbpBuilder = BytesBytesPair.newBuilder();
- bbpBuilder.setFirst(ByteStringer.wrap(fuzzyData.getFirst()));
- bbpBuilder.setSecond(ByteStringer.wrap(fuzzyData.getSecond()));
+ bbpBuilder.setFirst(UnsafeByteOperations.unsafeWrap(fuzzyData.getFirst()));
+ bbpBuilder.setSecond(UnsafeByteOperations.unsafeWrap(fuzzyData.getSecond()));
builder.addFuzzyKeysData(bbpBuilder);
}
return builder.build().toByteArray();
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java
index 53e116e..7aa807c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java
@@ -26,12 +26,12 @@ import org.apache.hadoop.hbase.CellComparator;
import org.apache.hadoop.hbase.classification.InterfaceAudience;
import org.apache.hadoop.hbase.classification.InterfaceStability;
import org.apache.hadoop.hbase.exceptions.DeserializationException;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
import org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.util.Bytes;
import com.google.common.base.Preconditions;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
/**
* A Filter that stops after the given row. There is no "RowStopFilter" because
@@ -84,7 +84,8 @@ public class InclusiveStopFilter extends FilterBase {
public byte [] toByteArray() {
FilterProtos.InclusiveStopFilter.Builder builder =
FilterProtos.InclusiveStopFilter.newBuilder();
- if (this.stopRowKey != null) builder.setStopRowKey(ByteStringer.wrap(this.stopRowKey));
+ if (this.stopRowKey != null) builder.setStopRowKey(
+ UnsafeByteOperations.unsafeWrap(this.stopRowKey));
return builder.build().toByteArray();
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java
index 9e59e9a..328498d 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java
@@ -28,12 +28,11 @@ import org.apache.hadoop.hbase.HConstants;
import org.apache.hadoop.hbase.classification.InterfaceAudience;
import org.apache.hadoop.hbase.classification.InterfaceStability;
import org.apache.hadoop.hbase.exceptions.DeserializationException;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
import org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.util.Bytes;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
-
/**
* Filter to support scan multiple row key ranges. It can construct the row key ranges from the
* passed list which can be accessed by each region server.
@@ -151,10 +150,10 @@ public class MultiRowRangeFilter extends FilterBase {
if (range != null) {
FilterProtos.RowRange.Builder rangebuilder = FilterProtos.RowRange.newBuilder();
if (range.startRow != null)
- rangebuilder.setStartRow(ByteStringer.wrap(range.startRow));
+ rangebuilder.setStartRow(UnsafeByteOperations.unsafeWrap(range.startRow));
rangebuilder.setStartRowInclusive(range.startRowInclusive);
if (range.stopRow != null)
- rangebuilder.setStopRow(ByteStringer.wrap(range.stopRow));
+ rangebuilder.setStopRow(UnsafeByteOperations.unsafeWrap(range.stopRow));
rangebuilder.setStopRowInclusive(range.stopRowInclusive);
range.isScan = Bytes.equals(range.startRow, range.stopRow) ? 1 : 0;
builder.addRowRangeList(rangebuilder.build());
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java
index cc97e75..9909d1f 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java
@@ -27,12 +27,11 @@ import org.apache.hadoop.hbase.CellUtil;
import org.apache.hadoop.hbase.classification.InterfaceAudience;
import org.apache.hadoop.hbase.classification.InterfaceStability;
import org.apache.hadoop.hbase.exceptions.DeserializationException;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
import org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.util.Bytes;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
-
/**
* This filter is used for selecting only those keys with columns that matches
* a particular prefix. For example, if prefix is 'an', it will pass keys will
@@ -118,7 +117,7 @@ public class MultipleColumnPrefixFilter extends FilterBase {
FilterProtos.MultipleColumnPrefixFilter.Builder builder =
FilterProtos.MultipleColumnPrefixFilter.newBuilder();
for (byte [] element : sortedPrefixes) {
- if (element != null) builder.addSortedPrefixes(ByteStringer.wrap(element));
+ if (element != null) builder.addSortedPrefixes(UnsafeByteOperations.unsafeWrap(element));
}
return builder.build().toByteArray();
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java
index d651f02..9bc6236 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java
@@ -28,11 +28,11 @@ import org.apache.hadoop.hbase.classification.InterfaceStability;
import org.apache.hadoop.hbase.exceptions.DeserializationException;
import org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
import org.apache.hadoop.hbase.util.ByteBufferUtils;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.util.Bytes;
import com.google.common.base.Preconditions;
import org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
/**
* Pass results that have same row prefix.
@@ -108,7 +108,7 @@ public class PrefixFilter extends FilterBase {
public byte [] toByteArray() {
FilterProtos.PrefixFilter.Builder builder =
FilterProtos.PrefixFilter.newBuilder();
- if (this.prefix != null) builder.setPrefix(ByteStringer.wrap(this.prefix));
+ if (this.prefix != null) builder.setPrefix(UnsafeByteOperations.unsafeWrap(this.prefix));
return builder.build().toByteArray();
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java
index 91149f5..3bb80cb 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java
@@ -29,12 +29,12 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience;
import org.apache.hadoop.hbase.classification.InterfaceStability;
import org.apache.hadoop.hbase.exceptions.DeserializationException;
import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
import org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.CompareType;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.util.Bytes;
import com.google.common.base.Preconditions;
@@ -307,10 +307,10 @@ public class SingleColumnValueFilter extends FilterBase {
FilterProtos.SingleColumnValueFilter.Builder builder =
FilterProtos.SingleColumnValueFilter.newBuilder();
if (this.columnFamily != null) {
- builder.setColumnFamily(ByteStringer.wrap(this.columnFamily));
+ builder.setColumnFamily(UnsafeByteOperations.unsafeWrap(this.columnFamily));
}
if (this.columnQualifier != null) {
- builder.setColumnQualifier(ByteStringer.wrap(this.columnQualifier));
+ builder.setColumnQualifier(UnsafeByteOperations.unsafeWrap(this.columnQualifier));
}
HBaseProtos.CompareType compareOp = CompareType.valueOf(this.compareOp.name());
builder.setCompareOp(compareOp);
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-client/src/main/java/org/apache/hadoop/hbase/security/EncryptionUtil.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/EncryptionUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/EncryptionUtil.java
index 29ef57d..b5009e0 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/EncryptionUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/EncryptionUtil.java
@@ -35,8 +35,8 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience;
import org.apache.hadoop.hbase.classification.InterfaceStability;
import org.apache.hadoop.hbase.io.crypto.Cipher;
import org.apache.hadoop.hbase.io.crypto.Encryption;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
import org.apache.hadoop.hbase.shaded.protobuf.generated.EncryptionProtos;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.util.Bytes;
/**
@@ -94,15 +94,15 @@ public final class EncryptionUtil {
if (cipher.getIvLength() > 0) {
iv = new byte[cipher.getIvLength()];
RNG.nextBytes(iv);
- builder.setIv(ByteStringer.wrap(iv));
+ builder.setIv(UnsafeByteOperations.unsafeWrap(iv));
}
byte[] keyBytes = key.getEncoded();
builder.setLength(keyBytes.length);
- builder.setHash(ByteStringer.wrap(Encryption.hash128(keyBytes)));
+ builder.setHash(UnsafeByteOperations.unsafeWrap(Encryption.hash128(keyBytes)));
ByteArrayOutputStream out = new ByteArrayOutputStream();
Encryption.encryptWithSubjectKey(out, new ByteArrayInputStream(keyBytes), subject,
conf, cipher, iv);
- builder.setData(ByteStringer.wrap(out.toByteArray()));
+ builder.setData(UnsafeByteOperations.unsafeWrap(out.toByteArray()));
// Build and return the protobuf message
out.reset();
builder.build().writeDelimitedTo(out);
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java
index 5a56ee3..2f72eaa 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java
@@ -102,6 +102,7 @@ import org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController;
import org.apache.hadoop.hbase.shaded.com.google.protobuf.Service;
import org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
import org.apache.hadoop.hbase.shaded.com.google.protobuf.TextFormat;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionRequest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionResponse;
@@ -159,7 +160,6 @@ import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.FlushDescript
import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.RegionEventDescriptor;
import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.RegionEventDescriptor.EventType;
import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.StoreDescriptor;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.util.Addressing;
import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.hbase.util.DynamicClassLoader;
@@ -311,7 +311,7 @@ public final class ProtobufUtil {
public static ComparatorProtos.ByteArrayComparable toByteArrayComparable(final byte [] value) {
ComparatorProtos.ByteArrayComparable.Builder builder =
ComparatorProtos.ByteArrayComparable.newBuilder();
- if (value != null) builder.setValue(ByteStringer.wrap(value));
+ if (value != null) builder.setValue(UnsafeByteOperations.unsafeWrap(value));
return builder.build();
}
@@ -949,7 +949,7 @@ public final class ProtobufUtil {
scanBuilder.setMaxVersions(scan.getMaxVersions());
for (Entry<byte[], TimeRange> cftr : scan.getColumnFamilyTimeRange().entrySet()) {
HBaseProtos.ColumnFamilyTimeRange.Builder b = HBaseProtos.ColumnFamilyTimeRange.newBuilder();
- b.setColumnFamily(ByteStringer.wrap(cftr.getKey()));
+ b.setColumnFamily(UnsafeByteOperations.unsafeWrap(cftr.getKey()));
b.setTimeRange(timeRangeToProto(cftr.getValue()));
scanBuilder.addCfTimeRange(b);
}
@@ -966,17 +966,17 @@ public final class ProtobufUtil {
NameBytesPair.Builder attributeBuilder = NameBytesPair.newBuilder();
for (Map.Entry<String, byte[]> attribute: attributes.entrySet()) {
attributeBuilder.setName(attribute.getKey());
- attributeBuilder.setValue(ByteStringer.wrap(attribute.getValue()));
+ attributeBuilder.setValue(UnsafeByteOperations.unsafeWrap(attribute.getValue()));
scanBuilder.addAttribute(attributeBuilder.build());
}
}
byte[] startRow = scan.getStartRow();
if (startRow != null && startRow.length > 0) {
- scanBuilder.setStartRow(ByteStringer.wrap(startRow));
+ scanBuilder.setStartRow(UnsafeByteOperations.unsafeWrap(startRow));
}
byte[] stopRow = scan.getStopRow();
if (stopRow != null && stopRow.length > 0) {
- scanBuilder.setStopRow(ByteStringer.wrap(stopRow));
+ scanBuilder.setStopRow(UnsafeByteOperations.unsafeWrap(stopRow));
}
if (scan.hasFilter()) {
scanBuilder.setFilter(ProtobufUtil.toFilter(scan.getFilter()));
@@ -985,12 +985,12 @@ public final class ProtobufUtil {
Column.Builder columnBuilder = Column.newBuilder();
for (Map.Entry<byte[],NavigableSet<byte []>>
family: scan.getFamilyMap().entrySet()) {
- columnBuilder.setFamily(ByteStringer.wrap(family.getKey()));
+ columnBuilder.setFamily(UnsafeByteOperations.unsafeWrap(family.getKey()));
NavigableSet<byte []> qualifiers = family.getValue();
columnBuilder.clearQualifier();
if (qualifiers != null && qualifiers.size() > 0) {
for (byte [] qualifier: qualifiers) {
- columnBuilder.addQualifier(ByteStringer.wrap(qualifier));
+ columnBuilder.addQualifier(UnsafeByteOperations.unsafeWrap(qualifier));
}
}
scanBuilder.addColumn(columnBuilder.build());
@@ -1112,7 +1112,7 @@ public final class ProtobufUtil {
final Get get) throws IOException {
ClientProtos.Get.Builder builder =
ClientProtos.Get.newBuilder();
- builder.setRow(ByteStringer.wrap(get.getRow()));
+ builder.setRow(UnsafeByteOperations.unsafeWrap(get.getRow()));
builder.setCacheBlocks(get.getCacheBlocks());
builder.setMaxVersions(get.getMaxVersions());
if (get.getFilter() != null) {
@@ -1120,7 +1120,7 @@ public final class ProtobufUtil {
}
for (Entry<byte[], TimeRange> cftr : get.getColumnFamilyTimeRange().entrySet()) {
HBaseProtos.ColumnFamilyTimeRange.Builder b = HBaseProtos.ColumnFamilyTimeRange.newBuilder();
- b.setColumnFamily(ByteStringer.wrap(cftr.getKey()));
+ b.setColumnFamily(UnsafeByteOperations.unsafeWrap(cftr.getKey()));
b.setTimeRange(timeRangeToProto(cftr.getValue()));
builder.addCfTimeRange(b);
}
@@ -1137,7 +1137,7 @@ public final class ProtobufUtil {
NameBytesPair.Builder attributeBuilder = NameBytesPair.newBuilder();
for (Map.Entry<String, byte[]> attribute: attributes.entrySet()) {
attributeBuilder.setName(attribute.getKey());
- attributeBuilder.setValue(ByteStringer.wrap(attribute.getValue()));
+ attributeBuilder.setValue(UnsafeByteOperations.unsafeWrap(attribute.getValue()));
builder.addAttribute(attributeBuilder.build());
}
}
@@ -1146,11 +1146,11 @@ public final class ProtobufUtil {
Map<byte[], NavigableSet<byte[]>> families = get.getFamilyMap();
for (Map.Entry<byte[], NavigableSet<byte[]>> family: families.entrySet()) {
NavigableSet<byte[]> qualifiers = family.getValue();
- columnBuilder.setFamily(ByteStringer.wrap(family.getKey()));
+ columnBuilder.setFamily(UnsafeByteOperations.unsafeWrap(family.getKey()));
columnBuilder.clearQualifier();
if (qualifiers != null && qualifiers.size() > 0) {
for (byte[] qualifier: qualifiers) {
- columnBuilder.addQualifier(ByteStringer.wrap(qualifier));
+ columnBuilder.addQualifier(UnsafeByteOperations.unsafeWrap(qualifier));
}
}
builder.addColumn(columnBuilder.build());
@@ -1190,7 +1190,7 @@ public final class ProtobufUtil {
*/
public static MutationProto toMutation(
final Increment increment, final MutationProto.Builder builder, long nonce) {
- builder.setRow(ByteStringer.wrap(increment.getRow()));
+ builder.setRow(UnsafeByteOperations.unsafeWrap(increment.getRow()));
builder.setMutateType(MutationType.INCREMENT);
builder.setDurability(toDurability(increment.getDurability()));
if (nonce != HConstants.NO_NONCE) {
@@ -1201,18 +1201,18 @@ public final class ProtobufUtil {
ColumnValue.Builder columnBuilder = ColumnValue.newBuilder();
QualifierValue.Builder valueBuilder = QualifierValue.newBuilder();
for (Map.Entry<byte[], List<Cell>> family: increment.getFamilyCellMap().entrySet()) {
- columnBuilder.setFamily(ByteStringer.wrap(family.getKey()));
+ columnBuilder.setFamily(UnsafeByteOperations.unsafeWrap(family.getKey()));
columnBuilder.clearQualifierValue();
List<Cell> values = family.getValue();
if (values != null && values.size() > 0) {
for (Cell cell: values) {
valueBuilder.clear();
- valueBuilder.setQualifier(ByteStringer.wrap(
+ valueBuilder.setQualifier(UnsafeByteOperations.unsafeWrap(
cell.getQualifierArray(), cell.getQualifierOffset(), cell.getQualifierLength()));
- valueBuilder.setValue(ByteStringer.wrap(
+ valueBuilder.setValue(UnsafeByteOperations.unsafeWrap(
cell.getValueArray(), cell.getValueOffset(), cell.getValueLength()));
if (cell.getTagsLength() > 0) {
- valueBuilder.setTags(ByteStringer.wrap(cell.getTagsArray(),
+ valueBuilder.setTags(UnsafeByteOperations.unsafeWrap(cell.getTagsArray(),
cell.getTagsOffset(), cell.getTagsLength()));
}
columnBuilder.addQualifierValue(valueBuilder.build());
@@ -1225,7 +1225,7 @@ public final class ProtobufUtil {
NameBytesPair.Builder attributeBuilder = NameBytesPair.newBuilder();
for (Map.Entry<String, byte[]> attribute : attributes.entrySet()) {
attributeBuilder.setName(attribute.getKey());
- attributeBuilder.setValue(ByteStringer.wrap(attribute.getValue()));
+ attributeBuilder.setValue(UnsafeByteOperations.unsafeWrap(attribute.getValue()));
builder.addAttribute(attributeBuilder.build());
}
}
@@ -1266,12 +1266,12 @@ public final class ProtobufUtil {
QualifierValue.Builder valueBuilder = QualifierValue.newBuilder();
for (Map.Entry<byte[],List<Cell>> family: mutation.getFamilyCellMap().entrySet()) {
columnBuilder.clear();
- columnBuilder.setFamily(ByteStringer.wrap(family.getKey()));
+ columnBuilder.setFamily(UnsafeByteOperations.unsafeWrap(family.getKey()));
for (Cell cell: family.getValue()) {
valueBuilder.clear();
- valueBuilder.setQualifier(ByteStringer.wrap(
+ valueBuilder.setQualifier(UnsafeByteOperations.unsafeWrap(
cell.getQualifierArray(), cell.getQualifierOffset(), cell.getQualifierLength()));
- valueBuilder.setValue(ByteStringer.wrap(
+ valueBuilder.setValue(UnsafeByteOperations.unsafeWrap(
cell.getValueArray(), cell.getValueOffset(), cell.getValueLength()));
valueBuilder.setTimestamp(cell.getTimestamp());
if (type == MutationType.DELETE || (type == MutationType.PUT && CellUtil.isDelete(cell))) {
@@ -1335,7 +1335,7 @@ public final class ProtobufUtil {
*/
private static MutationProto.Builder getMutationBuilderAndSetCommonFields(final MutationType type,
final Mutation mutation, MutationProto.Builder builder) {
- builder.setRow(ByteStringer.wrap(mutation.getRow()));
+ builder.setRow(UnsafeByteOperations.unsafeWrap(mutation.getRow()));
builder.setMutateType(type);
builder.setDurability(toDurability(mutation.getDurability()));
builder.setTimestamp(mutation.getTimeStamp());
@@ -1344,7 +1344,7 @@ public final class ProtobufUtil {
NameBytesPair.Builder attributeBuilder = NameBytesPair.newBuilder();
for (Map.Entry<String, byte[]> attribute: attributes.entrySet()) {
attributeBuilder.setName(attribute.getKey());
- attributeBuilder.setValue(ByteStringer.wrap(attribute.getValue()));
+ attributeBuilder.setValue(UnsafeByteOperations.unsafeWrap(attribute.getValue()));
builder.addAttribute(attributeBuilder.build());
}
}
@@ -1491,7 +1491,7 @@ public final class ProtobufUtil {
public static ComparatorProtos.Comparator toComparator(ByteArrayComparable comparator) {
ComparatorProtos.Comparator.Builder builder = ComparatorProtos.Comparator.newBuilder();
builder.setName(comparator.getClass().getName());
- builder.setSerializedComparator(ByteStringer.wrap(comparator.toByteArray()));
+ builder.setSerializedComparator(UnsafeByteOperations.unsafeWrap(comparator.toByteArray()));
return builder.build();
}
@@ -1555,7 +1555,7 @@ public final class ProtobufUtil {
public static FilterProtos.Filter toFilter(Filter filter) throws IOException {
FilterProtos.Filter.Builder builder = FilterProtos.Filter.newBuilder();
builder.setName(filter.getClass().getName());
- builder.setSerializedFilter(ByteStringer.wrap(filter.toByteArray()));
+ builder.setSerializedFilter(UnsafeByteOperations.unsafeWrap(filter.toByteArray()));
return builder.build();
}
@@ -2007,15 +2007,15 @@ public final class ProtobufUtil {
// Doing this is going to kill us if we do it for all data passed.
// St.Ack 20121205
CellProtos.Cell.Builder kvbuilder = CellProtos.Cell.newBuilder();
- kvbuilder.setRow(ByteStringer.wrap(kv.getRowArray(), kv.getRowOffset(),
+ kvbuilder.setRow(UnsafeByteOperations.unsafeWrap(kv.getRowArray(), kv.getRowOffset(),
kv.getRowLength()));
- kvbuilder.setFamily(ByteStringer.wrap(kv.getFamilyArray(),
+ kvbuilder.setFamily(UnsafeByteOperations.unsafeWrap(kv.getFamilyArray(),
kv.getFamilyOffset(), kv.getFamilyLength()));
- kvbuilder.setQualifier(ByteStringer.wrap(kv.getQualifierArray(),
+ kvbuilder.setQualifier(UnsafeByteOperations.unsafeWrap(kv.getQualifierArray(),
kv.getQualifierOffset(), kv.getQualifierLength()));
kvbuilder.setCellType(CellProtos.CellType.valueOf(kv.getTypeByte()));
kvbuilder.setTimestamp(kv.getTimestamp());
- kvbuilder.setValue(ByteStringer.wrap(kv.getValueArray(), kv.getValueOffset(),
+ kvbuilder.setValue(UnsafeByteOperations.unsafeWrap(kv.getValueArray(), kv.getValueOffset(),
kv.getValueLength()));
return kvbuilder.build();
}
@@ -2064,10 +2064,10 @@ public final class ProtobufUtil {
// input / output paths are relative to the store dir
// store dir is relative to region dir
CompactionDescriptor.Builder builder = CompactionDescriptor.newBuilder()
- .setTableName(ByteStringer.wrap(info.getTable().toBytes()))
- .setEncodedRegionName(ByteStringer.wrap(
+ .setTableName(UnsafeByteOperations.unsafeWrap(info.getTable().toBytes()))
+ .setEncodedRegionName(UnsafeByteOperations.unsafeWrap(
regionName == null ? info.getEncodedNameAsBytes() : regionName))
- .setFamilyName(ByteStringer.wrap(family))
+ .setFamilyName(UnsafeByteOperations.unsafeWrap(family))
.setStoreHomeDir(storeDir.getName()); //make relative
for (Path inputPath : inputPaths) {
builder.addCompactionInput(inputPath.getName()); //relative path
@@ -2075,7 +2075,7 @@ public final class ProtobufUtil {
for (Path outputPath : outputPaths) {
builder.addCompactionOutput(outputPath.getName());
}
- builder.setRegionName(ByteStringer.wrap(info.getRegionName()));
+ builder.setRegionName(UnsafeByteOperations.unsafeWrap(info.getRegionName()));
return builder.build();
}
@@ -2083,15 +2083,15 @@ public final class ProtobufUtil {
long flushSeqId, Map<byte[], List<Path>> committedFiles) {
FlushDescriptor.Builder desc = FlushDescriptor.newBuilder()
.setAction(action)
- .setEncodedRegionName(ByteStringer.wrap(hri.getEncodedNameAsBytes()))
- .setRegionName(ByteStringer.wrap(hri.getRegionName()))
+ .setEncodedRegionName(UnsafeByteOperations.unsafeWrap(hri.getEncodedNameAsBytes()))
+ .setRegionName(UnsafeByteOperations.unsafeWrap(hri.getRegionName()))
.setFlushSequenceNumber(flushSeqId)
- .setTableName(ByteStringer.wrap(hri.getTable().getName()));
+ .setTableName(UnsafeByteOperations.unsafeWrap(hri.getTable().getName()));
for (Map.Entry<byte[], List<Path>> entry : committedFiles.entrySet()) {
WALProtos.FlushDescriptor.StoreFlushDescriptor.Builder builder =
WALProtos.FlushDescriptor.StoreFlushDescriptor.newBuilder()
- .setFamilyName(ByteStringer.wrap(entry.getKey()))
+ .setFamilyName(UnsafeByteOperations.unsafeWrap(entry.getKey()))
.setStoreHomeDir(Bytes.toString(entry.getKey())); //relative to region
if (entry.getValue() != null) {
for (Path path : entry.getValue()) {
@@ -2129,15 +2129,15 @@ public final class ProtobufUtil {
Map<byte[], List<Path>> storeFiles) {
RegionEventDescriptor.Builder desc = RegionEventDescriptor.newBuilder()
.setEventType(eventType)
- .setTableName(ByteStringer.wrap(tableNameAsBytes))
- .setEncodedRegionName(ByteStringer.wrap(encodedNameAsBytes))
- .setRegionName(ByteStringer.wrap(regionNameAsBytes))
+ .setTableName(UnsafeByteOperations.unsafeWrap(tableNameAsBytes))
+ .setEncodedRegionName(UnsafeByteOperations.unsafeWrap(encodedNameAsBytes))
+ .setRegionName(UnsafeByteOperations.unsafeWrap(regionNameAsBytes))
.setLogSequenceNumber(seqId)
.setServer(toServerName(server));
for (Entry<byte[], List<Path>> entry : storeFiles.entrySet()) {
StoreDescriptor.Builder builder = StoreDescriptor.newBuilder()
- .setFamilyName(ByteStringer.wrap(entry.getKey()))
+ .setFamilyName(UnsafeByteOperations.unsafeWrap(entry.getKey()))
.setStoreHomeDir(Bytes.toString(entry.getKey()));
for (Path path : entry.getValue()) {
builder.addStoreFile(path.getName());
@@ -2214,8 +2214,8 @@ public final class ProtobufUtil {
public static HBaseProtos.TableName toProtoTableName(TableName tableName) {
return HBaseProtos.TableName.newBuilder()
- .setNamespace(ByteStringer.wrap(tableName.getNamespace()))
- .setQualifier(ByteStringer.wrap(tableName.getQualifier())).build();
+ .setNamespace(UnsafeByteOperations.unsafeWrap(tableName.getNamespace()))
+ .setQualifier(UnsafeByteOperations.unsafeWrap(tableName.getQualifier())).build();
}
public static TableName[] getTableNameArray(List<HBaseProtos.TableName> tableNamesList) {
@@ -2484,7 +2484,7 @@ public final class ProtobufUtil {
for (Map.Entry<byte[], List<Path>> entry : storeFiles.entrySet()) {
WALProtos.StoreDescriptor.Builder builder = StoreDescriptor.newBuilder()
- .setFamilyName(ByteStringer.wrap(entry.getKey()))
+ .setFamilyName(UnsafeByteOperations.unsafeWrap(entry.getKey()))
.setStoreHomeDir(Bytes.toString(entry.getKey())); // relative to region
for (Path path : entry.getValue()) {
String name = path.getName();
@@ -2697,11 +2697,11 @@ public final class ProtobufUtil {
*/
public static ColumnFamilySchema convertToColumnFamilySchema(HColumnDescriptor hcd) {
ColumnFamilySchema.Builder builder = ColumnFamilySchema.newBuilder();
- builder.setName(ByteStringer.wrap(hcd.getName()));
+ builder.setName(UnsafeByteOperations.unsafeWrap(hcd.getName()));
for (Map.Entry<Bytes, Bytes> e : hcd.getValues().entrySet()) {
BytesBytesPair.Builder aBuilder = BytesBytesPair.newBuilder();
- aBuilder.setFirst(ByteStringer.wrap(e.getKey().get()));
- aBuilder.setSecond(ByteStringer.wrap(e.getValue().get()));
+ aBuilder.setFirst(UnsafeByteOperations.unsafeWrap(e.getKey().get()));
+ aBuilder.setSecond(UnsafeByteOperations.unsafeWrap(e.getValue().get()));
builder.addAttributes(aBuilder.build());
}
for (Map.Entry<String, String> e : hcd.getConfiguration().entrySet()) {
@@ -2742,8 +2742,8 @@ public final class ProtobufUtil {
builder.setTableName(toProtoTableName(htd.getTableName()));
for (Map.Entry<Bytes, Bytes> e : htd.getValues().entrySet()) {
BytesBytesPair.Builder aBuilder = BytesBytesPair.newBuilder();
- aBuilder.setFirst(ByteStringer.wrap(e.getKey().get()));
- aBuilder.setSecond(ByteStringer.wrap(e.getValue().get()));
+ aBuilder.setFirst(UnsafeByteOperations.unsafeWrap(e.getKey().get()));
+ aBuilder.setSecond(UnsafeByteOperations.unsafeWrap(e.getValue().get()));
builder.addAttributes(aBuilder.build());
}
for (HColumnDescriptor hcd : htd.getColumnFamilies()) {
@@ -2948,7 +2948,7 @@ public final class ProtobufUtil {
ClusterStatusProtos.RegionState rs = rit.convert();
RegionSpecifier.Builder spec =
RegionSpecifier.newBuilder().setType(RegionSpecifierType.REGION_NAME);
- spec.setValue(ByteStringer.wrap(rit.getRegion().getRegionName()));
+ spec.setValue(UnsafeByteOperations.unsafeWrap(rit.getRegion().getRegionName()));
RegionInTransition pbRIT =
RegionInTransition.newBuilder().setSpec(spec.build()).setRegionState(rs).build();
@@ -3025,7 +3025,7 @@ public final class ProtobufUtil {
RegionSpecifier region = RequestConverter.buildRegionSpecifier(
RegionSpecifierType.REGION_NAME, regionName);
builder.setRegion(region);
- builder.addFamily(ByteStringer.wrap(family));
+ builder.addFamily(UnsafeByteOperations.unsafeWrap(family));
return builder.build();
}
@@ -3088,7 +3088,7 @@ public final class ProtobufUtil {
RegionSpecifierType.REGION_NAME, regionName);
builder.setRegion(region);
if (splitPoint != null) {
- builder.setSplitPoint(ByteStringer.wrap(splitPoint));
+ builder.setSplitPoint(UnsafeByteOperations.unsafeWrap(splitPoint));
}
return builder.build();
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java
index baec3e4..3f91ee0 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java
@@ -46,6 +46,7 @@ import org.apache.hadoop.hbase.client.RowMutations;
import org.apache.hadoop.hbase.client.Scan;
import org.apache.hadoop.hbase.exceptions.DeserializationException;
import org.apache.hadoop.hbase.filter.ByteArrayComparable;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionRequest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionRequest;
@@ -105,7 +106,6 @@ import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOr
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.GetLastFlushedSequenceIdRequest;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
import org.apache.hadoop.hbase.util.Pair;
@@ -160,14 +160,14 @@ public final class RequestConverter {
builder.setRegion(region);
MutationProto.Builder mutateBuilder = MutationProto.newBuilder();
- mutateBuilder.setRow(ByteStringer.wrap(row));
+ mutateBuilder.setRow(UnsafeByteOperations.unsafeWrap(row));
mutateBuilder.setMutateType(MutationType.INCREMENT);
mutateBuilder.setDurability(ProtobufUtil.toDurability(durability));
ColumnValue.Builder columnBuilder = ColumnValue.newBuilder();
- columnBuilder.setFamily(ByteStringer.wrap(family));
+ columnBuilder.setFamily(UnsafeByteOperations.unsafeWrap(family));
QualifierValue.Builder valueBuilder = QualifierValue.newBuilder();
- valueBuilder.setValue(ByteStringer.wrap(Bytes.toBytes(amount)));
- valueBuilder.setQualifier(ByteStringer.wrap(qualifier));
+ valueBuilder.setValue(UnsafeByteOperations.unsafeWrap(Bytes.toBytes(amount)));
+ valueBuilder.setQualifier(UnsafeByteOperations.unsafeWrap(qualifier));
columnBuilder.addQualifierValue(valueBuilder.build());
mutateBuilder.addColumnValue(columnBuilder.build());
if (nonce != HConstants.NO_NONCE) {
@@ -549,8 +549,8 @@ public final class RequestConverter {
if (userToken != null) {
protoDT =
ClientProtos.DelegationToken.newBuilder()
- .setIdentifier(ByteStringer.wrap(userToken.getIdentifier()))
- .setPassword(ByteStringer.wrap(userToken.getPassword()))
+ .setIdentifier(UnsafeByteOperations.unsafeWrap(userToken.getIdentifier()))
+ .setPassword(UnsafeByteOperations.unsafeWrap(userToken.getPassword()))
.setKind(userToken.getKind().toString())
.setService(userToken.getService().toString()).build();
}
@@ -559,7 +559,7 @@ public final class RequestConverter {
new ArrayList<ClientProtos.BulkLoadHFileRequest.FamilyPath>(familyPaths.size());
for(Pair<byte[], String> el: familyPaths) {
protoFamilyPaths.add(ClientProtos.BulkLoadHFileRequest.FamilyPath.newBuilder()
- .setFamily(ByteStringer.wrap(el.getFirst()))
+ .setFamily(UnsafeByteOperations.unsafeWrap(el.getFirst()))
.setPath(el.getSecond()).build());
}
@@ -619,7 +619,7 @@ public final class RequestConverter {
exec.getRequest().toByteArray());
regionActionBuilder.addAction(actionBuilder.setServiceCall(
ClientProtos.CoprocessorServiceCall.newBuilder()
- .setRow(ByteStringer.wrap(exec.getRow()))
+ .setRow(UnsafeByteOperations.unsafeWrap(exec.getRow()))
.setServiceName(exec.getMethod().getService().getFullName())
.setMethodName(exec.getMethod().getName())
.setRequest(value)));
@@ -702,7 +702,7 @@ public final class RequestConverter {
exec.getRequest().toByteArray());
builder.addAction(actionBuilder.setServiceCall(
ClientProtos.CoprocessorServiceCall.newBuilder()
- .setRow(ByteStringer.wrap(exec.getRow()))
+ .setRow(UnsafeByteOperations.unsafeWrap(exec.getRow()))
.setServiceName(exec.getMethod().getService().getFullName())
.setMethodName(exec.getMethod().getName())
.setRequest(value)));
@@ -875,7 +875,7 @@ public final class RequestConverter {
builder.setRegion(region);
builder.setMajor(major);
if (family != null) {
- builder.setFamily(ByteStringer.wrap(family));
+ builder.setFamily(UnsafeByteOperations.unsafeWrap(family));
}
return builder.build();
}
@@ -934,7 +934,7 @@ public final class RequestConverter {
public static RegionSpecifier buildRegionSpecifier(
final RegionSpecifierType type, final byte[] value) {
RegionSpecifier.Builder regionBuilder = RegionSpecifier.newBuilder();
- regionBuilder.setValue(ByteStringer.wrap(value));
+ regionBuilder.setValue(UnsafeByteOperations.unsafeWrap(value));
regionBuilder.setType(type);
return regionBuilder.build();
}
@@ -955,9 +955,9 @@ public final class RequestConverter {
final ByteArrayComparable comparator,
final CompareType compareType) throws IOException {
Condition.Builder builder = Condition.newBuilder();
- builder.setRow(ByteStringer.wrap(row));
- builder.setFamily(ByteStringer.wrap(family));
- builder.setQualifier(ByteStringer.wrap(qualifier));
+ builder.setRow(UnsafeByteOperations.unsafeWrap(row));
+ builder.setFamily(UnsafeByteOperations.unsafeWrap(family));
+ builder.setQualifier(UnsafeByteOperations.unsafeWrap(qualifier));
builder.setComparator(ProtobufUtil.toComparator(comparator));
builder.setCompareType(compareType);
return builder.build();
@@ -997,7 +997,7 @@ public final class RequestConverter {
final long nonce) {
DeleteColumnRequest.Builder builder = DeleteColumnRequest.newBuilder();
builder.setTableName(ProtobufUtil.toProtoTableName((tableName)));
- builder.setColumnName(ByteStringer.wrap(columnName));
+ builder.setColumnName(UnsafeByteOperations.unsafeWrap(columnName));
builder.setNonceGroup(nonceGroup);
builder.setNonce(nonce);
return builder.build();
@@ -1187,7 +1187,7 @@ public final class RequestConverter {
builder.setTableSchema(ProtobufUtil.convertToTableSchema(hTableDesc));
if (splitKeys != null) {
for (byte [] splitKey : splitKeys) {
- builder.addSplitKeys(ByteStringer.wrap(splitKey));
+ builder.addSplitKeys(UnsafeByteOperations.unsafeWrap(splitKey));
}
}
builder.setNonceGroup(nonceGroup);
@@ -1401,7 +1401,7 @@ public final class RequestConverter {
public static GetLastFlushedSequenceIdRequest buildGetLastFlushedSequenceIdRequest(
byte[] regionName) {
return GetLastFlushedSequenceIdRequest.newBuilder().setRegionName(
- ByteStringer.wrap(regionName)).build();
+ UnsafeByteOperations.unsafeWrap(regionName)).build();
}
/**
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java
index 8018bd7..cf11641 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java
@@ -42,16 +42,16 @@ import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hbase.AuthUtil;
import org.apache.hadoop.hbase.HBaseConfiguration;
import org.apache.hadoop.hbase.HConstants;
-import org.apache.hadoop.hbase.ServerName;
import org.apache.hadoop.hbase.classification.InterfaceAudience;
import org.apache.hadoop.hbase.exceptions.DeserializationException;
+import org.apache.hadoop.hbase.replication.ReplicationStateZKBase;
+import org.apache.hadoop.hbase.security.Superusers;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
import org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos;
import org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStoreSequenceIds;
import org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos;
-import org.apache.hadoop.hbase.replication.ReplicationStateZKBase;
-import org.apache.hadoop.hbase.security.Superusers;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.hbase.util.Threads;
import org.apache.hadoop.hbase.zookeeper.ZKUtil.ZKUtilOp.CreateAndFailSilent;
@@ -76,7 +76,6 @@ import org.apache.zookeeper.proto.CreateRequest;
import org.apache.zookeeper.proto.DeleteRequest;
import org.apache.zookeeper.proto.SetDataRequest;
import org.apache.zookeeper.server.ZooKeeperSaslServer;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
/**
* Internal HBase utility class for ZooKeeper.
@@ -2107,7 +2106,7 @@ public class ZKUtil {
for (Map.Entry<byte[], Long> e : storeSequenceIds.entrySet()){
byte[] columnFamilyName = e.getKey();
Long curSeqId = e.getValue();
- storeSequenceIdBuilder.setFamilyName(ByteStringer.wrap(columnFamilyName));
+ storeSequenceIdBuilder.setFamilyName(UnsafeByteOperations.unsafeWrap(columnFamilyName));
storeSequenceIdBuilder.setSequenceId(curSeqId);
regionSequenceIdsBuilder.addStoreSequenceId(storeSequenceIdBuilder.build());
storeSequenceIdBuilder.clear();
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java b/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java
index df6de42..41c9a56 100644
--- a/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java
+++ b/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java
@@ -77,7 +77,6 @@ import org.apache.hadoop.hbase.regionserver.RegionServerStoppedException;
import org.apache.hadoop.hbase.security.User;
import org.apache.hadoop.hbase.testclassification.ClientTests;
import org.apache.hadoop.hbase.testclassification.SmallTests;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.hbase.util.Pair;
import org.apache.hadoop.hbase.util.Threads;
@@ -93,6 +92,7 @@ import com.google.common.base.Stopwatch;
import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString;
import org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController;
import org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
/**
* Test client behavior w/o setting up a cluster.
@@ -509,7 +509,7 @@ public class TestClientNoCluster extends Configured implements Tool {
if (max <= 0) break;
if (++count > max) break;
HRegionInfo hri = e.getValue().getFirst();
- ByteString row = ByteStringer.wrap(hri.getRegionName());
+ ByteString row = UnsafeByteOperations.unsafeWrap(hri.getRegionName());
resultBuilder.clear();
resultBuilder.addCell(getRegionInfo(row, hri));
resultBuilder.addCell(getServer(row, e.getValue().getSecond()));
@@ -556,11 +556,11 @@ public class TestClientNoCluster extends Configured implements Tool {
}
private final static ByteString CATALOG_FAMILY_BYTESTRING =
- ByteStringer.wrap(HConstants.CATALOG_FAMILY);
+ UnsafeByteOperations.unsafeWrap(HConstants.CATALOG_FAMILY);
private final static ByteString REGIONINFO_QUALIFIER_BYTESTRING =
- ByteStringer.wrap(HConstants.REGIONINFO_QUALIFIER);
+ UnsafeByteOperations.unsafeWrap(HConstants.REGIONINFO_QUALIFIER);
private final static ByteString SERVER_QUALIFIER_BYTESTRING =
- ByteStringer.wrap(HConstants.SERVER_QUALIFIER);
+ UnsafeByteOperations.unsafeWrap(HConstants.SERVER_QUALIFIER);
static CellProtos.Cell.Builder getBaseCellBuilder(final ByteString row) {
CellProtos.Cell.Builder cellBuilder = CellProtos.Cell.newBuilder();
@@ -573,7 +573,7 @@ public class TestClientNoCluster extends Configured implements Tool {
static CellProtos.Cell getRegionInfo(final ByteString row, final HRegionInfo hri) {
CellProtos.Cell.Builder cellBuilder = getBaseCellBuilder(row);
cellBuilder.setQualifier(REGIONINFO_QUALIFIER_BYTESTRING);
- cellBuilder.setValue(ByteStringer.wrap(hri.toByteArray()));
+ cellBuilder.setValue(UnsafeByteOperations.unsafeWrap(hri.toByteArray()));
return cellBuilder.build();
}
@@ -586,9 +586,10 @@ public class TestClientNoCluster extends Configured implements Tool {
static CellProtos.Cell getStartCode(final ByteString row) {
CellProtos.Cell.Builder cellBuilder = getBaseCellBuilder(row);
- cellBuilder.setQualifier(ByteStringer.wrap(HConstants.STARTCODE_QUALIFIER));
+ cellBuilder.setQualifier(UnsafeByteOperations.unsafeWrap(HConstants.STARTCODE_QUALIFIER));
// TODO:
- cellBuilder.setValue(ByteStringer.wrap(Bytes.toBytes(META_SERVERNAME.getStartcode())));
+ cellBuilder.setValue(UnsafeByteOperations.unsafeWrap(
+ Bytes.toBytes(META_SERVERNAME.getStartcode())));
return cellBuilder.build();
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureUtil.java
----------------------------------------------------------------------
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureUtil.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureUtil.java
index c2985fd..05e8e09 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureUtil.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureUtil.java
@@ -17,13 +17,13 @@
*/
package org.apache.hadoop.hbase.procedure2;
-import org.apache.hadoop.hbase.classification.InterfaceAudience;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos;
import org.apache.hadoop.hbase.HConstants;
import org.apache.hadoop.hbase.ProcedureInfo;
import org.apache.hadoop.hbase.ProcedureState;
+import org.apache.hadoop.hbase.classification.InterfaceAudience;
+import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
+import org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos;
import org.apache.hadoop.hbase.util.ForeignExceptionUtil;
-import org.apache.hadoop.hbase.shaded.util.ByteStringer;
import org.apache.hadoop.hbase.util.NonceKey;
/**
@@ -60,7 +60,7 @@ public final class ProcedureUtil {
}
if (procInfo.hasResultData()) {
- builder.setResult(ByteStringer.wrap(procInfo.getResult()));
+ builder.setResult(UnsafeByteOperations.unsafeWrap(procInfo.getResult()));
}
return builder.build();
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/README.txt
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/README.txt b/hbase-protocol-shaded/README.txt
index 387ada9..5a4b83b 100644
--- a/hbase-protocol-shaded/README.txt
+++ b/hbase-protocol-shaded/README.txt
@@ -1,41 +1,44 @@
-These are the protobuf definition files used by core hbase. This modules
-does shading of all to do with protobuf. All of core relies on this module.
-All core references in core to protobuf are to the protobuf this module
-includes but offset by the package prefix of org.apache.hadoop.hbase.shaded.*
-as in org.apache.hadoop.hbase.shaded.protobuf.generated.* and
-org.apache.hadoop.hbase.shaded.com.google.protobuf.*.
-
-NOTE: the .protos in here are copied in an adjacent module, hbase-protocol.
-There they are non-shaded. If you make changes here, consider making them
-over in the adjacent module too. Be careful, the .proto files are not
-exactly the same; they differ in one line at least -- the location the file
-gets generated to; i.e. those in here get the extra 'shaded' in their
-package name.
-
-The produced java classes are generated and then checked in. The reasoning
-is that they change infrequently.
-
-To regenerate the classes after making definition file changes, ensure first that
-the protobuf protoc tool is in your $PATH. You may need to download it and build
-it first; its part of the protobuf package. For example, if using v2.5.0 of
-protobuf, it is obtainable from here:
-
- https://github.com/google/protobuf/releases/tag/v2.5.0
-
-HBase uses hadoop-maven-plugins:protoc goal to invoke the protoc command. You can
-compile the protoc definitions by invoking maven with profile compile-protobuf or
-passing in compile-protobuf property.
-
-mvn compile -Dcompile-protobuf
-or
-mvn compile -Pcompile-protobuf
+Please read carefully as the 'menu options' have changed.
+
+This module has proto files used by core. These protos
+overlap with protos that are used by coprocessor endpoints
+(CPEP) in the module hbase-protocol. So the core versions have
+a different name, the generated classes are relocated
+-- i.e. shaded -- to a new location; they are moved from
+org.apache.hadoop.hbase.* to org.apache.hadoop.hbase.shaded.
+
+This module also includes the protobuf that hbase core depends
+on again relocated to live at an offset of
+org.apache.hadoop.hbase.shaded so as to avoid clashes with other
+versions of protobuf resident on our CLASSPATH included,
+transitively or otherwise, by dependencies: i.e. the shaded
+protobuf Message class is at
+org.apache.hadoop.hbase.shaded.com.google.protobuf.Message
+rather than at com.google.protobuf.Message.
+
+Below we describe how to generate the java files for this
+module. Run this step any time you change the proto files
+in this module or if you change the protobuf version. If you
+add a new file, be sure to add mention of the proto in the
+pom.xml (scroll till you see the listing of protos to consider).
+
+First ensure that the appropriate protobuf protoc tool is in
+your $PATH (or pass -Dprotoc.path=PATH_TO_PROTOC when running
+the below mvn commands). You may need to download protobuf and
+build protoc first.
+
+Run:
+
+ $ mvn install -Dgenerate-shaded-classes
-You may also want to define protoc.path for the protoc binary
+or
-mvn compile -Dcompile-protobuf -Dprotoc.path=/opt/local/bin/protoc
+ $ mvn install -Pgenerate-shaded-classes
-If you have added a new proto file, you should add it to the pom.xml file first.
-Other modules also support the maven profile.
+to build and trigger the special generate-shaded-classes
+profile. When finished, the content of
+src/main/java/org/apache/hadoop/hbase/shaded will have
+been updated. Check in the changes.
-After you've done the above, check it in and then check it in (or post a patch
-on a JIRA with your definition file changes and the generated files).
+See the pom.xml under the generate-shaded-classes profile
+for more info on how this step works.