You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2016/09/30 21:56:54 UTC
[05/29] hbase git commit: HBASE-16741 Amend the generate protobufs
out-of-band build step to include shade,
pulling in protobuf source and a hook for patching protobuf
http://git-wip-us.apache.org/repos/asf/hbase/blob/32be831c/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/SnapshotProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/SnapshotProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/SnapshotProtos.java
index 4d3b294..d40086b 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/SnapshotProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/SnapshotProtos.java
@@ -6,17 +6,17 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class SnapshotProtos {
private SnapshotProtos() {}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistryLite registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
- (com.google.protobuf.ExtensionRegistryLite) registry);
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
}
public interface SnapshotFileInfoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.SnapshotFileInfo)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required .hbase.pb.SnapshotFileInfo.Type type = 1;</code>
@@ -38,7 +38,7 @@ public final class SnapshotProtos {
/**
* <code>optional string hfile = 3;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getHfileBytes();
/**
@@ -52,7 +52,7 @@ public final class SnapshotProtos {
/**
* <code>optional string wal_server = 4;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getWalServerBytes();
/**
@@ -66,18 +66,18 @@ public final class SnapshotProtos {
/**
* <code>optional string wal_name = 5;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getWalNameBytes();
}
/**
* Protobuf type {@code hbase.pb.SnapshotFileInfo}
*/
public static final class SnapshotFileInfo extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.SnapshotFileInfo)
SnapshotFileInfoOrBuilder {
// Use SnapshotFileInfo.newBuilder() to construct.
- private SnapshotFileInfo(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private SnapshotFileInfo(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private SnapshotFileInfo() {
@@ -88,18 +88,18 @@ public final class SnapshotProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private SnapshotFileInfo(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -127,41 +127,41 @@ public final class SnapshotProtos {
break;
}
case 26: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000002;
hfile_ = bs;
break;
}
case 34: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000004;
walServer_ = bs;
break;
}
case 42: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000008;
walName_ = bs;
break;
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotFileInfo_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotFileInfo_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -172,7 +172,7 @@ public final class SnapshotProtos {
* Protobuf enum {@code hbase.pb.SnapshotFileInfo.Type}
*/
public enum Type
- implements com.google.protobuf.ProtocolMessageEnum {
+ implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
/**
* <code>HFILE = 1;</code>
*/
@@ -213,27 +213,27 @@ public final class SnapshotProtos {
}
}
- public static com.google.protobuf.Internal.EnumLiteMap<Type>
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<Type>
internalGetValueMap() {
return internalValueMap;
}
- private static final com.google.protobuf.Internal.EnumLiteMap<
+ private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
Type> internalValueMap =
- new com.google.protobuf.Internal.EnumLiteMap<Type>() {
+ new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<Type>() {
public Type findValueByNumber(int number) {
return Type.forNumber(number);
}
};
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
- public final com.google.protobuf.Descriptors.EnumDescriptor
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
- public static final com.google.protobuf.Descriptors.EnumDescriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo.getDescriptor().getEnumTypes().get(0);
}
@@ -241,7 +241,7 @@ public final class SnapshotProtos {
private static final Type[] VALUES = values();
public static Type valueOf(
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
@@ -291,8 +291,8 @@ public final class SnapshotProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
hfile_ = s;
@@ -303,17 +303,17 @@ public final class SnapshotProtos {
/**
* <code>optional string hfile = 3;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getHfileBytes() {
java.lang.Object ref = hfile_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
hfile_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -333,8 +333,8 @@ public final class SnapshotProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
walServer_ = s;
@@ -345,17 +345,17 @@ public final class SnapshotProtos {
/**
* <code>optional string wal_server = 4;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getWalServerBytes() {
java.lang.Object ref = walServer_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
walServer_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -375,8 +375,8 @@ public final class SnapshotProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
walName_ = s;
@@ -387,17 +387,17 @@ public final class SnapshotProtos {
/**
* <code>optional string wal_name = 5;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getWalNameBytes() {
java.lang.Object ref = walName_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
walName_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -415,19 +415,19 @@ public final class SnapshotProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeEnum(1, type_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 3, hfile_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 3, hfile_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 4, walServer_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 4, walServer_);
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 5, walName_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 5, walName_);
}
unknownFields.writeTo(output);
}
@@ -438,17 +438,17 @@ public final class SnapshotProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeEnumSize(1, type_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(3, hfile_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(3, hfile_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(4, walServer_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(4, walServer_);
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(5, walName_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(5, walName_);
}
size += unknownFields.getSerializedSize();
memoizedSize = size;
@@ -519,61 +519,61 @@ public final class SnapshotProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -591,7 +591,7 @@ public final class SnapshotProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -599,15 +599,15 @@ public final class SnapshotProtos {
* Protobuf type {@code hbase.pb.SnapshotFileInfo}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.SnapshotFileInfo)
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfoOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotFileInfo_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotFileInfo_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -620,12 +620,12 @@ public final class SnapshotProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
@@ -642,7 +642,7 @@ public final class SnapshotProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotFileInfo_descriptor;
}
@@ -688,29 +688,29 @@ public final class SnapshotProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo)other);
} else {
@@ -752,13 +752,13 @@ public final class SnapshotProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -819,8 +819,8 @@ public final class SnapshotProtos {
public java.lang.String getHfile() {
java.lang.Object ref = hfile_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
hfile_ = s;
@@ -833,17 +833,17 @@ public final class SnapshotProtos {
/**
* <code>optional string hfile = 3;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getHfileBytes() {
java.lang.Object ref = hfile_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
hfile_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -872,7 +872,7 @@ public final class SnapshotProtos {
* <code>optional string hfile = 3;</code>
*/
public Builder setHfileBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -895,8 +895,8 @@ public final class SnapshotProtos {
public java.lang.String getWalServer() {
java.lang.Object ref = walServer_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
walServer_ = s;
@@ -909,17 +909,17 @@ public final class SnapshotProtos {
/**
* <code>optional string wal_server = 4;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getWalServerBytes() {
java.lang.Object ref = walServer_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
walServer_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -948,7 +948,7 @@ public final class SnapshotProtos {
* <code>optional string wal_server = 4;</code>
*/
public Builder setWalServerBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -971,8 +971,8 @@ public final class SnapshotProtos {
public java.lang.String getWalName() {
java.lang.Object ref = walName_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
walName_ = s;
@@ -985,17 +985,17 @@ public final class SnapshotProtos {
/**
* <code>optional string wal_name = 5;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getWalNameBytes() {
java.lang.Object ref = walName_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
walName_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -1024,7 +1024,7 @@ public final class SnapshotProtos {
* <code>optional string wal_name = 5;</code>
*/
public Builder setWalNameBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1034,12 +1034,12 @@ public final class SnapshotProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -1057,22 +1057,22 @@ public final class SnapshotProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<SnapshotFileInfo>
- PARSER = new com.google.protobuf.AbstractParser<SnapshotFileInfo>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<SnapshotFileInfo>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<SnapshotFileInfo>() {
public SnapshotFileInfo parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new SnapshotFileInfo(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<SnapshotFileInfo> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<SnapshotFileInfo> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<SnapshotFileInfo> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<SnapshotFileInfo> getParserForType() {
return PARSER;
}
@@ -1084,7 +1084,7 @@ public final class SnapshotProtos {
public interface SnapshotRegionManifestOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.SnapshotRegionManifest)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>optional int32 version = 1;</code>
@@ -1136,11 +1136,11 @@ public final class SnapshotProtos {
* Protobuf type {@code hbase.pb.SnapshotRegionManifest}
*/
public static final class SnapshotRegionManifest extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.SnapshotRegionManifest)
SnapshotRegionManifestOrBuilder {
// Use SnapshotRegionManifest.newBuilder() to construct.
- private SnapshotRegionManifest(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private SnapshotRegionManifest(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private SnapshotRegionManifest() {
@@ -1149,18 +1149,18 @@ public final class SnapshotProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private SnapshotRegionManifest(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -1205,10 +1205,10 @@ public final class SnapshotProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
@@ -1218,12 +1218,12 @@ public final class SnapshotProtos {
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotRegionManifest_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotRegionManifest_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1232,7 +1232,7 @@ public final class SnapshotProtos {
public interface StoreFileOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.SnapshotRegionManifest.StoreFile)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required string name = 1;</code>
@@ -1245,7 +1245,7 @@ public final class SnapshotProtos {
/**
* <code>required string name = 1;</code>
*/
- com.google.protobuf.ByteString
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getNameBytes();
/**
@@ -1282,11 +1282,11 @@ public final class SnapshotProtos {
* Protobuf type {@code hbase.pb.SnapshotRegionManifest.StoreFile}
*/
public static final class StoreFile extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.SnapshotRegionManifest.StoreFile)
StoreFileOrBuilder {
// Use StoreFile.newBuilder() to construct.
- private StoreFile(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private StoreFile(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private StoreFile() {
@@ -1295,18 +1295,18 @@ public final class SnapshotProtos {
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private StoreFile(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -1323,7 +1323,7 @@ public final class SnapshotProtos {
break;
}
case 10: {
- com.google.protobuf.ByteString bs = input.readBytes();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000001;
name_ = bs;
break;
@@ -1348,22 +1348,22 @@ public final class SnapshotProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotRegionManifest_StoreFile_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotRegionManifest_StoreFile_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1387,8 +1387,8 @@ public final class SnapshotProtos {
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
name_ = s;
@@ -1399,17 +1399,17 @@ public final class SnapshotProtos {
/**
* <code>required string name = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getNameBytes() {
java.lang.Object ref = name_;
if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
name_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
@@ -1477,10 +1477,10 @@ public final class SnapshotProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- com.google.protobuf.GeneratedMessageV3.writeString(output, 1, name_);
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, name_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
output.writeMessage(2, getReference());
@@ -1497,14 +1497,14 @@ public final class SnapshotProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, name_);
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, name_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(2, getReference());
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeUInt64Size(3, fileSize_);
}
size += unknownFields.getSerializedSize();
@@ -1560,7 +1560,7 @@ public final class SnapshotProtos {
}
if (hasFileSize()) {
hash = (37 * hash) + FILE_SIZE_FIELD_NUMBER;
- hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
getFileSize());
}
hash = (29 * hash) + unknownFields.hashCode();
@@ -1569,61 +1569,61 @@ public final class SnapshotProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -1641,7 +1641,7 @@ public final class SnapshotProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1649,15 +1649,15 @@ public final class SnapshotProtos {
* Protobuf type {@code hbase.pb.SnapshotRegionManifest.StoreFile}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.SnapshotRegionManifest.StoreFile)
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFileOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotRegionManifest_StoreFile_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotRegionManifest_StoreFile_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1670,12 +1670,12 @@ public final class SnapshotProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getReferenceFieldBuilder();
}
@@ -1695,7 +1695,7 @@ public final class SnapshotProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotRegionManifest_StoreFile_descriptor;
}
@@ -1741,29 +1741,29 @@ public final class SnapshotProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile)other);
} else {
@@ -1803,13 +1803,13 @@ public final class SnapshotProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -1834,8 +1834,8 @@ public final class SnapshotProtos {
public java.lang.String getName() {
java.lang.Object ref = name_;
if (!(ref instanceof java.lang.String)) {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
+ (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
name_ = s;
@@ -1848,17 +1848,17 @@ public final class SnapshotProtos {
/**
* <code>required string name = 1;</code>
*/
- public com.google.protobuf.ByteString
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
getNameBytes() {
java.lang.Object ref = name_;
if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
name_ = b;
return b;
} else {
- return (com.google.protobuf.ByteString) ref;
+ return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
}
}
/**
@@ -1887,7 +1887,7 @@ public final class SnapshotProtos {
* <code>required string name = 1;</code>
*/
public Builder setNameBytes(
- com.google.protobuf.ByteString value) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1898,7 +1898,7 @@ public final class SnapshotProtos {
}
private org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference reference_ = null;
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference, org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.ReferenceOrBuilder> referenceBuilder_;
/**
* <code>optional .hbase.pb.Reference reference = 2;</code>
@@ -2001,11 +2001,11 @@ public final class SnapshotProtos {
/**
* <code>optional .hbase.pb.Reference reference = 2;</code>
*/
- private com.google.protobuf.SingleFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference, org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.ReferenceOrBuilder>
getReferenceFieldBuilder() {
if (referenceBuilder_ == null) {
- referenceBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ referenceBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference, org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.ReferenceOrBuilder>(
getReference(),
getParentForChildren(),
@@ -2063,12 +2063,12 @@ public final class SnapshotProtos {
return this;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -2086,22 +2086,22 @@ public final class SnapshotProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<StoreFile>
- PARSER = new com.google.protobuf.AbstractParser<StoreFile>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<StoreFile>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<StoreFile>() {
public StoreFile parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new StoreFile(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<StoreFile> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<StoreFile> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<StoreFile> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<StoreFile> getParserForType() {
return PARSER;
}
@@ -2113,7 +2113,7 @@ public final class SnapshotProtos {
public interface FamilyFilesOrBuilder extends
// @@protoc_insertion_point(interface_extends:hbase.pb.SnapshotRegionManifest.FamilyFiles)
- com.google.protobuf.MessageOrBuilder {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
/**
* <code>required bytes family_name = 1;</code>
@@ -2122,7 +2122,7 @@ public final class SnapshotProtos {
/**
* <code>required bytes family_name = 1;</code>
*/
- com.google.protobuf.ByteString getFamilyName();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamilyName();
/**
* <code>repeated .hbase.pb.SnapshotRegionManifest.StoreFile store_files = 2;</code>
@@ -2152,31 +2152,31 @@ public final class SnapshotProtos {
* Protobuf type {@code hbase.pb.SnapshotRegionManifest.FamilyFiles}
*/
public static final class FamilyFiles extends
- com.google.protobuf.GeneratedMessageV3 implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hbase.pb.SnapshotRegionManifest.FamilyFiles)
FamilyFilesOrBuilder {
// Use FamilyFiles.newBuilder() to construct.
- private FamilyFiles(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
+ private FamilyFiles(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private FamilyFiles() {
- familyName_ = com.google.protobuf.ByteString.EMPTY;
+ familyName_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
storeFiles_ = java.util.Collections.emptyList();
}
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
+ public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private FamilyFiles(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
this();
int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
@@ -2208,10 +2208,10 @@ public final class SnapshotProtos {
}
}
}
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
+ throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
@@ -2221,12 +2221,12 @@ public final class SnapshotProtos {
makeExtensionsImmutable();
}
}
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotRegionManifest_FamilyFiles_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotRegionManifest_FamilyFiles_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2235,7 +2235,7 @@ public final class SnapshotProtos {
private int bitField0_;
public static final int FAMILY_NAME_FIELD_NUMBER = 1;
- private com.google.protobuf.ByteString familyName_;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString familyName_;
/**
* <code>required bytes family_name = 1;</code>
*/
@@ -2245,7 +2245,7 @@ public final class SnapshotProtos {
/**
* <code>required bytes family_name = 1;</code>
*/
- public com.google.protobuf.ByteString getFamilyName() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamilyName() {
return familyName_;
}
@@ -2304,7 +2304,7 @@ public final class SnapshotProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, familyName_);
@@ -2321,11 +2321,11 @@ public final class SnapshotProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeBytesSize(1, familyName_);
}
for (int i = 0; i < storeFiles_.size(); i++) {
- size += com.google.protobuf.CodedOutputStream
+ size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
.computeMessageSize(2, storeFiles_.get(i));
}
size += unknownFields.getSerializedSize();
@@ -2377,61 +2377,61 @@ public final class SnapshotProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles parseFrom(
byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles parseFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles parseDelimitedFrom(
java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles parseFrom(
- com.google.protobuf.CodedInputStream input)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return com.google.protobuf.GeneratedMessageV3
+ return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@@ -2449,7 +2449,7 @@ public final class SnapshotProtos {
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -2457,15 +2457,15 @@ public final class SnapshotProtos {
* Protobuf type {@code hbase.pb.SnapshotRegionManifest.FamilyFiles}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hbase.pb.SnapshotRegionManifest.FamilyFiles)
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFilesOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
+ public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotRegionManifest_FamilyFiles_descriptor;
}
- protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
+ protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotRegionManifest_FamilyFiles_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2478,19 +2478,19 @@ public final class SnapshotProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessageV3
+ if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getStoreFilesFieldBuilder();
}
}
public Builder clear() {
super.clear();
- familyName_ = com.google.protobuf.ByteString.EMPTY;
+ familyName_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000001);
if (storeFilesBuilder_ == null) {
storeFiles_ = java.util.Collections.emptyList();
@@ -2501,7 +2501,7 @@ public final class SnapshotProtos {
return this;
}
- public com.google.protobuf.Descriptors.Descriptor
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotRegionManifest_FamilyFiles_descriptor;
}
@@ -2544,29 +2544,29 @@ public final class SnapshotProtos {
return (Builder) super.clone();
}
public Builder setField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
- com.google.protobuf.Descriptors.FieldDescriptor field) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
- com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
int index, Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
- com.google.protobuf.Descriptors.FieldDescriptor field,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) {
return (Builder) super.addRepeatedField(field, value);
}
- public Builder mergeFrom(com.google.protobuf.Message other) {
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles)other);
} else {
@@ -2599,7 +2599,7 @@ public final class SnapshotProtos {
storeFiles_ = other.storeFiles_;
bitField0_ = (bitField0_ & ~0x00000002);
storeFilesBuilder_ =
- com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getStoreFilesFieldBuilder() : null;
} else {
storeFilesBuilder_.addAllMessages(other.storeFiles_);
@@ -2624,13 +2624,13 @@ public final class SnapshotProtos {
}
public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.FamilyFiles) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
@@ -2642,7 +2642,7 @@ public final class SnapshotProtos {
}
private int bitField0_;
- private com.google.protobuf.ByteString familyName_ = com.google.protobuf.ByteString.EMPTY;
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString familyName_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
/**
* <code>required bytes family_name = 1;</code>
*/
@@ -2652,13 +2652,13 @@ public final class SnapshotProtos {
/**
* <code>required bytes family_name = 1;</code>
*/
- public com.google.protobuf.ByteString getFamilyName() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamilyName() {
return familyName_;
}
/**
* <code>required bytes family_name = 1;</code>
*/
- public Builder setFamilyName(com.google.protobuf.ByteString value) {
+ public Builder setFamilyName(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
@@ -2686,7 +2686,7 @@ public final class SnapshotProtos {
}
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFileOrBuilder> storeFilesBuilder_;
/**
@@ -2818,7 +2818,7 @@ public final class SnapshotProtos {
java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile> values) {
if (storeFilesBuilder_ == null) {
ensureStoreFilesIsMutable();
- com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
values, storeFiles_);
onChanged();
} else {
@@ -2902,11 +2902,11 @@ public final class SnapshotProtos {
getStoreFilesBuilderList() {
return getStoreFilesFieldBuilder().getBuilderList();
}
- private com.google.protobuf.RepeatedFieldBuilderV3<
+ private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFileOrBuilder>
getStoreFilesFieldBuilder() {
if (storeFilesBuilder_ == null) {
- storeFilesBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3<
+ storeFilesBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFile.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest.StoreFileOrBuilder>(
storeFiles_,
((bitField0_ & 0x00000002) == 0x00000002),
@@ -2917,12 +2917,12 @@ public final class SnapshotProtos {
return storeFilesBuilder_;
}
public final Builder setUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
public final Builder mergeUnknownFields(
- final com.google.protobuf.UnknownFieldSet unknownFields) {
+ final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
@@ -2940,22 +2940,22 @@ public final class SnapshotProtos {
return DEFAULT_INSTANCE;
}
- @java.lang.Deprecated public static final com.google.protobuf.Parser<FamilyFiles>
- PARSER = new com.google.protobuf.AbstractParser<FamilyFiles>() {
+ @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<FamilyFiles>
+ PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<FamilyFiles>() {
public FamilyFiles parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
+ org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
return new FamilyFiles(input, extensionRegistry);
}
};
- public static com.google.protobuf.Parser<FamilyFiles> parser() {
+ public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<FamilyFiles> parser() {
return PARSER;
}
@java.lang.Override
- public com.google.protobuf.Parser<FamilyFiles> getParserForType() {
+ public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<FamilyFiles> getParserForType() {
return PARSER;
}
@@ -3061,7 +3061,7 @@ public final class SnapshotProtos {
return true;
}
- public void writeTo(com.google.protobuf.CodedOutputStream output)
+ public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
<TRUNCATED>