You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2016/09/30 00:30:54 UTC
[25/29] hbase git commit: HBASE-16567 Upgrade to protobuf-3.1.x
Regenerate all protos in this module with protoc3. Redo ByteStringer to use
new pb3.1.0 unsafebytesutil instead of HBaseZeroCopyByteString
http://git-wip-us.apache.org/repos/asf/hbase/blob/b4a729ed/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AdminProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AdminProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AdminProtos.java
index f54d238..e15dbc4 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AdminProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AdminProtos.java
@@ -6,12 +6,18 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class AdminProtos {
private AdminProtos() {}
public static void registerAllExtensions(
+ com.google.protobuf.ExtensionRegistryLite registry) {
+ }
+
+ public static void registerAllExtensions(
com.google.protobuf.ExtensionRegistry registry) {
+ registerAllExtensions(
+ (com.google.protobuf.ExtensionRegistryLite) registry);
}
- public interface GetRegionInfoRequestOrBuilder
- extends com.google.protobuf.MessageOrBuilder {
+ public interface GetRegionInfoRequestOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:hbase.pb.GetRegionInfoRequest)
+ com.google.protobuf.MessageOrBuilder {
- // required .hbase.pb.RegionSpecifier region = 1;
/**
* <code>required .hbase.pb.RegionSpecifier region = 1;</code>
*/
@@ -25,7 +31,6 @@ public final class AdminProtos {
*/
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
- // optional bool compaction_state = 2;
/**
* <code>optional bool compaction_state = 2;</code>
*/
@@ -38,36 +43,28 @@ public final class AdminProtos {
/**
* Protobuf type {@code hbase.pb.GetRegionInfoRequest}
*/
- public static final class GetRegionInfoRequest extends
- com.google.protobuf.GeneratedMessage
- implements GetRegionInfoRequestOrBuilder {
+ public static final class GetRegionInfoRequest extends
+ com.google.protobuf.GeneratedMessageV3 implements
+ // @@protoc_insertion_point(message_implements:hbase.pb.GetRegionInfoRequest)
+ GetRegionInfoRequestOrBuilder {
// Use GetRegionInfoRequest.newBuilder() to construct.
- private GetRegionInfoRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ private GetRegionInfoRequest(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
- this.unknownFields = builder.getUnknownFields();
- }
- private GetRegionInfoRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
- private static final GetRegionInfoRequest defaultInstance;
- public static GetRegionInfoRequest getDefaultInstance() {
- return defaultInstance;
}
-
- public GetRegionInfoRequest getDefaultInstanceForType() {
- return defaultInstance;
+ private GetRegionInfoRequest() {
+ compactionState_ = false;
}
- private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
+ getUnknownFields() {
return this.unknownFields;
}
private GetRegionInfoRequest(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- initFields();
+ this();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
@@ -110,7 +107,7 @@ public final class AdminProtos {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
- e.getMessage()).setUnfinishedMessage(this);
+ e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
@@ -121,30 +118,14 @@ public final class AdminProtos {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoRequest_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoRequest_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest.Builder.class);
}
- public static com.google.protobuf.Parser<GetRegionInfoRequest> PARSER =
- new com.google.protobuf.AbstractParser<GetRegionInfoRequest>() {
- public GetRegionInfoRequest parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return new GetRegionInfoRequest(input, extensionRegistry);
- }
- };
-
- @java.lang.Override
- public com.google.protobuf.Parser<GetRegionInfoRequest> getParserForType() {
- return PARSER;
- }
-
private int bitField0_;
- // required .hbase.pb.RegionSpecifier region = 1;
public static final int REGION_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier region_;
/**
@@ -157,16 +138,15 @@ public final class AdminProtos {
* <code>required .hbase.pb.RegionSpecifier region = 1;</code>
*/
public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() {
- return region_;
+ return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_;
}
/**
* <code>required .hbase.pb.RegionSpecifier region = 1;</code>
*/
public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() {
- return region_;
+ return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_;
}
- // optional bool compaction_state = 2;
public static final int COMPACTION_STATE_FIELD_NUMBER = 2;
private boolean compactionState_;
/**
@@ -182,14 +162,11 @@ public final class AdminProtos {
return compactionState_;
}
- private void initFields() {
- region_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
- compactionState_ = false;
- }
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
- if (isInitialized != -1) return isInitialized == 1;
+ if (isInitialized == 1) return true;
+ if (isInitialized == 0) return false;
if (!hasRegion()) {
memoizedIsInitialized = 0;
@@ -205,43 +182,35 @@ public final class AdminProtos {
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
- getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- output.writeMessage(1, region_);
+ output.writeMessage(1, getRegion());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
output.writeBool(2, compactionState_);
}
- getUnknownFields().writeTo(output);
+ unknownFields.writeTo(output);
}
- private int memoizedSerializedSize = -1;
public int getSerializedSize() {
- int size = memoizedSerializedSize;
+ int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
- .computeMessageSize(1, region_);
+ .computeMessageSize(1, getRegion());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
size += com.google.protobuf.CodedOutputStream
.computeBoolSize(2, compactionState_);
}
- size += getUnknownFields().getSerializedSize();
- memoizedSerializedSize = size;
+ size += unknownFields.getSerializedSize();
+ memoizedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
- protected java.lang.Object writeReplace()
- throws java.io.ObjectStreamException {
- return super.writeReplace();
- }
-
- @java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
@@ -262,12 +231,10 @@ public final class AdminProtos {
result = result && (getCompactionState()
== other.getCompactionState());
}
- result = result &&
- getUnknownFields().equals(other.getUnknownFields());
+ result = result && unknownFields.equals(other.unknownFields);
return result;
}
- private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
@@ -281,9 +248,10 @@ public final class AdminProtos {
}
if (hasCompactionState()) {
hash = (37 * hash) + COMPACTION_STATE_FIELD_NUMBER;
- hash = (53 * hash) + hashBoolean(getCompactionState());
+ hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean(
+ getCompactionState());
}
- hash = (29 * hash) + getUnknownFields().hashCode();
+ hash = (29 * hash) + unknownFields.hashCode();
memoizedHashCode = hash;
return hash;
}
@@ -311,46 +279,57 @@ public final class AdminProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
- public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder() {
+ return DEFAULT_INSTANCE.toBuilder();
+ }
public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest prototype) {
- return newBuilder().mergeFrom(prototype);
+ return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+ }
+ public Builder toBuilder() {
+ return this == DEFAULT_INSTANCE
+ ? new Builder() : new Builder().mergeFrom(this);
}
- public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -358,14 +337,15 @@ public final class AdminProtos {
* Protobuf type {@code hbase.pb.GetRegionInfoRequest}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequestOrBuilder {
+ com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ // @@protoc_insertion_point(builder_implements:hbase.pb.GetRegionInfoRequest)
+ org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequestOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoRequest_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoRequest_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -378,23 +358,20 @@ public final class AdminProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ if (com.google.protobuf.GeneratedMessageV3
+ .alwaysUseFieldBuilders) {
getRegionFieldBuilder();
}
}
- private static Builder create() {
- return new Builder();
- }
-
public Builder clear() {
super.clear();
if (regionBuilder_ == null) {
- region_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
+ region_ = null;
} else {
regionBuilder_.clear();
}
@@ -404,10 +381,6 @@ public final class AdminProtos {
return this;
}
- public Builder clone() {
- return create().mergeFrom(buildPartial());
- }
-
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoRequest_descriptor;
@@ -446,6 +419,32 @@ public final class AdminProtos {
return result;
}
+ public Builder clone() {
+ return (Builder) super.clone();
+ }
+ public Builder setField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ Object value) {
+ return (Builder) super.setField(field, value);
+ }
+ public Builder clearField(
+ com.google.protobuf.Descriptors.FieldDescriptor field) {
+ return (Builder) super.clearField(field);
+ }
+ public Builder clearOneof(
+ com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ return (Builder) super.clearOneof(oneof);
+ }
+ public Builder setRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ int index, Object value) {
+ return (Builder) super.setRepeatedField(field, index, value);
+ }
+ public Builder addRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ Object value) {
+ return (Builder) super.addRepeatedField(field, value);
+ }
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest)other);
@@ -463,17 +462,16 @@ public final class AdminProtos {
if (other.hasCompactionState()) {
setCompactionState(other.getCompactionState());
}
- this.mergeUnknownFields(other.getUnknownFields());
+ this.mergeUnknownFields(other.unknownFields);
+ onChanged();
return this;
}
public final boolean isInitialized() {
if (!hasRegion()) {
-
return false;
}
if (!getRegion().isInitialized()) {
-
return false;
}
return true;
@@ -488,7 +486,7 @@ public final class AdminProtos {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest) e.getUnfinishedMessage();
- throw e;
+ throw e.unwrapIOException();
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
@@ -498,9 +496,8 @@ public final class AdminProtos {
}
private int bitField0_;
- // required .hbase.pb.RegionSpecifier region = 1;
- private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier region_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
- private com.google.protobuf.SingleFieldBuilder<
+ private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier region_ = null;
+ private com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBuilder_;
/**
* <code>required .hbase.pb.RegionSpecifier region = 1;</code>
@@ -513,7 +510,7 @@ public final class AdminProtos {
*/
public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() {
if (regionBuilder_ == null) {
- return region_;
+ return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_;
} else {
return regionBuilder_.getMessage();
}
@@ -554,6 +551,7 @@ public final class AdminProtos {
public Builder mergeRegion(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) {
if (regionBuilder_ == null) {
if (((bitField0_ & 0x00000001) == 0x00000001) &&
+ region_ != null &&
region_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) {
region_ =
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(region_).mergeFrom(value).buildPartial();
@@ -572,7 +570,7 @@ public final class AdminProtos {
*/
public Builder clearRegion() {
if (regionBuilder_ == null) {
- region_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
+ region_ = null;
onChanged();
} else {
regionBuilder_.clear();
@@ -595,19 +593,20 @@ public final class AdminProtos {
if (regionBuilder_ != null) {
return regionBuilder_.getMessageOrBuilder();
} else {
- return region_;
+ return region_ == null ?
+ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_;
}
}
/**
* <code>required .hbase.pb.RegionSpecifier region = 1;</code>
*/
- private com.google.protobuf.SingleFieldBuilder<
+ private com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>
getRegionFieldBuilder() {
if (regionBuilder_ == null) {
- regionBuilder_ = new com.google.protobuf.SingleFieldBuilder<
+ regionBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>(
- region_,
+ getRegion(),
getParentForChildren(),
isClean());
region_ = null;
@@ -615,7 +614,6 @@ public final class AdminProtos {
return regionBuilder_;
}
- // optional bool compaction_state = 2;
private boolean compactionState_ ;
/**
* <code>optional bool compaction_state = 2;</code>
@@ -647,22 +645,59 @@ public final class AdminProtos {
onChanged();
return this;
}
+ public final Builder setUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.setUnknownFields(unknownFields);
+ }
+
+ public final Builder mergeUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.mergeUnknownFields(unknownFields);
+ }
+
// @@protoc_insertion_point(builder_scope:hbase.pb.GetRegionInfoRequest)
}
+ // @@protoc_insertion_point(class_scope:hbase.pb.GetRegionInfoRequest)
+ private static final org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest DEFAULT_INSTANCE;
static {
- defaultInstance = new GetRegionInfoRequest(true);
- defaultInstance.initFields();
+ DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest();
+ }
+
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest getDefaultInstance() {
+ return DEFAULT_INSTANCE;
+ }
+
+ @java.lang.Deprecated public static final com.google.protobuf.Parser<GetRegionInfoRequest>
+ PARSER = new com.google.protobuf.AbstractParser<GetRegionInfoRequest>() {
+ public GetRegionInfoRequest parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new GetRegionInfoRequest(input, extensionRegistry);
+ }
+ };
+
+ public static com.google.protobuf.Parser<GetRegionInfoRequest> parser() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<GetRegionInfoRequest> getParserForType() {
+ return PARSER;
+ }
+
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest getDefaultInstanceForType() {
+ return DEFAULT_INSTANCE;
}
- // @@protoc_insertion_point(class_scope:hbase.pb.GetRegionInfoRequest)
}
- public interface GetRegionInfoResponseOrBuilder
- extends com.google.protobuf.MessageOrBuilder {
+ public interface GetRegionInfoResponseOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:hbase.pb.GetRegionInfoResponse)
+ com.google.protobuf.MessageOrBuilder {
- // required .hbase.pb.RegionInfo region_info = 1;
/**
* <code>required .hbase.pb.RegionInfo region_info = 1;</code>
*/
@@ -676,7 +711,6 @@ public final class AdminProtos {
*/
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionInfoOrBuilder();
- // optional .hbase.pb.GetRegionInfoResponse.CompactionState compaction_state = 2;
/**
* <code>optional .hbase.pb.GetRegionInfoResponse.CompactionState compaction_state = 2;</code>
*/
@@ -686,7 +720,6 @@ public final class AdminProtos {
*/
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState getCompactionState();
- // optional bool isRecovering = 3;
/**
* <code>optional bool isRecovering = 3;</code>
*/
@@ -699,36 +732,29 @@ public final class AdminProtos {
/**
* Protobuf type {@code hbase.pb.GetRegionInfoResponse}
*/
- public static final class GetRegionInfoResponse extends
- com.google.protobuf.GeneratedMessage
- implements GetRegionInfoResponseOrBuilder {
+ public static final class GetRegionInfoResponse extends
+ com.google.protobuf.GeneratedMessageV3 implements
+ // @@protoc_insertion_point(message_implements:hbase.pb.GetRegionInfoResponse)
+ GetRegionInfoResponseOrBuilder {
// Use GetRegionInfoResponse.newBuilder() to construct.
- private GetRegionInfoResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ private GetRegionInfoResponse(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
- this.unknownFields = builder.getUnknownFields();
- }
- private GetRegionInfoResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
- private static final GetRegionInfoResponse defaultInstance;
- public static GetRegionInfoResponse getDefaultInstance() {
- return defaultInstance;
}
-
- public GetRegionInfoResponse getDefaultInstanceForType() {
- return defaultInstance;
+ private GetRegionInfoResponse() {
+ compactionState_ = 0;
+ isRecovering_ = false;
}
- private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
+ getUnknownFields() {
return this.unknownFields;
}
private GetRegionInfoResponse(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- initFields();
+ this();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
@@ -767,7 +793,7 @@ public final class AdminProtos {
unknownFields.mergeVarintField(2, rawValue);
} else {
bitField0_ |= 0x00000002;
- compactionState_ = value;
+ compactionState_ = rawValue;
}
break;
}
@@ -782,7 +808,7 @@ public final class AdminProtos {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
- e.getMessage()).setUnfinishedMessage(this);
+ e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
@@ -793,28 +819,13 @@ public final class AdminProtos {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoResponse_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoResponse_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.Builder.class);
}
- public static com.google.protobuf.Parser<GetRegionInfoResponse> PARSER =
- new com.google.protobuf.AbstractParser<GetRegionInfoResponse>() {
- public GetRegionInfoResponse parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return new GetRegionInfoResponse(input, extensionRegistry);
- }
- };
-
- @java.lang.Override
- public com.google.protobuf.Parser<GetRegionInfoResponse> getParserForType() {
- return PARSER;
- }
-
/**
* Protobuf enum {@code hbase.pb.GetRegionInfoResponse.CompactionState}
*/
@@ -823,19 +834,19 @@ public final class AdminProtos {
/**
* <code>NONE = 0;</code>
*/
- NONE(0, 0),
+ NONE(0),
/**
* <code>MINOR = 1;</code>
*/
- MINOR(1, 1),
+ MINOR(1),
/**
* <code>MAJOR = 2;</code>
*/
- MAJOR(2, 2),
+ MAJOR(2),
/**
* <code>MAJOR_AND_MINOR = 3;</code>
*/
- MAJOR_AND_MINOR(3, 3),
+ MAJOR_AND_MINOR(3),
;
/**
@@ -856,9 +867,19 @@ public final class AdminProtos {
public static final int MAJOR_AND_MINOR_VALUE = 3;
- public final int getNumber() { return value; }
+ public final int getNumber() {
+ return value;
+ }
+ /**
+ * @deprecated Use {@link #forNumber(int)} instead.
+ */
+ @java.lang.Deprecated
public static CompactionState valueOf(int value) {
+ return forNumber(value);
+ }
+
+ public static CompactionState forNumber(int value) {
switch (value) {
case 0: return NONE;
case 1: return MINOR;
@@ -872,17 +893,17 @@ public final class AdminProtos {
internalGetValueMap() {
return internalValueMap;
}
- private static com.google.protobuf.Internal.EnumLiteMap<CompactionState>
- internalValueMap =
+ private static final com.google.protobuf.Internal.EnumLiteMap<
+ CompactionState> internalValueMap =
new com.google.protobuf.Internal.EnumLiteMap<CompactionState>() {
public CompactionState findValueByNumber(int number) {
- return CompactionState.valueOf(number);
+ return CompactionState.forNumber(number);
}
};
public final com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
- return getDescriptor().getValues().get(index);
+ return getDescriptor().getValues().get(ordinal());
}
public final com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
@@ -904,11 +925,9 @@ public final class AdminProtos {
return VALUES[desc.getIndex()];
}
- private final int index;
private final int value;
- private CompactionState(int index, int value) {
- this.index = index;
+ private CompactionState(int value) {
this.value = value;
}
@@ -916,7 +935,6 @@ public final class AdminProtos {
}
private int bitField0_;
- // required .hbase.pb.RegionInfo region_info = 1;
public static final int REGION_INFO_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo regionInfo_;
/**
@@ -929,18 +947,17 @@ public final class AdminProtos {
* <code>required .hbase.pb.RegionInfo region_info = 1;</code>
*/
public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo() {
- return regionInfo_;
+ return regionInfo_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance() : regionInfo_;
}
/**
* <code>required .hbase.pb.RegionInfo region_info = 1;</code>
*/
public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionInfoOrBuilder() {
- return regionInfo_;
+ return regionInfo_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance() : regionInfo_;
}
- // optional .hbase.pb.GetRegionInfoResponse.CompactionState compaction_state = 2;
public static final int COMPACTION_STATE_FIELD_NUMBER = 2;
- private org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState compactionState_;
+ private int compactionState_;
/**
* <code>optional .hbase.pb.GetRegionInfoResponse.CompactionState compaction_state = 2;</code>
*/
@@ -951,10 +968,10 @@ public final class AdminProtos {
* <code>optional .hbase.pb.GetRegionInfoResponse.CompactionState compaction_state = 2;</code>
*/
public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState getCompactionState() {
- return compactionState_;
+ org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState result = org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState.valueOf(compactionState_);
+ return result == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState.NONE : result;
}
- // optional bool isRecovering = 3;
public static final int ISRECOVERING_FIELD_NUMBER = 3;
private boolean isRecovering_;
/**
@@ -970,15 +987,11 @@ public final class AdminProtos {
return isRecovering_;
}
- private void initFields() {
- regionInfo_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance();
- compactionState_ = org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState.NONE;
- isRecovering_ = false;
- }
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
- if (isInitialized != -1) return isInitialized == 1;
+ if (isInitialized == 1) return true;
+ if (isInitialized == 0) return false;
if (!hasRegionInfo()) {
memoizedIsInitialized = 0;
@@ -994,50 +1007,42 @@ public final class AdminProtos {
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
- getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- output.writeMessage(1, regionInfo_);
+ output.writeMessage(1, getRegionInfo());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- output.writeEnum(2, compactionState_.getNumber());
+ output.writeEnum(2, compactionState_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
output.writeBool(3, isRecovering_);
}
- getUnknownFields().writeTo(output);
+ unknownFields.writeTo(output);
}
- private int memoizedSerializedSize = -1;
public int getSerializedSize() {
- int size = memoizedSerializedSize;
+ int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
- .computeMessageSize(1, regionInfo_);
+ .computeMessageSize(1, getRegionInfo());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
size += com.google.protobuf.CodedOutputStream
- .computeEnumSize(2, compactionState_.getNumber());
+ .computeEnumSize(2, compactionState_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
size += com.google.protobuf.CodedOutputStream
.computeBoolSize(3, isRecovering_);
}
- size += getUnknownFields().getSerializedSize();
- memoizedSerializedSize = size;
+ size += unknownFields.getSerializedSize();
+ memoizedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
- protected java.lang.Object writeReplace()
- throws java.io.ObjectStreamException {
- return super.writeReplace();
- }
-
- @java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
@@ -1055,20 +1060,17 @@ public final class AdminProtos {
}
result = result && (hasCompactionState() == other.hasCompactionState());
if (hasCompactionState()) {
- result = result &&
- (getCompactionState() == other.getCompactionState());
+ result = result && compactionState_ == other.compactionState_;
}
result = result && (hasIsRecovering() == other.hasIsRecovering());
if (hasIsRecovering()) {
result = result && (getIsRecovering()
== other.getIsRecovering());
}
- result = result &&
- getUnknownFields().equals(other.getUnknownFields());
+ result = result && unknownFields.equals(other.unknownFields);
return result;
}
- private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
@@ -1082,13 +1084,14 @@ public final class AdminProtos {
}
if (hasCompactionState()) {
hash = (37 * hash) + COMPACTION_STATE_FIELD_NUMBER;
- hash = (53 * hash) + hashEnum(getCompactionState());
+ hash = (53 * hash) + compactionState_;
}
if (hasIsRecovering()) {
hash = (37 * hash) + ISRECOVERING_FIELD_NUMBER;
- hash = (53 * hash) + hashBoolean(getIsRecovering());
+ hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean(
+ getIsRecovering());
}
- hash = (29 * hash) + getUnknownFields().hashCode();
+ hash = (29 * hash) + unknownFields.hashCode();
memoizedHashCode = hash;
return hash;
}
@@ -1116,46 +1119,57 @@ public final class AdminProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
- public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder() {
+ return DEFAULT_INSTANCE.toBuilder();
+ }
public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse prototype) {
- return newBuilder().mergeFrom(prototype);
+ return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+ }
+ public Builder toBuilder() {
+ return this == DEFAULT_INSTANCE
+ ? new Builder() : new Builder().mergeFrom(this);
}
- public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1163,14 +1177,15 @@ public final class AdminProtos {
* Protobuf type {@code hbase.pb.GetRegionInfoResponse}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponseOrBuilder {
+ com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ // @@protoc_insertion_point(builder_implements:hbase.pb.GetRegionInfoResponse)
+ org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponseOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoResponse_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoResponse_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1183,38 +1198,31 @@ public final class AdminProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ if (com.google.protobuf.GeneratedMessageV3
+ .alwaysUseFieldBuilders) {
getRegionInfoFieldBuilder();
}
}
- private static Builder create() {
- return new Builder();
- }
-
public Builder clear() {
super.clear();
if (regionInfoBuilder_ == null) {
- regionInfo_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance();
+ regionInfo_ = null;
} else {
regionInfoBuilder_.clear();
}
bitField0_ = (bitField0_ & ~0x00000001);
- compactionState_ = org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState.NONE;
+ compactionState_ = 0;
bitField0_ = (bitField0_ & ~0x00000002);
isRecovering_ = false;
bitField0_ = (bitField0_ & ~0x00000004);
return this;
}
- public Builder clone() {
- return create().mergeFrom(buildPartial());
- }
-
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoResponse_descriptor;
@@ -1257,6 +1265,32 @@ public final class AdminProtos {
return result;
}
+ public Builder clone() {
+ return (Builder) super.clone();
+ }
+ public Builder setField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ Object value) {
+ return (Builder) super.setField(field, value);
+ }
+ public Builder clearField(
+ com.google.protobuf.Descriptors.FieldDescriptor field) {
+ return (Builder) super.clearField(field);
+ }
+ public Builder clearOneof(
+ com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ return (Builder) super.clearOneof(oneof);
+ }
+ public Builder setRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ int index, Object value) {
+ return (Builder) super.setRepeatedField(field, index, value);
+ }
+ public Builder addRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ Object value) {
+ return (Builder) super.addRepeatedField(field, value);
+ }
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse)other);
@@ -1277,17 +1311,16 @@ public final class AdminProtos {
if (other.hasIsRecovering()) {
setIsRecovering(other.getIsRecovering());
}
- this.mergeUnknownFields(other.getUnknownFields());
+ this.mergeUnknownFields(other.unknownFields);
+ onChanged();
return this;
}
public final boolean isInitialized() {
if (!hasRegionInfo()) {
-
return false;
}
if (!getRegionInfo().isInitialized()) {
-
return false;
}
return true;
@@ -1302,7 +1335,7 @@ public final class AdminProtos {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse) e.getUnfinishedMessage();
- throw e;
+ throw e.unwrapIOException();
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
@@ -1312,9 +1345,8 @@ public final class AdminProtos {
}
private int bitField0_;
- // required .hbase.pb.RegionInfo region_info = 1;
- private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo regionInfo_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance();
- private com.google.protobuf.SingleFieldBuilder<
+ private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo regionInfo_ = null;
+ private com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder> regionInfoBuilder_;
/**
* <code>required .hbase.pb.RegionInfo region_info = 1;</code>
@@ -1327,7 +1359,7 @@ public final class AdminProtos {
*/
public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo() {
if (regionInfoBuilder_ == null) {
- return regionInfo_;
+ return regionInfo_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance() : regionInfo_;
} else {
return regionInfoBuilder_.getMessage();
}
@@ -1368,6 +1400,7 @@ public final class AdminProtos {
public Builder mergeRegionInfo(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo value) {
if (regionInfoBuilder_ == null) {
if (((bitField0_ & 0x00000001) == 0x00000001) &&
+ regionInfo_ != null &&
regionInfo_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance()) {
regionInfo_ =
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.newBuilder(regionInfo_).mergeFrom(value).buildPartial();
@@ -1386,7 +1419,7 @@ public final class AdminProtos {
*/
public Builder clearRegionInfo() {
if (regionInfoBuilder_ == null) {
- regionInfo_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance();
+ regionInfo_ = null;
onChanged();
} else {
regionInfoBuilder_.clear();
@@ -1409,19 +1442,20 @@ public final class AdminProtos {
if (regionInfoBuilder_ != null) {
return regionInfoBuilder_.getMessageOrBuilder();
} else {
- return regionInfo_;
+ return regionInfo_ == null ?
+ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance() : regionInfo_;
}
}
/**
* <code>required .hbase.pb.RegionInfo region_info = 1;</code>
*/
- private com.google.protobuf.SingleFieldBuilder<
+ private com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder>
getRegionInfoFieldBuilder() {
if (regionInfoBuilder_ == null) {
- regionInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
+ regionInfoBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder>(
- regionInfo_,
+ getRegionInfo(),
getParentForChildren(),
isClean());
regionInfo_ = null;
@@ -1429,8 +1463,7 @@ public final class AdminProtos {
return regionInfoBuilder_;
}
- // optional .hbase.pb.GetRegionInfoResponse.CompactionState compaction_state = 2;
- private org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState compactionState_ = org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState.NONE;
+ private int compactionState_ = 0;
/**
* <code>optional .hbase.pb.GetRegionInfoResponse.CompactionState compaction_state = 2;</code>
*/
@@ -1441,7 +1474,8 @@ public final class AdminProtos {
* <code>optional .hbase.pb.GetRegionInfoResponse.CompactionState compaction_state = 2;</code>
*/
public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState getCompactionState() {
- return compactionState_;
+ org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState result = org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState.valueOf(compactionState_);
+ return result == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState.NONE : result;
}
/**
* <code>optional .hbase.pb.GetRegionInfoResponse.CompactionState compaction_state = 2;</code>
@@ -1451,7 +1485,7 @@ public final class AdminProtos {
throw new NullPointerException();
}
bitField0_ |= 0x00000002;
- compactionState_ = value;
+ compactionState_ = value.getNumber();
onChanged();
return this;
}
@@ -1460,12 +1494,11 @@ public final class AdminProtos {
*/
public Builder clearCompactionState() {
bitField0_ = (bitField0_ & ~0x00000002);
- compactionState_ = org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState.NONE;
+ compactionState_ = 0;
onChanged();
return this;
}
- // optional bool isRecovering = 3;
private boolean isRecovering_ ;
/**
* <code>optional bool isRecovering = 3;</code>
@@ -1497,22 +1530,59 @@ public final class AdminProtos {
onChanged();
return this;
}
+ public final Builder setUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.setUnknownFields(unknownFields);
+ }
+
+ public final Builder mergeUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.mergeUnknownFields(unknownFields);
+ }
+
// @@protoc_insertion_point(builder_scope:hbase.pb.GetRegionInfoResponse)
}
+ // @@protoc_insertion_point(class_scope:hbase.pb.GetRegionInfoResponse)
+ private static final org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse DEFAULT_INSTANCE;
static {
- defaultInstance = new GetRegionInfoResponse(true);
- defaultInstance.initFields();
+ DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse();
+ }
+
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse getDefaultInstance() {
+ return DEFAULT_INSTANCE;
+ }
+
+ @java.lang.Deprecated public static final com.google.protobuf.Parser<GetRegionInfoResponse>
+ PARSER = new com.google.protobuf.AbstractParser<GetRegionInfoResponse>() {
+ public GetRegionInfoResponse parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new GetRegionInfoResponse(input, extensionRegistry);
+ }
+ };
+
+ public static com.google.protobuf.Parser<GetRegionInfoResponse> parser() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<GetRegionInfoResponse> getParserForType() {
+ return PARSER;
+ }
+
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse getDefaultInstanceForType() {
+ return DEFAULT_INSTANCE;
}
- // @@protoc_insertion_point(class_scope:hbase.pb.GetRegionInfoResponse)
}
- public interface GetStoreFileRequestOrBuilder
- extends com.google.protobuf.MessageOrBuilder {
+ public interface GetStoreFileRequestOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:hbase.pb.GetStoreFileRequest)
+ com.google.protobuf.MessageOrBuilder {
- // required .hbase.pb.RegionSpecifier region = 1;
/**
* <code>required .hbase.pb.RegionSpecifier region = 1;</code>
*/
@@ -1526,7 +1596,6 @@ public final class AdminProtos {
*/
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
- // repeated bytes family = 2;
/**
* <code>repeated bytes family = 2;</code>
*/
@@ -1541,44 +1610,36 @@ public final class AdminProtos {
com.google.protobuf.ByteString getFamily(int index);
}
/**
- * Protobuf type {@code hbase.pb.GetStoreFileRequest}
- *
* <pre>
**
* Get a list of store files for a set of column families in a particular region.
* If no column family is specified, get the store files for all column families.
* </pre>
+ *
+ * Protobuf type {@code hbase.pb.GetStoreFileRequest}
*/
- public static final class GetStoreFileRequest extends
- com.google.protobuf.GeneratedMessage
- implements GetStoreFileRequestOrBuilder {
+ public static final class GetStoreFileRequest extends
+ com.google.protobuf.GeneratedMessageV3 implements
+ // @@protoc_insertion_point(message_implements:hbase.pb.GetStoreFileRequest)
+ GetStoreFileRequestOrBuilder {
// Use GetStoreFileRequest.newBuilder() to construct.
- private GetStoreFileRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ private GetStoreFileRequest(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
- this.unknownFields = builder.getUnknownFields();
}
- private GetStoreFileRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
- private static final GetStoreFileRequest defaultInstance;
- public static GetStoreFileRequest getDefaultInstance() {
- return defaultInstance;
- }
-
- public GetStoreFileRequest getDefaultInstanceForType() {
- return defaultInstance;
+ private GetStoreFileRequest() {
+ family_ = java.util.Collections.emptyList();
}
- private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
+ getUnknownFields() {
return this.unknownFields;
}
private GetStoreFileRequest(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- initFields();
+ this();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
@@ -1624,7 +1685,7 @@ public final class AdminProtos {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
- e.getMessage()).setUnfinishedMessage(this);
+ e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
family_ = java.util.Collections.unmodifiableList(family_);
@@ -1638,30 +1699,14 @@ public final class AdminProtos {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetStoreFileRequest_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetStoreFileRequest_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest.Builder.class);
}
- public static com.google.protobuf.Parser<GetStoreFileRequest> PARSER =
- new com.google.protobuf.AbstractParser<GetStoreFileRequest>() {
- public GetStoreFileRequest parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return new GetStoreFileRequest(input, extensionRegistry);
- }
- };
-
- @java.lang.Override
- public com.google.protobuf.Parser<GetStoreFileRequest> getParserForType() {
- return PARSER;
- }
-
private int bitField0_;
- // required .hbase.pb.RegionSpecifier region = 1;
public static final int REGION_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier region_;
/**
@@ -1674,16 +1719,15 @@ public final class AdminProtos {
* <code>required .hbase.pb.RegionSpecifier region = 1;</code>
*/
public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() {
- return region_;
+ return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_;
}
/**
* <code>required .hbase.pb.RegionSpecifier region = 1;</code>
*/
public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() {
- return region_;
+ return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_;
}
- // repeated bytes family = 2;
public static final int FAMILY_FIELD_NUMBER = 2;
private java.util.List<com.google.protobuf.ByteString> family_;
/**
@@ -1706,14 +1750,11 @@ public final class AdminProtos {
return family_.get(index);
}
- private void initFields() {
- region_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
- family_ = java.util.Collections.emptyList();
- }
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
- if (isInitialized != -1) return isInitialized == 1;
+ if (isInitialized == 1) return true;
+ if (isInitialized == 0) return false;
if (!hasRegion()) {
memoizedIsInitialized = 0;
@@ -1729,25 +1770,23 @@ public final class AdminProtos {
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
- getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- output.writeMessage(1, region_);
+ output.writeMessage(1, getRegion());
}
for (int i = 0; i < family_.size(); i++) {
output.writeBytes(2, family_.get(i));
}
- getUnknownFields().writeTo(output);
+ unknownFields.writeTo(output);
}
- private int memoizedSerializedSize = -1;
public int getSerializedSize() {
- int size = memoizedSerializedSize;
+ int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
- .computeMessageSize(1, region_);
+ .computeMessageSize(1, getRegion());
}
{
int dataSize = 0;
@@ -1758,19 +1797,13 @@ public final class AdminProtos {
size += dataSize;
size += 1 * getFamilyList().size();
}
- size += getUnknownFields().getSerializedSize();
- memoizedSerializedSize = size;
+ size += unknownFields.getSerializedSize();
+ memoizedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
- protected java.lang.Object writeReplace()
- throws java.io.ObjectStreamException {
- return super.writeReplace();
- }
-
- @java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
@@ -1788,12 +1821,10 @@ public final class AdminProtos {
}
result = result && getFamilyList()
.equals(other.getFamilyList());
- result = result &&
- getUnknownFields().equals(other.getUnknownFields());
+ result = result && unknownFields.equals(other.unknownFields);
return result;
}
- private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
@@ -1809,7 +1840,7 @@ public final class AdminProtos {
hash = (37 * hash) + FAMILY_FIELD_NUMBER;
hash = (53 * hash) + getFamilyList().hashCode();
}
- hash = (29 * hash) + getUnknownFields().hashCode();
+ hash = (29 * hash) + unknownFields.hashCode();
memoizedHashCode = hash;
return hash;
}
@@ -1837,67 +1868,79 @@ public final class AdminProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
- public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder() {
+ return DEFAULT_INSTANCE.toBuilder();
+ }
public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest prototype) {
- return newBuilder().mergeFrom(prototype);
+ return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+ }
+ public Builder toBuilder() {
+ return this == DEFAULT_INSTANCE
+ ? new Builder() : new Builder().mergeFrom(this);
}
- public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
- * Protobuf type {@code hbase.pb.GetStoreFileRequest}
- *
* <pre>
**
* Get a list of store files for a set of column families in a particular region.
* If no column family is specified, get the store files for all column families.
* </pre>
+ *
+ * Protobuf type {@code hbase.pb.GetStoreFileRequest}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequestOrBuilder {
+ com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ // @@protoc_insertion_point(builder_implements:hbase.pb.GetStoreFileRequest)
+ org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequestOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetStoreFileRequest_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetStoreFileRequest_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1910,23 +1953,20 @@ public final class AdminProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ if (com.google.protobuf.GeneratedMessageV3
+ .alwaysUseFieldBuilders) {
getRegionFieldBuilder();
}
}
- private static Builder create() {
- return new Builder();
- }
-
public Builder clear() {
super.clear();
if (regionBuilder_ == null) {
- region_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
+ region_ = null;
} else {
regionBuilder_.clear();
}
@@ -1936,10 +1976,6 @@ public final class AdminProtos {
return this;
}
- public Builder clone() {
- return create().mergeFrom(buildPartial());
- }
-
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetStoreFileRequest_descriptor;
@@ -1979,6 +2015,32 @@ public final class AdminProtos {
return result;
}
+ public Builder clone() {
+ return (Builder) super.clone();
+ }
+ public Builder setField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ Object value) {
+ return (Builder) super.setField(field, value);
+ }
+ public Builder clearField(
+ com.google.protobuf.Descriptors.FieldDescriptor field) {
+ return (Builder) super.clearField(field);
+ }
+ public Builder clearOneof(
+ com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ return (Builder) super.clearOneof(oneof);
+ }
+ public Builder setRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ int index, Object value) {
+ return (Builder) super.setRepeatedField(field, index, value);
+ }
+ public Builder addRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ Object value) {
+ return (Builder) super.addRepeatedField(field, value);
+ }
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest)other);
@@ -2003,17 +2065,16 @@ public final class AdminProtos {
}
onChanged();
}
- this.mergeUnknownFields(other.getUnknownFields());
+ this.mergeUnknownFields(other.unknownFields);
+ onChanged();
return this;
}
public final boolean isInitialized() {
if (!hasRegion()) {
-
return false;
}
if (!getRegion().isInitialized()) {
-
return false;
}
return true;
@@ -2028,7 +2089,7 @@ public final class AdminProtos {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest) e.getUnfinishedMessage();
- throw e;
+ throw e.unwrapIOException();
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
@@ -2038,9 +2099,8 @@ public final class AdminProtos {
}
private int bitField0_;
- // required .hbase.pb.RegionSpecifier region = 1;
- private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier region_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
- private com.google.protobuf.SingleFieldBuilder<
+ private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier region_ = null;
+ private com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBuilder_;
/**
* <code>required .hbase.pb.RegionSpecifier region = 1;</code>
@@ -2053,7 +2113,7 @@ public final class AdminProtos {
*/
public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() {
if (regionBuilder_ == null) {
- return region_;
+ return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_;
} else {
return regionBuilder_.getMessage();
}
@@ -2094,6 +2154,7 @@ public final class AdminProtos {
public Builder mergeRegion(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) {
if (regionBuilder_ == null) {
if (((bitField0_ & 0x00000001) == 0x00000001) &&
+ region_ != null &&
region_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) {
region_ =
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(region_).mergeFrom(value).buildPartial();
@@ -2112,7 +2173,7 @@ public final class AdminProtos {
*/
public Builder clearRegion() {
if (regionBuilder_ == null) {
- region_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
+ region_ = null;
onChanged();
} else {
regionBuilder_.clear();
@@ -2135,19 +2196,20 @@ public final class AdminProtos {
if (regionBuilder_ != null) {
return regionBuilder_.getMessageOrBuilder();
} else {
- return region_;
+ return region_ == null ?
+ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_;
}
}
/**
* <code>required .hbase.pb.RegionSpecifier region = 1;</code>
*/
- private com.google.protobuf.SingleFieldBuilder<
+ private com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>
getRegionFieldBuilder() {
if (regionBuilder_ == null) {
- regionBuilder_ = new com.google.protobuf.SingleFieldBuilder<
+ regionBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>(
- region_,
+ getRegion(),
getParentForChildren(),
isClean());
region_ = null;
@@ -2155,7 +2217,6 @@ public final class AdminProtos {
return regionBuilder_;
}
- // repeated bytes family = 2;
private java.util.List<com.google.protobuf.ByteString> family_ = java.util.Collections.emptyList();
private void ensureFamilyIsMutable() {
if (!((bitField0_ & 0x00000002) == 0x00000002)) {
@@ -2213,7 +2274,8 @@ public final class AdminProtos {
public Builder addAllFamily(
java.lang.Iterable<? extends com.google.protobuf.ByteString> values) {
ensureFamilyIsMutable();
- super.addAll(values, family_);
+ com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ values, family_);
onChanged();
return this;
}
@@ -2226,27 +2288,64 @@ public final class AdminProtos {
onChanged();
return this;
}
+ public final Builder setUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.setUnknownFields(unknownFields);
+ }
+
+ public final Builder mergeUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.mergeUnknownFields(unknownFields);
+ }
+
// @@protoc_insertion_point(builder_scope:hbase.pb.GetStoreFileRequest)
}
+ // @@protoc_insertion_point(class_scope:hbase.pb.GetStoreFileRequest)
+ private static final org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest DEFAULT_INSTANCE;
static {
- defaultInstance = new GetStoreFileRequest(true);
- defaultInstance.initFields();
+ DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest();
+ }
+
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest getDefaultInstance() {
+ return DEFAULT_INSTANCE;
+ }
+
+ @java.lang.Deprecated public static final com.google.protobuf.Parser<GetStoreFileRequest>
+ PARSER = new com.google.protobuf.AbstractParser<GetStoreFileRequest>() {
+ public GetStoreFileRequest parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new GetStoreFileRequest(input, extensionRegistry);
+ }
+ };
+
+ public static com.google.protobuf.Parser<GetStoreFileRequest> parser() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<GetStoreFileRequest> getParserForType() {
+ return PARSER;
+ }
+
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest getDefaultInstanceForType() {
+ return DEFAULT_INSTANCE;
}
- // @@protoc_insertion_point(class_scope:hbase.pb.GetStoreFileRequest)
}
- public interface GetStoreFileResponseOrBuilder
- extends com.google.protobuf.MessageOrBuilder {
+ public interface GetStoreFileResponseOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:hbase.pb.GetStoreFileResponse)
+ com.google.protobuf.MessageOrBuilder {
- // repeated string store_file = 1;
/**
* <code>repeated string store_file = 1;</code>
*/
java.util.List<java.lang.String>
- getStoreFileList();
+ getStoreFileList();
/**
* <code>repeated string store_file = 1;</code>
*/
@@ -2264,36 +2363,28 @@ public final class AdminProtos {
/**
* Protobuf type {@code hbase.pb.GetStoreFileResponse}
*/
- public static final class GetStoreFileResponse extends
- com.google.protobuf.GeneratedMessage
- implements GetStoreFileResponseOrBuilder {
+ public static final class GetStoreFileResponse extends
+ com.google.protobuf.GeneratedMessageV3 implements
+ // @@protoc_insertion_point(message_implements:hbase.pb.GetStoreFileResponse)
+ GetStoreFileResponseOrBuilder {
// Use GetStoreFileResponse.newBuilder() to construct.
- private GetStoreFileResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ private GetStoreFileResponse(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
- this.unknownFields = builder.getUnknownFields();
- }
- private GetStoreFileResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
- private static final GetStoreFileResponse defaultInstance;
- public static GetStoreFileResponse getDefaultInstance() {
- return defaultInstance;
}
-
- public GetStoreFileResponse getDefaultInstanceForType() {
- return defaultInstance;
+ private GetStoreFileResponse() {
+ storeFile_ = com.google.protobuf.LazyStringArrayList.EMPTY;
}
- private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
+ getUnknownFields() {
return this.unknownFields;
}
private GetStoreFileResponse(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- initFields();
+ this();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
@@ -2313,11 +2404,12 @@ public final class AdminProtos {
break;
}
case 10: {
+ com.google.protobuf.ByteString bs = input.readBytes();
if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
storeFile_ = new com.google.protobuf.LazyStringArrayList();
mutable_bitField0_ |= 0x00000001;
}
- storeFile_.add(input.readBytes());
+ storeFile_.add(bs);
break;
}
}
@@ -2326,10 +2418,10 @@ public final class AdminProtos {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
- e.getMessage()).setUnfinishedMessage(this);
+ e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
- storeFile_ = new com.google.protobuf.UnmodifiableLazyStringList(storeFile_);
+ storeFile_ = storeFile_.getUnmodifiableView();
}
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
@@ -2340,35 +2432,19 @@ public final class AdminProtos {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetStoreFileResponse_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetStoreFileResponse_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileResponse.Builder.class);
}
- public static com.google.protobuf.Parser<GetStoreFileResponse> PARSER =
- new com.google.protobuf.AbstractParser<GetStoreFileResponse>() {
- public GetStoreFileResponse parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return new GetStoreFileResponse(input, extensionRegistry);
- }
- };
-
- @java.lang.Override
- public com.google.protobuf.Parser<GetStoreFileResponse> getParserForType() {
- return PARSER;
- }
-
- // repeated string store_file = 1;
public static final int STORE_FILE_FIELD_NUMBER = 1;
private com.google.protobuf.LazyStringList storeFile_;
/**
* <code>repeated string store_file = 1;</code>
*/
- public java.util.List<java.lang.String>
+ public com.google.protobuf.ProtocolStringList
getStoreFileList() {
return storeFile_;
}
@@ -2392,13 +2468,11 @@ public final class AdminProtos {
return storeFile_.getByteString(index);
}
- private void initFields() {
- storeFile_ = com.google.protobuf.LazyStringArrayList.EMPTY;
- }
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
- if (isInitialized != -1) return isInitialized == 1;
+ if (isInitialized == 1) return true;
+ if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
@@ -2406,41 +2480,32 @@ public final class AdminProtos {
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
- getSerializedSize();
for (int i = 0; i < storeFile_.size(); i++) {
- output.writeBytes(1, storeFile_.getByteString(i));
+ com.google.protobuf.GeneratedMessageV3.writeString(output, 1, storeFile_.getRaw(i));
}
- getUnknownFields().writeTo(output);
+ unknownFields.writeTo(output);
}
- private int memoizedSerializedSize = -1;
public int getSerializedSize() {
- int size = memoizedSerializedSize;
+ int size = memoizedSize;
if (size != -1) return size;
size = 0;
{
int dataSize = 0;
for (int i = 0; i < storeFile_.size(); i++) {
- dataSize += com.google.protobuf.CodedOutputStream
- .computeBytesSizeNoTag(storeFile_.getByteString(i));
+ dataSize += computeStringSizeNoTag(storeFile_.getRaw(i));
}
size += dataSize;
size += 1 * getStoreFileList().size();
}
- size += getUnknownFields().getSerializedSize();
- memoizedSerializedSize = size;
+ size += unknownFields.getSerializedSize();
+ memoizedSize = size;
return size;
}
<TRUNCATED>