You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2016/09/30 00:30:45 UTC
[16/29] hbase git commit: HBASE-16567 Upgrade to protobuf-3.1.x
Regenerate all protos in this module with protoc3. Redo ByteStringer to use
new pb3.1.0 unsafebytesutil instead of HBaseZeroCopyByteString
http://git-wip-us.apache.org/repos/asf/hbase/blob/b4a729ed/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FilterProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FilterProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FilterProtos.java
index b63acb2..7204813 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FilterProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FilterProtos.java
@@ -6,12 +6,18 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class FilterProtos {
private FilterProtos() {}
public static void registerAllExtensions(
+ com.google.protobuf.ExtensionRegistryLite registry) {
+ }
+
+ public static void registerAllExtensions(
com.google.protobuf.ExtensionRegistry registry) {
+ registerAllExtensions(
+ (com.google.protobuf.ExtensionRegistryLite) registry);
}
- public interface FilterOrBuilder
- extends com.google.protobuf.MessageOrBuilder {
+ public interface FilterOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:hbase.pb.Filter)
+ com.google.protobuf.MessageOrBuilder {
- // required string name = 1;
/**
* <code>required string name = 1;</code>
*/
@@ -26,7 +32,6 @@ public final class FilterProtos {
com.google.protobuf.ByteString
getNameBytes();
- // optional bytes serialized_filter = 2;
/**
* <code>optional bytes serialized_filter = 2;</code>
*/
@@ -39,36 +44,29 @@ public final class FilterProtos {
/**
* Protobuf type {@code hbase.pb.Filter}
*/
- public static final class Filter extends
- com.google.protobuf.GeneratedMessage
- implements FilterOrBuilder {
+ public static final class Filter extends
+ com.google.protobuf.GeneratedMessageV3 implements
+ // @@protoc_insertion_point(message_implements:hbase.pb.Filter)
+ FilterOrBuilder {
// Use Filter.newBuilder() to construct.
- private Filter(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ private Filter(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
- this.unknownFields = builder.getUnknownFields();
- }
- private Filter(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
- private static final Filter defaultInstance;
- public static Filter getDefaultInstance() {
- return defaultInstance;
}
-
- public Filter getDefaultInstanceForType() {
- return defaultInstance;
+ private Filter() {
+ name_ = "";
+ serializedFilter_ = com.google.protobuf.ByteString.EMPTY;
}
- private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
+ getUnknownFields() {
return this.unknownFields;
}
private Filter(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- initFields();
+ this();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
@@ -88,8 +86,9 @@ public final class FilterProtos {
break;
}
case 10: {
+ com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000001;
- name_ = input.readBytes();
+ name_ = bs;
break;
}
case 18: {
@@ -103,7 +102,7 @@ public final class FilterProtos {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
- e.getMessage()).setUnfinishedMessage(this);
+ e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
@@ -114,32 +113,16 @@ public final class FilterProtos {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_Filter_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_Filter_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter.class, org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter.Builder.class);
}
- public static com.google.protobuf.Parser<Filter> PARSER =
- new com.google.protobuf.AbstractParser<Filter>() {
- public Filter parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return new Filter(input, extensionRegistry);
- }
- };
-
- @java.lang.Override
- public com.google.protobuf.Parser<Filter> getParserForType() {
- return PARSER;
- }
-
private int bitField0_;
- // required string name = 1;
public static final int NAME_FIELD_NUMBER = 1;
- private java.lang.Object name_;
+ private volatile java.lang.Object name_;
/**
* <code>required string name = 1;</code>
*/
@@ -180,7 +163,6 @@ public final class FilterProtos {
}
}
- // optional bytes serialized_filter = 2;
public static final int SERIALIZED_FILTER_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString serializedFilter_;
/**
@@ -196,14 +178,11 @@ public final class FilterProtos {
return serializedFilter_;
}
- private void initFields() {
- name_ = "";
- serializedFilter_ = com.google.protobuf.ByteString.EMPTY;
- }
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
- if (isInitialized != -1) return isInitialized == 1;
+ if (isInitialized == 1) return true;
+ if (isInitialized == 0) return false;
if (!hasName()) {
memoizedIsInitialized = 0;
@@ -215,43 +194,34 @@ public final class FilterProtos {
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
- getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- output.writeBytes(1, getNameBytes());
+ com.google.protobuf.GeneratedMessageV3.writeString(output, 1, name_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
output.writeBytes(2, serializedFilter_);
}
- getUnknownFields().writeTo(output);
+ unknownFields.writeTo(output);
}
- private int memoizedSerializedSize = -1;
public int getSerializedSize() {
- int size = memoizedSerializedSize;
+ int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
- .computeBytesSize(1, getNameBytes());
+ size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, name_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
size += com.google.protobuf.CodedOutputStream
.computeBytesSize(2, serializedFilter_);
}
- size += getUnknownFields().getSerializedSize();
- memoizedSerializedSize = size;
+ size += unknownFields.getSerializedSize();
+ memoizedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
- protected java.lang.Object writeReplace()
- throws java.io.ObjectStreamException {
- return super.writeReplace();
- }
-
- @java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
@@ -272,12 +242,10 @@ public final class FilterProtos {
result = result && getSerializedFilter()
.equals(other.getSerializedFilter());
}
- result = result &&
- getUnknownFields().equals(other.getUnknownFields());
+ result = result && unknownFields.equals(other.unknownFields);
return result;
}
- private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
@@ -293,7 +261,7 @@ public final class FilterProtos {
hash = (37 * hash) + SERIALIZED_FILTER_FIELD_NUMBER;
hash = (53 * hash) + getSerializedFilter().hashCode();
}
- hash = (29 * hash) + getUnknownFields().hashCode();
+ hash = (29 * hash) + unknownFields.hashCode();
memoizedHashCode = hash;
return hash;
}
@@ -321,46 +289,57 @@ public final class FilterProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
- public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder() {
+ return DEFAULT_INSTANCE.toBuilder();
+ }
public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter prototype) {
- return newBuilder().mergeFrom(prototype);
+ return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+ }
+ public Builder toBuilder() {
+ return this == DEFAULT_INSTANCE
+ ? new Builder() : new Builder().mergeFrom(this);
}
- public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -368,14 +347,15 @@ public final class FilterProtos {
* Protobuf type {@code hbase.pb.Filter}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.FilterOrBuilder {
+ com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ // @@protoc_insertion_point(builder_implements:hbase.pb.Filter)
+ org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.FilterOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_Filter_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_Filter_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -388,18 +368,15 @@ public final class FilterProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ if (com.google.protobuf.GeneratedMessageV3
+ .alwaysUseFieldBuilders) {
}
}
- private static Builder create() {
- return new Builder();
- }
-
public Builder clear() {
super.clear();
name_ = "";
@@ -409,10 +386,6 @@ public final class FilterProtos {
return this;
}
- public Builder clone() {
- return create().mergeFrom(buildPartial());
- }
-
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_Filter_descriptor;
@@ -447,6 +420,32 @@ public final class FilterProtos {
return result;
}
+ public Builder clone() {
+ return (Builder) super.clone();
+ }
+ public Builder setField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ Object value) {
+ return (Builder) super.setField(field, value);
+ }
+ public Builder clearField(
+ com.google.protobuf.Descriptors.FieldDescriptor field) {
+ return (Builder) super.clearField(field);
+ }
+ public Builder clearOneof(
+ com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ return (Builder) super.clearOneof(oneof);
+ }
+ public Builder setRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ int index, Object value) {
+ return (Builder) super.setRepeatedField(field, index, value);
+ }
+ public Builder addRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ Object value) {
+ return (Builder) super.addRepeatedField(field, value);
+ }
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter)other);
@@ -466,13 +465,13 @@ public final class FilterProtos {
if (other.hasSerializedFilter()) {
setSerializedFilter(other.getSerializedFilter());
}
- this.mergeUnknownFields(other.getUnknownFields());
+ this.mergeUnknownFields(other.unknownFields);
+ onChanged();
return this;
}
public final boolean isInitialized() {
if (!hasName()) {
-
return false;
}
return true;
@@ -487,7 +486,7 @@ public final class FilterProtos {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter) e.getUnfinishedMessage();
- throw e;
+ throw e.unwrapIOException();
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
@@ -497,7 +496,6 @@ public final class FilterProtos {
}
private int bitField0_;
- // required string name = 1;
private java.lang.Object name_ = "";
/**
* <code>required string name = 1;</code>
@@ -511,9 +509,12 @@ public final class FilterProtos {
public java.lang.String getName() {
java.lang.Object ref = name_;
if (!(ref instanceof java.lang.String)) {
- java.lang.String s = ((com.google.protobuf.ByteString) ref)
- .toStringUtf8();
- name_ = s;
+ com.google.protobuf.ByteString bs =
+ (com.google.protobuf.ByteString) ref;
+ java.lang.String s = bs.toStringUtf8();
+ if (bs.isValidUtf8()) {
+ name_ = s;
+ }
return s;
} else {
return (java.lang.String) ref;
@@ -571,7 +572,6 @@ public final class FilterProtos {
return this;
}
- // optional bytes serialized_filter = 2;
private com.google.protobuf.ByteString serializedFilter_ = com.google.protobuf.ByteString.EMPTY;
/**
* <code>optional bytes serialized_filter = 2;</code>
@@ -606,22 +606,59 @@ public final class FilterProtos {
onChanged();
return this;
}
+ public final Builder setUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.setUnknownFields(unknownFields);
+ }
+
+ public final Builder mergeUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.mergeUnknownFields(unknownFields);
+ }
+
// @@protoc_insertion_point(builder_scope:hbase.pb.Filter)
}
+ // @@protoc_insertion_point(class_scope:hbase.pb.Filter)
+ private static final org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter DEFAULT_INSTANCE;
static {
- defaultInstance = new Filter(true);
- defaultInstance.initFields();
+ DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter();
+ }
+
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter getDefaultInstance() {
+ return DEFAULT_INSTANCE;
+ }
+
+ @java.lang.Deprecated public static final com.google.protobuf.Parser<Filter>
+ PARSER = new com.google.protobuf.AbstractParser<Filter>() {
+ public Filter parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new Filter(input, extensionRegistry);
+ }
+ };
+
+ public static com.google.protobuf.Parser<Filter> parser() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<Filter> getParserForType() {
+ return PARSER;
+ }
+
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter getDefaultInstanceForType() {
+ return DEFAULT_INSTANCE;
}
- // @@protoc_insertion_point(class_scope:hbase.pb.Filter)
}
- public interface ColumnCountGetFilterOrBuilder
- extends com.google.protobuf.MessageOrBuilder {
+ public interface ColumnCountGetFilterOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:hbase.pb.ColumnCountGetFilter)
+ com.google.protobuf.MessageOrBuilder {
- // required int32 limit = 1;
/**
* <code>required int32 limit = 1;</code>
*/
@@ -634,36 +671,28 @@ public final class FilterProtos {
/**
* Protobuf type {@code hbase.pb.ColumnCountGetFilter}
*/
- public static final class ColumnCountGetFilter extends
- com.google.protobuf.GeneratedMessage
- implements ColumnCountGetFilterOrBuilder {
+ public static final class ColumnCountGetFilter extends
+ com.google.protobuf.GeneratedMessageV3 implements
+ // @@protoc_insertion_point(message_implements:hbase.pb.ColumnCountGetFilter)
+ ColumnCountGetFilterOrBuilder {
// Use ColumnCountGetFilter.newBuilder() to construct.
- private ColumnCountGetFilter(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ private ColumnCountGetFilter(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
- this.unknownFields = builder.getUnknownFields();
- }
- private ColumnCountGetFilter(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
- private static final ColumnCountGetFilter defaultInstance;
- public static ColumnCountGetFilter getDefaultInstance() {
- return defaultInstance;
}
-
- public ColumnCountGetFilter getDefaultInstanceForType() {
- return defaultInstance;
+ private ColumnCountGetFilter() {
+ limit_ = 0;
}
- private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
+ getUnknownFields() {
return this.unknownFields;
}
private ColumnCountGetFilter(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- initFields();
+ this();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
@@ -693,7 +722,7 @@ public final class FilterProtos {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
- e.getMessage()).setUnfinishedMessage(this);
+ e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
@@ -704,30 +733,14 @@ public final class FilterProtos {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnCountGetFilter_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnCountGetFilter_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter.class, org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter.Builder.class);
}
- public static com.google.protobuf.Parser<ColumnCountGetFilter> PARSER =
- new com.google.protobuf.AbstractParser<ColumnCountGetFilter>() {
- public ColumnCountGetFilter parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return new ColumnCountGetFilter(input, extensionRegistry);
- }
- };
-
- @java.lang.Override
- public com.google.protobuf.Parser<ColumnCountGetFilter> getParserForType() {
- return PARSER;
- }
-
private int bitField0_;
- // required int32 limit = 1;
public static final int LIMIT_FIELD_NUMBER = 1;
private int limit_;
/**
@@ -743,13 +756,11 @@ public final class FilterProtos {
return limit_;
}
- private void initFields() {
- limit_ = 0;
- }
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
- if (isInitialized != -1) return isInitialized == 1;
+ if (isInitialized == 1) return true;
+ if (isInitialized == 0) return false;
if (!hasLimit()) {
memoizedIsInitialized = 0;
@@ -761,16 +772,14 @@ public final class FilterProtos {
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
- getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeInt32(1, limit_);
}
- getUnknownFields().writeTo(output);
+ unknownFields.writeTo(output);
}
- private int memoizedSerializedSize = -1;
public int getSerializedSize() {
- int size = memoizedSerializedSize;
+ int size = memoizedSize;
if (size != -1) return size;
size = 0;
@@ -778,19 +787,13 @@ public final class FilterProtos {
size += com.google.protobuf.CodedOutputStream
.computeInt32Size(1, limit_);
}
- size += getUnknownFields().getSerializedSize();
- memoizedSerializedSize = size;
+ size += unknownFields.getSerializedSize();
+ memoizedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
- protected java.lang.Object writeReplace()
- throws java.io.ObjectStreamException {
- return super.writeReplace();
- }
-
- @java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
@@ -806,12 +809,10 @@ public final class FilterProtos {
result = result && (getLimit()
== other.getLimit());
}
- result = result &&
- getUnknownFields().equals(other.getUnknownFields());
+ result = result && unknownFields.equals(other.unknownFields);
return result;
}
- private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
@@ -823,7 +824,7 @@ public final class FilterProtos {
hash = (37 * hash) + LIMIT_FIELD_NUMBER;
hash = (53 * hash) + getLimit();
}
- hash = (29 * hash) + getUnknownFields().hashCode();
+ hash = (29 * hash) + unknownFields.hashCode();
memoizedHashCode = hash;
return hash;
}
@@ -851,46 +852,57 @@ public final class FilterProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
- public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder() {
+ return DEFAULT_INSTANCE.toBuilder();
+ }
public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter prototype) {
- return newBuilder().mergeFrom(prototype);
+ return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+ }
+ public Builder toBuilder() {
+ return this == DEFAULT_INSTANCE
+ ? new Builder() : new Builder().mergeFrom(this);
}
- public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -898,14 +910,15 @@ public final class FilterProtos {
* Protobuf type {@code hbase.pb.ColumnCountGetFilter}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilterOrBuilder {
+ com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ // @@protoc_insertion_point(builder_implements:hbase.pb.ColumnCountGetFilter)
+ org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilterOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnCountGetFilter_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnCountGetFilter_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -918,18 +931,15 @@ public final class FilterProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ if (com.google.protobuf.GeneratedMessageV3
+ .alwaysUseFieldBuilders) {
}
}
- private static Builder create() {
- return new Builder();
- }
-
public Builder clear() {
super.clear();
limit_ = 0;
@@ -937,10 +947,6 @@ public final class FilterProtos {
return this;
}
- public Builder clone() {
- return create().mergeFrom(buildPartial());
- }
-
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnCountGetFilter_descriptor;
@@ -971,6 +977,32 @@ public final class FilterProtos {
return result;
}
+ public Builder clone() {
+ return (Builder) super.clone();
+ }
+ public Builder setField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ Object value) {
+ return (Builder) super.setField(field, value);
+ }
+ public Builder clearField(
+ com.google.protobuf.Descriptors.FieldDescriptor field) {
+ return (Builder) super.clearField(field);
+ }
+ public Builder clearOneof(
+ com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ return (Builder) super.clearOneof(oneof);
+ }
+ public Builder setRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ int index, Object value) {
+ return (Builder) super.setRepeatedField(field, index, value);
+ }
+ public Builder addRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ Object value) {
+ return (Builder) super.addRepeatedField(field, value);
+ }
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter)other);
@@ -985,13 +1017,13 @@ public final class FilterProtos {
if (other.hasLimit()) {
setLimit(other.getLimit());
}
- this.mergeUnknownFields(other.getUnknownFields());
+ this.mergeUnknownFields(other.unknownFields);
+ onChanged();
return this;
}
public final boolean isInitialized() {
if (!hasLimit()) {
-
return false;
}
return true;
@@ -1006,7 +1038,7 @@ public final class FilterProtos {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter) e.getUnfinishedMessage();
- throw e;
+ throw e.unwrapIOException();
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
@@ -1016,7 +1048,6 @@ public final class FilterProtos {
}
private int bitField0_;
- // required int32 limit = 1;
private int limit_ ;
/**
* <code>required int32 limit = 1;</code>
@@ -1048,22 +1079,59 @@ public final class FilterProtos {
onChanged();
return this;
}
+ public final Builder setUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.setUnknownFields(unknownFields);
+ }
+
+ public final Builder mergeUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.mergeUnknownFields(unknownFields);
+ }
+
// @@protoc_insertion_point(builder_scope:hbase.pb.ColumnCountGetFilter)
}
+ // @@protoc_insertion_point(class_scope:hbase.pb.ColumnCountGetFilter)
+ private static final org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter DEFAULT_INSTANCE;
static {
- defaultInstance = new ColumnCountGetFilter(true);
- defaultInstance.initFields();
+ DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter();
+ }
+
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter getDefaultInstance() {
+ return DEFAULT_INSTANCE;
+ }
+
+ @java.lang.Deprecated public static final com.google.protobuf.Parser<ColumnCountGetFilter>
+ PARSER = new com.google.protobuf.AbstractParser<ColumnCountGetFilter>() {
+ public ColumnCountGetFilter parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new ColumnCountGetFilter(input, extensionRegistry);
+ }
+ };
+
+ public static com.google.protobuf.Parser<ColumnCountGetFilter> parser() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<ColumnCountGetFilter> getParserForType() {
+ return PARSER;
+ }
+
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter getDefaultInstanceForType() {
+ return DEFAULT_INSTANCE;
}
- // @@protoc_insertion_point(class_scope:hbase.pb.ColumnCountGetFilter)
}
- public interface ColumnPaginationFilterOrBuilder
- extends com.google.protobuf.MessageOrBuilder {
+ public interface ColumnPaginationFilterOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:hbase.pb.ColumnPaginationFilter)
+ com.google.protobuf.MessageOrBuilder {
- // required int32 limit = 1;
/**
* <code>required int32 limit = 1;</code>
*/
@@ -1073,7 +1141,6 @@ public final class FilterProtos {
*/
int getLimit();
- // optional int32 offset = 2;
/**
* <code>optional int32 offset = 2;</code>
*/
@@ -1083,7 +1150,6 @@ public final class FilterProtos {
*/
int getOffset();
- // optional bytes column_offset = 3;
/**
* <code>optional bytes column_offset = 3;</code>
*/
@@ -1096,36 +1162,30 @@ public final class FilterProtos {
/**
* Protobuf type {@code hbase.pb.ColumnPaginationFilter}
*/
- public static final class ColumnPaginationFilter extends
- com.google.protobuf.GeneratedMessage
- implements ColumnPaginationFilterOrBuilder {
+ public static final class ColumnPaginationFilter extends
+ com.google.protobuf.GeneratedMessageV3 implements
+ // @@protoc_insertion_point(message_implements:hbase.pb.ColumnPaginationFilter)
+ ColumnPaginationFilterOrBuilder {
// Use ColumnPaginationFilter.newBuilder() to construct.
- private ColumnPaginationFilter(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ private ColumnPaginationFilter(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
- this.unknownFields = builder.getUnknownFields();
}
- private ColumnPaginationFilter(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
- private static final ColumnPaginationFilter defaultInstance;
- public static ColumnPaginationFilter getDefaultInstance() {
- return defaultInstance;
- }
-
- public ColumnPaginationFilter getDefaultInstanceForType() {
- return defaultInstance;
+ private ColumnPaginationFilter() {
+ limit_ = 0;
+ offset_ = 0;
+ columnOffset_ = com.google.protobuf.ByteString.EMPTY;
}
- private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
+ getUnknownFields() {
return this.unknownFields;
}
private ColumnPaginationFilter(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- initFields();
+ this();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
@@ -1165,7 +1225,7 @@ public final class FilterProtos {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
- e.getMessage()).setUnfinishedMessage(this);
+ e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
@@ -1176,30 +1236,14 @@ public final class FilterProtos {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPaginationFilter_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPaginationFilter_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter.class, org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter.Builder.class);
}
- public static com.google.protobuf.Parser<ColumnPaginationFilter> PARSER =
- new com.google.protobuf.AbstractParser<ColumnPaginationFilter>() {
- public ColumnPaginationFilter parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return new ColumnPaginationFilter(input, extensionRegistry);
- }
- };
-
- @java.lang.Override
- public com.google.protobuf.Parser<ColumnPaginationFilter> getParserForType() {
- return PARSER;
- }
-
private int bitField0_;
- // required int32 limit = 1;
public static final int LIMIT_FIELD_NUMBER = 1;
private int limit_;
/**
@@ -1215,7 +1259,6 @@ public final class FilterProtos {
return limit_;
}
- // optional int32 offset = 2;
public static final int OFFSET_FIELD_NUMBER = 2;
private int offset_;
/**
@@ -1231,7 +1274,6 @@ public final class FilterProtos {
return offset_;
}
- // optional bytes column_offset = 3;
public static final int COLUMN_OFFSET_FIELD_NUMBER = 3;
private com.google.protobuf.ByteString columnOffset_;
/**
@@ -1247,15 +1289,11 @@ public final class FilterProtos {
return columnOffset_;
}
- private void initFields() {
- limit_ = 0;
- offset_ = 0;
- columnOffset_ = com.google.protobuf.ByteString.EMPTY;
- }
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
- if (isInitialized != -1) return isInitialized == 1;
+ if (isInitialized == 1) return true;
+ if (isInitialized == 0) return false;
if (!hasLimit()) {
memoizedIsInitialized = 0;
@@ -1267,7 +1305,6 @@ public final class FilterProtos {
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
- getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeInt32(1, limit_);
}
@@ -1277,12 +1314,11 @@ public final class FilterProtos {
if (((bitField0_ & 0x00000004) == 0x00000004)) {
output.writeBytes(3, columnOffset_);
}
- getUnknownFields().writeTo(output);
+ unknownFields.writeTo(output);
}
- private int memoizedSerializedSize = -1;
public int getSerializedSize() {
- int size = memoizedSerializedSize;
+ int size = memoizedSize;
if (size != -1) return size;
size = 0;
@@ -1298,19 +1334,13 @@ public final class FilterProtos {
size += com.google.protobuf.CodedOutputStream
.computeBytesSize(3, columnOffset_);
}
- size += getUnknownFields().getSerializedSize();
- memoizedSerializedSize = size;
+ size += unknownFields.getSerializedSize();
+ memoizedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
- protected java.lang.Object writeReplace()
- throws java.io.ObjectStreamException {
- return super.writeReplace();
- }
-
- @java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
@@ -1336,12 +1366,10 @@ public final class FilterProtos {
result = result && getColumnOffset()
.equals(other.getColumnOffset());
}
- result = result &&
- getUnknownFields().equals(other.getUnknownFields());
+ result = result && unknownFields.equals(other.unknownFields);
return result;
}
- private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
@@ -1361,7 +1389,7 @@ public final class FilterProtos {
hash = (37 * hash) + COLUMN_OFFSET_FIELD_NUMBER;
hash = (53 * hash) + getColumnOffset().hashCode();
}
- hash = (29 * hash) + getUnknownFields().hashCode();
+ hash = (29 * hash) + unknownFields.hashCode();
memoizedHashCode = hash;
return hash;
}
@@ -1389,46 +1417,57 @@ public final class FilterProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
- public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder() {
+ return DEFAULT_INSTANCE.toBuilder();
+ }
public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter prototype) {
- return newBuilder().mergeFrom(prototype);
+ return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+ }
+ public Builder toBuilder() {
+ return this == DEFAULT_INSTANCE
+ ? new Builder() : new Builder().mergeFrom(this);
}
- public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1436,14 +1475,15 @@ public final class FilterProtos {
* Protobuf type {@code hbase.pb.ColumnPaginationFilter}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilterOrBuilder {
+ com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ // @@protoc_insertion_point(builder_implements:hbase.pb.ColumnPaginationFilter)
+ org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilterOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPaginationFilter_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPaginationFilter_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1456,18 +1496,15 @@ public final class FilterProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ if (com.google.protobuf.GeneratedMessageV3
+ .alwaysUseFieldBuilders) {
}
}
- private static Builder create() {
- return new Builder();
- }
-
public Builder clear() {
super.clear();
limit_ = 0;
@@ -1479,10 +1516,6 @@ public final class FilterProtos {
return this;
}
- public Builder clone() {
- return create().mergeFrom(buildPartial());
- }
-
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPaginationFilter_descriptor;
@@ -1521,6 +1554,32 @@ public final class FilterProtos {
return result;
}
+ public Builder clone() {
+ return (Builder) super.clone();
+ }
+ public Builder setField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ Object value) {
+ return (Builder) super.setField(field, value);
+ }
+ public Builder clearField(
+ com.google.protobuf.Descriptors.FieldDescriptor field) {
+ return (Builder) super.clearField(field);
+ }
+ public Builder clearOneof(
+ com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ return (Builder) super.clearOneof(oneof);
+ }
+ public Builder setRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ int index, Object value) {
+ return (Builder) super.setRepeatedField(field, index, value);
+ }
+ public Builder addRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ Object value) {
+ return (Builder) super.addRepeatedField(field, value);
+ }
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter)other);
@@ -1541,13 +1600,13 @@ public final class FilterProtos {
if (other.hasColumnOffset()) {
setColumnOffset(other.getColumnOffset());
}
- this.mergeUnknownFields(other.getUnknownFields());
+ this.mergeUnknownFields(other.unknownFields);
+ onChanged();
return this;
}
public final boolean isInitialized() {
if (!hasLimit()) {
-
return false;
}
return true;
@@ -1562,7 +1621,7 @@ public final class FilterProtos {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter) e.getUnfinishedMessage();
- throw e;
+ throw e.unwrapIOException();
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
@@ -1572,7 +1631,6 @@ public final class FilterProtos {
}
private int bitField0_;
- // required int32 limit = 1;
private int limit_ ;
/**
* <code>required int32 limit = 1;</code>
@@ -1605,7 +1663,6 @@ public final class FilterProtos {
return this;
}
- // optional int32 offset = 2;
private int offset_ ;
/**
* <code>optional int32 offset = 2;</code>
@@ -1638,7 +1695,6 @@ public final class FilterProtos {
return this;
}
- // optional bytes column_offset = 3;
private com.google.protobuf.ByteString columnOffset_ = com.google.protobuf.ByteString.EMPTY;
/**
* <code>optional bytes column_offset = 3;</code>
@@ -1673,22 +1729,59 @@ public final class FilterProtos {
onChanged();
return this;
}
+ public final Builder setUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.setUnknownFields(unknownFields);
+ }
+
+ public final Builder mergeUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.mergeUnknownFields(unknownFields);
+ }
+
// @@protoc_insertion_point(builder_scope:hbase.pb.ColumnPaginationFilter)
}
+ // @@protoc_insertion_point(class_scope:hbase.pb.ColumnPaginationFilter)
+ private static final org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter DEFAULT_INSTANCE;
static {
- defaultInstance = new ColumnPaginationFilter(true);
- defaultInstance.initFields();
+ DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter();
+ }
+
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter getDefaultInstance() {
+ return DEFAULT_INSTANCE;
+ }
+
+ @java.lang.Deprecated public static final com.google.protobuf.Parser<ColumnPaginationFilter>
+ PARSER = new com.google.protobuf.AbstractParser<ColumnPaginationFilter>() {
+ public ColumnPaginationFilter parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new ColumnPaginationFilter(input, extensionRegistry);
+ }
+ };
+
+ public static com.google.protobuf.Parser<ColumnPaginationFilter> parser() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<ColumnPaginationFilter> getParserForType() {
+ return PARSER;
+ }
+
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter getDefaultInstanceForType() {
+ return DEFAULT_INSTANCE;
}
- // @@protoc_insertion_point(class_scope:hbase.pb.ColumnPaginationFilter)
}
- public interface ColumnPrefixFilterOrBuilder
- extends com.google.protobuf.MessageOrBuilder {
+ public interface ColumnPrefixFilterOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:hbase.pb.ColumnPrefixFilter)
+ com.google.protobuf.MessageOrBuilder {
- // required bytes prefix = 1;
/**
* <code>required bytes prefix = 1;</code>
*/
@@ -1701,36 +1794,28 @@ public final class FilterProtos {
/**
* Protobuf type {@code hbase.pb.ColumnPrefixFilter}
*/
- public static final class ColumnPrefixFilter extends
- com.google.protobuf.GeneratedMessage
- implements ColumnPrefixFilterOrBuilder {
+ public static final class ColumnPrefixFilter extends
+ com.google.protobuf.GeneratedMessageV3 implements
+ // @@protoc_insertion_point(message_implements:hbase.pb.ColumnPrefixFilter)
+ ColumnPrefixFilterOrBuilder {
// Use ColumnPrefixFilter.newBuilder() to construct.
- private ColumnPrefixFilter(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ private ColumnPrefixFilter(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
- this.unknownFields = builder.getUnknownFields();
- }
- private ColumnPrefixFilter(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
- private static final ColumnPrefixFilter defaultInstance;
- public static ColumnPrefixFilter getDefaultInstance() {
- return defaultInstance;
}
-
- public ColumnPrefixFilter getDefaultInstanceForType() {
- return defaultInstance;
+ private ColumnPrefixFilter() {
+ prefix_ = com.google.protobuf.ByteString.EMPTY;
}
- private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
+ getUnknownFields() {
return this.unknownFields;
}
private ColumnPrefixFilter(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- initFields();
+ this();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
@@ -1760,7 +1845,7 @@ public final class FilterProtos {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
- e.getMessage()).setUnfinishedMessage(this);
+ e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
@@ -1771,30 +1856,14 @@ public final class FilterProtos {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPrefixFilter_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPrefixFilter_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter.class, org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter.Builder.class);
}
- public static com.google.protobuf.Parser<ColumnPrefixFilter> PARSER =
- new com.google.protobuf.AbstractParser<ColumnPrefixFilter>() {
- public ColumnPrefixFilter parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return new ColumnPrefixFilter(input, extensionRegistry);
- }
- };
-
- @java.lang.Override
- public com.google.protobuf.Parser<ColumnPrefixFilter> getParserForType() {
- return PARSER;
- }
-
private int bitField0_;
- // required bytes prefix = 1;
public static final int PREFIX_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString prefix_;
/**
@@ -1810,13 +1879,11 @@ public final class FilterProtos {
return prefix_;
}
- private void initFields() {
- prefix_ = com.google.protobuf.ByteString.EMPTY;
- }
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
- if (isInitialized != -1) return isInitialized == 1;
+ if (isInitialized == 1) return true;
+ if (isInitialized == 0) return false;
if (!hasPrefix()) {
memoizedIsInitialized = 0;
@@ -1828,16 +1895,14 @@ public final class FilterProtos {
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
- getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, prefix_);
}
- getUnknownFields().writeTo(output);
+ unknownFields.writeTo(output);
}
- private int memoizedSerializedSize = -1;
public int getSerializedSize() {
- int size = memoizedSerializedSize;
+ int size = memoizedSize;
if (size != -1) return size;
size = 0;
@@ -1845,19 +1910,13 @@ public final class FilterProtos {
size += com.google.protobuf.CodedOutputStream
.computeBytesSize(1, prefix_);
}
- size += getUnknownFields().getSerializedSize();
- memoizedSerializedSize = size;
+ size += unknownFields.getSerializedSize();
+ memoizedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
- protected java.lang.Object writeReplace()
- throws java.io.ObjectStreamException {
- return super.writeReplace();
- }
-
- @java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
@@ -1873,12 +1932,10 @@ public final class FilterProtos {
result = result && getPrefix()
.equals(other.getPrefix());
}
- result = result &&
- getUnknownFields().equals(other.getUnknownFields());
+ result = result && unknownFields.equals(other.unknownFields);
return result;
}
- private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
@@ -1890,7 +1947,7 @@ public final class FilterProtos {
hash = (37 * hash) + PREFIX_FIELD_NUMBER;
hash = (53 * hash) + getPrefix().hashCode();
}
- hash = (29 * hash) + getUnknownFields().hashCode();
+ hash = (29 * hash) + unknownFields.hashCode();
memoizedHashCode = hash;
return hash;
}
@@ -1918,46 +1975,57 @@ public final class FilterProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
- public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder() {
+ return DEFAULT_INSTANCE.toBuilder();
+ }
public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter prototype) {
- return newBuilder().mergeFrom(prototype);
+ return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+ }
+ public Builder toBuilder() {
+ return this == DEFAULT_INSTANCE
+ ? new Builder() : new Builder().mergeFrom(this);
}
- public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1965,14 +2033,15 @@ public final class FilterProtos {
* Protobuf type {@code hbase.pb.ColumnPrefixFilter}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilterOrBuilder {
+ com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ // @@protoc_insertion_point(builder_implements:hbase.pb.ColumnPrefixFilter)
+ org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilterOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPrefixFilter_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPrefixFilter_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1985,18 +2054,15 @@ public final class FilterProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ if (com.google.protobuf.GeneratedMessageV3
+ .alwaysUseFieldBuilders) {
}
}
- private static Builder create() {
- return new Builder();
- }
-
public Builder clear() {
super.clear();
prefix_ = com.google.protobuf.ByteString.EMPTY;
@@ -2004,10 +2070,6 @@ public final class FilterProtos {
return this;
}
- public Builder clone() {
- return create().mergeFrom(buildPartial());
- }
-
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPrefixFilter_descriptor;
@@ -2038,6 +2100,32 @@ public final class FilterProtos {
return result;
}
+ public Builder clone() {
+ return (Builder) super.clone();
+ }
+ public Builder setField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ Object value) {
+ return (Builder) super.setField(field, value);
+ }
+ public Builder clearField(
+ com.google.protobuf.Descriptors.FieldDescriptor field) {
+ return (Builder) super.clearField(field);
+ }
+ public Builder clearOneof(
+ com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ return (Builder) super.clearOneof(oneof);
+ }
+ public Builder setRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ int index, Object value) {
+ return (Builder) super.setRepeatedField(field, index, value);
+ }
+ public Builder addRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ Object value) {
+ return (Builder) super.addRepeatedField(field, value);
+ }
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter)other);
@@ -2052,13 +2140,13 @@ public final class FilterProtos {
if (other.hasPrefix()) {
setPrefix(other.getPrefix());
}
- this.mergeUnknownFields(other.getUnknownFields());
+ this.mergeUnknownFields(other.unknownFields);
+ onChanged();
return this;
}
public final boolean isInitialized() {
if (!hasPrefix()) {
-
return false;
}
return true;
@@ -2073,7 +2161,7 @@ public final class FilterProtos {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter) e.getUnfinishedMessage();
- throw e;
+ throw e.unwrapIOException();
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
@@ -2083,7 +2171,6 @@ public final class FilterProtos {
}
private int bitField0_;
- // required bytes prefix = 1;
private com.google.protobuf.ByteString prefix_ = com.google.protobuf.ByteString.EMPTY;
/**
* <code>required bytes prefix = 1;</code>
@@ -2118,32 +2205,68 @@ public final class FilterProtos {
onChanged();
return this;
}
+ public final Builder setUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.setUnknownFields(unknownFields);
+ }
+
+ public final Builder mergeUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.mergeUnknownFields(unknownFields);
+ }
+
// @@protoc_insertion_point(builder_scope:hbase.pb.ColumnPrefixFilter)
}
+ // @@protoc_insertion_point(class_scope:hbase.pb.ColumnPrefixFilter)
+ private static final org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter DEFAULT_INSTANCE;
static {
- defaultInstance = new ColumnPrefixFilter(true);
- defaultInstance.initFields();
+ DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter();
}
- // @@protoc_insertion_point(class_scope:hbase.pb.ColumnPrefixFilter)
- }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter getDefaultInstance() {
+ return DEFAULT_INSTANCE;
+ }
- public interface ColumnRangeFilterOrBuilder
- extends com.google.protobuf.MessageOrBuilder {
+ @java.lang.Deprecated public static final com.google.protobuf.Parser<ColumnPrefixFilter>
+ PARSER = new com.google.protobuf.AbstractParser<ColumnPrefixFilter>() {
+ public ColumnPrefixFilter parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new ColumnPrefixFilter(input, extensionRegistry);
+ }
+ };
- // optional bytes min_column = 1;
- /**
- * <code>optional bytes min_column = 1;</code>
- */
- boolean hasMinColumn();
- /**
- * <code>optional bytes min_column = 1;</code>
- */
+ public static com.google.protobuf.Parser<ColumnPrefixFilter> parser() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<ColumnPrefixFilter> getParserForType() {
+ return PARSER;
+ }
+
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter getDefaultInstanceForType() {
+ return DEFAULT_INSTANCE;
+ }
+
+ }
+
+ public interface ColumnRangeFilterOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:hbase.pb.ColumnRangeFilter)
+ com.google.protobuf.MessageOrBuilder {
+
+ /**
+ * <code>optional bytes min_column = 1;</code>
+ */
+ boolean hasMinColumn();
+ /**
+ * <code>optional bytes min_column = 1;</code>
+ */
com.google.protobuf.ByteString getMinColumn();
- // optional bool min_column_inclusive = 2;
/**
* <code>optional bool min_column_inclusive = 2;</code>
*/
@@ -2153,7 +2276,6 @@ public final class FilterProtos {
*/
boolean getMinColumnInclusive();
- // optional bytes max_column = 3;
/**
* <code>optional bytes max_column = 3;</code>
*/
@@ -2163,7 +2285,6 @@ public final class FilterProtos {
*/
com.google.protobuf.ByteString getMaxColumn();
- // optional bool max_column_inclusive = 4;
/**
* <code>optional bool max_column_inclusive = 4;</code>
*/
@@ -2176,36 +2297,31 @@ public final class FilterProtos {
/**
* Protobuf type {@code hbase.pb.ColumnRangeFilter}
*/
- public static final class ColumnRangeFilter extends
- com.google.protobuf.GeneratedMessage
- implements ColumnRangeFilterOrBuilder {
+ public static final class ColumnRangeFilter extends
+ com.google.protobuf.GeneratedMessageV3 implements
+ // @@protoc_insertion_point(message_implements:hbase.pb.ColumnRangeFilter)
+ ColumnRangeFilterOrBuilder {
// Use ColumnRangeFilter.newBuilder() to construct.
- private ColumnRangeFilter(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ private ColumnRangeFilter(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
- this.unknownFields = builder.getUnknownFields();
}
- private ColumnRangeFilter(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
- private static final ColumnRangeFilter defaultInstance;
- public static ColumnRangeFilter getDefaultInstance() {
- return defaultInstance;
- }
-
- public ColumnRangeFilter getDefaultInstanceForType() {
- return defaultInstance;
+ private ColumnRangeFilter() {
+ minColumn_ = com.google.protobuf.ByteString.EMPTY;
+ minColumnInclusive_ = false;
+ maxColumn_ = com.google.protobuf.ByteString.EMPTY;
+ maxColumnInclusive_ = false;
}
- private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
+ getUnknownFields() {
return this.unknownFields;
}
private ColumnRangeFilter(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- initFields();
+ this();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
@@ -2250,7 +2366,7 @@ public final class FilterProtos {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
- e.getMessage()).setUnfinishedMessage(this);
+ e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
@@ -2261,30 +2377,14 @@ public final class FilterProtos {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnRangeFilter_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnRangeFilter_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnRangeFilter.class, org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnRangeFilter.Builder.class);
}
- public static com.google.protobuf.Parser<ColumnRangeFilter> PARSER =
- new com.google.protobuf.AbstractParser<ColumnRangeFilter>() {
- public ColumnRangeFilter parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return new ColumnRangeFilter(input, extensionRegistry);
- }
- };
-
- @java.lang.Override
- public com.google.protobuf.Parser<ColumnRangeFilter> getParserForType() {
- return PARSER;
- }
-
private int bitField0_;
- // optional bytes min_column = 1;
public static final int MIN_COLUMN_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString minColumn_;
/**
@@ -2300,7 +2400,6 @@ public final class FilterProtos {
return minColumn_;
}
- // optional bool min_column_inclusive = 2;
public static final int MIN_COLUMN_INCLUSIVE_FIELD_NUMBER = 2;
private boolean minColumnInclusive_;
/**
@@ -2316,7 +2415,6 @@ public final class FilterProtos {
return minColumnInclusive_;
}
- // optional bytes max_column = 3;
public static final int MAX_COLUMN_FIELD_NUMBER = 3;
private com.google.protobuf.ByteString maxColumn_;
/**
@@ -2332,7 +2430,6 @@ public final class FilterProtos {
return maxColumn_;
}
- // optional bool max_column_inclusive = 4;
public static final int MAX_COLUMN_INCLUSIVE_FIELD_NUMBER = 4;
private boolean maxColumnInclusive_;
/**
@@ -2348,16 +2445,11 @@ public final class FilterProtos {
return maxColumnInclusive_;
}
- private void initFields() {
- minColumn_ = com.google.protobuf.ByteString.EMPTY;
- minColumnInclusive_ = false;
- maxColumn_ = com.google.protobuf.ByteString.EMPTY;
- maxColumnInclusive_ = false;
- }
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
- if (isInitialized != -1) return isInitialized == 1;
+ if (isInitialized == 1) return true;
+ if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
@@ -2365,7 +2457,6 @@ public final class FilterProtos {
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
- getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, minColumn_);
}
@@ -2378,12 +2469,11 @@ public final class FilterProtos {
if (((bitField0_ & 0x00000008) == 0x00000008)) {
output.writeBool(4, maxColumnInclusive_);
}
- getUnknownFields().writeTo(output);
+ unknownFields.writeTo(output);
}
- private int memoizedSerializedSize = -1;
public int getSerializedSize() {
- int size = memoizedSerializedSize;
+ int size = memoizedSize;
if (size != -1) return size;
size = 0;
@@ -2403,19 +2493,13 @@ public final class FilterProtos {
size += com.google.protobuf.CodedOutputStream
.computeBoolSize(4, maxColumnInclusive_);
}
- size += getUnknownFields().getSerializedSize();
- memoizedSerializedSize = size;
+ size += unknownFields.getSerializedSize();
+ memoizedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
- protected java.lang.Object writeReplace()
- throws java.io.ObjectStreamException {
- return super.writeReplace();
- }
-
- @java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
@@ -2446,12 +2530,10 @@ public final class FilterProtos {
result = result && (getMaxColumnInclusive()
== other.getMaxColumnInclusive());
}
- result = result &&
- getUnknownFields().equals(other.getUnknownFields());
+ result = result && unknownFields.equals(other.unknownFields);
return result;
}
- private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
@@ -2465,7 +2547,8 @@ public final class FilterProtos {
}
if (hasMinColumnInclusive()) {
hash = (37 * hash) + MIN_COLUMN_INCLUSIVE_FIELD_NUMBER;
- hash = (53 * hash) + hashBoolean(getMinColumnInclusive());
+ hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean(
+ getMinColumnInclusive());
}
if (hasMaxColumn()) {
hash = (37 * hash) + MAX_COLUMN_FIELD_NUMBER;
@@ -2473,9 +2556,10 @@ public final class FilterProtos {
}
if (hasMaxColumnInclusive()) {
hash = (37 * hash) + MAX_COLUMN_INCLUSIVE_FIELD_NUMBER;
- hash = (53 * hash) + hashBoolean(getMaxColumnInclusive());
+ hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean(
+ getMaxColumnInclusive());
}
- hash = (29 * hash) + getUnknownFields().hashCode();
+ hash = (29 * hash) + unknownFields.hashCode();
memoizedHashCode = hash;
return hash;
}
@@ -2503,46 +2587,57 @@ public final class FilterProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnRangeFilter parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnRangeFilter parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnRangeFilter parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnRangeFilter parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnRangeFilter parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnRangeFilter parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
- public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder() {
+ return DEFAULT_INSTANCE.toBuilder();
+ }
public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnRangeFilter prototype) {
- return newBuilder().mergeFrom(prototype);
+ return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+ }
+ public Builder toBuilder() {
+ return this == DEFAULT_INSTANCE
+ ? new Builder() : new Builder().mergeFrom(this);
}
- public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -2550,14 +2645,15 @@ public final class FilterProtos {
* Protobuf type {@code hbase.pb.ColumnRangeFilter}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnRangeFilterOrBuilder {
+ com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ // @@protoc_insertion_point(builder_implements:hbase.pb.ColumnRangeFilter)
+ org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnRangeFilterOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterP
<TRUNCATED>