You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2016/09/30 00:30:44 UTC
[15/29] hbase git commit: HBASE-16567 Upgrade to protobuf-3.1.x
Regenerate all protos in this module with protoc3. Redo ByteStringer to use
new pb3.1.0 unsafebytesutil instead of HBaseZeroCopyByteString
http://git-wip-us.apache.org/repos/asf/hbase/blob/b4a729ed/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/HBaseProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/HBaseProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/HBaseProtos.java
index 1259d3c..2ae9f9c 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/HBaseProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/HBaseProtos.java
@@ -6,45 +6,51 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class HBaseProtos {
private HBaseProtos() {}
public static void registerAllExtensions(
+ com.google.protobuf.ExtensionRegistryLite registry) {
+ }
+
+ public static void registerAllExtensions(
com.google.protobuf.ExtensionRegistry registry) {
+ registerAllExtensions(
+ (com.google.protobuf.ExtensionRegistryLite) registry);
}
/**
- * Protobuf enum {@code hbase.pb.CompareType}
- *
* <pre>
* Comparison operators
* </pre>
+ *
+ * Protobuf enum {@code hbase.pb.CompareType}
*/
public enum CompareType
implements com.google.protobuf.ProtocolMessageEnum {
/**
* <code>LESS = 0;</code>
*/
- LESS(0, 0),
+ LESS(0),
/**
* <code>LESS_OR_EQUAL = 1;</code>
*/
- LESS_OR_EQUAL(1, 1),
+ LESS_OR_EQUAL(1),
/**
* <code>EQUAL = 2;</code>
*/
- EQUAL(2, 2),
+ EQUAL(2),
/**
* <code>NOT_EQUAL = 3;</code>
*/
- NOT_EQUAL(3, 3),
+ NOT_EQUAL(3),
/**
* <code>GREATER_OR_EQUAL = 4;</code>
*/
- GREATER_OR_EQUAL(4, 4),
+ GREATER_OR_EQUAL(4),
/**
* <code>GREATER = 5;</code>
*/
- GREATER(5, 5),
+ GREATER(5),
/**
* <code>NO_OP = 6;</code>
*/
- NO_OP(6, 6),
+ NO_OP(6),
;
/**
@@ -77,9 +83,19 @@ public final class HBaseProtos {
public static final int NO_OP_VALUE = 6;
- public final int getNumber() { return value; }
+ public final int getNumber() {
+ return value;
+ }
+ /**
+ * @deprecated Use {@link #forNumber(int)} instead.
+ */
+ @java.lang.Deprecated
public static CompareType valueOf(int value) {
+ return forNumber(value);
+ }
+
+ public static CompareType forNumber(int value) {
switch (value) {
case 0: return LESS;
case 1: return LESS_OR_EQUAL;
@@ -96,17 +112,17 @@ public final class HBaseProtos {
internalGetValueMap() {
return internalValueMap;
}
- private static com.google.protobuf.Internal.EnumLiteMap<CompareType>
- internalValueMap =
+ private static final com.google.protobuf.Internal.EnumLiteMap<
+ CompareType> internalValueMap =
new com.google.protobuf.Internal.EnumLiteMap<CompareType>() {
public CompareType findValueByNumber(int number) {
- return CompareType.valueOf(number);
+ return CompareType.forNumber(number);
}
};
public final com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
- return getDescriptor().getValues().get(index);
+ return getDescriptor().getValues().get(ordinal());
}
public final com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
@@ -128,11 +144,9 @@ public final class HBaseProtos {
return VALUES[desc.getIndex()];
}
- private final int index;
private final int value;
- private CompareType(int index, int value) {
- this.index = index;
+ private CompareType(int value) {
this.value = value;
}
@@ -147,31 +161,31 @@ public final class HBaseProtos {
/**
* <code>NANOSECONDS = 1;</code>
*/
- NANOSECONDS(0, 1),
+ NANOSECONDS(1),
/**
* <code>MICROSECONDS = 2;</code>
*/
- MICROSECONDS(1, 2),
+ MICROSECONDS(2),
/**
* <code>MILLISECONDS = 3;</code>
*/
- MILLISECONDS(2, 3),
+ MILLISECONDS(3),
/**
* <code>SECONDS = 4;</code>
*/
- SECONDS(3, 4),
+ SECONDS(4),
/**
* <code>MINUTES = 5;</code>
*/
- MINUTES(4, 5),
+ MINUTES(5),
/**
* <code>HOURS = 6;</code>
*/
- HOURS(5, 6),
+ HOURS(6),
/**
* <code>DAYS = 7;</code>
*/
- DAYS(6, 7),
+ DAYS(7),
;
/**
@@ -204,9 +218,19 @@ public final class HBaseProtos {
public static final int DAYS_VALUE = 7;
- public final int getNumber() { return value; }
+ public final int getNumber() {
+ return value;
+ }
+ /**
+ * @deprecated Use {@link #forNumber(int)} instead.
+ */
+ @java.lang.Deprecated
public static TimeUnit valueOf(int value) {
+ return forNumber(value);
+ }
+
+ public static TimeUnit forNumber(int value) {
switch (value) {
case 1: return NANOSECONDS;
case 2: return MICROSECONDS;
@@ -223,17 +247,17 @@ public final class HBaseProtos {
internalGetValueMap() {
return internalValueMap;
}
- private static com.google.protobuf.Internal.EnumLiteMap<TimeUnit>
- internalValueMap =
+ private static final com.google.protobuf.Internal.EnumLiteMap<
+ TimeUnit> internalValueMap =
new com.google.protobuf.Internal.EnumLiteMap<TimeUnit>() {
public TimeUnit findValueByNumber(int number) {
- return TimeUnit.valueOf(number);
+ return TimeUnit.forNumber(number);
}
};
public final com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
- return getDescriptor().getValues().get(index);
+ return getDescriptor().getValues().get(ordinal());
}
public final com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
@@ -255,21 +279,19 @@ public final class HBaseProtos {
return VALUES[desc.getIndex()];
}
- private final int index;
private final int value;
- private TimeUnit(int index, int value) {
- this.index = index;
+ private TimeUnit(int value) {
this.value = value;
}
// @@protoc_insertion_point(enum_scope:hbase.pb.TimeUnit)
}
- public interface TableNameOrBuilder
- extends com.google.protobuf.MessageOrBuilder {
+ public interface TableNameOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:hbase.pb.TableName)
+ com.google.protobuf.MessageOrBuilder {
- // required bytes namespace = 1;
/**
* <code>required bytes namespace = 1;</code>
*/
@@ -279,7 +301,6 @@ public final class HBaseProtos {
*/
com.google.protobuf.ByteString getNamespace();
- // required bytes qualifier = 2;
/**
* <code>required bytes qualifier = 2;</code>
*/
@@ -290,43 +311,36 @@ public final class HBaseProtos {
com.google.protobuf.ByteString getQualifier();
}
/**
- * Protobuf type {@code hbase.pb.TableName}
- *
* <pre>
**
* Table Name
* </pre>
+ *
+ * Protobuf type {@code hbase.pb.TableName}
*/
- public static final class TableName extends
- com.google.protobuf.GeneratedMessage
- implements TableNameOrBuilder {
+ public static final class TableName extends
+ com.google.protobuf.GeneratedMessageV3 implements
+ // @@protoc_insertion_point(message_implements:hbase.pb.TableName)
+ TableNameOrBuilder {
// Use TableName.newBuilder() to construct.
- private TableName(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ private TableName(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
- this.unknownFields = builder.getUnknownFields();
- }
- private TableName(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
- private static final TableName defaultInstance;
- public static TableName getDefaultInstance() {
- return defaultInstance;
}
-
- public TableName getDefaultInstanceForType() {
- return defaultInstance;
+ private TableName() {
+ namespace_ = com.google.protobuf.ByteString.EMPTY;
+ qualifier_ = com.google.protobuf.ByteString.EMPTY;
}
- private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
+ getUnknownFields() {
return this.unknownFields;
}
private TableName(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- initFields();
+ this();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
@@ -361,7 +375,7 @@ public final class HBaseProtos {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
- e.getMessage()).setUnfinishedMessage(this);
+ e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
@@ -372,30 +386,14 @@ public final class HBaseProtos {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableName_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableName_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.class, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder.class);
}
- public static com.google.protobuf.Parser<TableName> PARSER =
- new com.google.protobuf.AbstractParser<TableName>() {
- public TableName parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return new TableName(input, extensionRegistry);
- }
- };
-
- @java.lang.Override
- public com.google.protobuf.Parser<TableName> getParserForType() {
- return PARSER;
- }
-
private int bitField0_;
- // required bytes namespace = 1;
public static final int NAMESPACE_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString namespace_;
/**
@@ -411,7 +409,6 @@ public final class HBaseProtos {
return namespace_;
}
- // required bytes qualifier = 2;
public static final int QUALIFIER_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString qualifier_;
/**
@@ -427,14 +424,11 @@ public final class HBaseProtos {
return qualifier_;
}
- private void initFields() {
- namespace_ = com.google.protobuf.ByteString.EMPTY;
- qualifier_ = com.google.protobuf.ByteString.EMPTY;
- }
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
- if (isInitialized != -1) return isInitialized == 1;
+ if (isInitialized == 1) return true;
+ if (isInitialized == 0) return false;
if (!hasNamespace()) {
memoizedIsInitialized = 0;
@@ -450,19 +444,17 @@ public final class HBaseProtos {
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
- getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, namespace_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
output.writeBytes(2, qualifier_);
}
- getUnknownFields().writeTo(output);
+ unknownFields.writeTo(output);
}
- private int memoizedSerializedSize = -1;
public int getSerializedSize() {
- int size = memoizedSerializedSize;
+ int size = memoizedSize;
if (size != -1) return size;
size = 0;
@@ -474,19 +466,13 @@ public final class HBaseProtos {
size += com.google.protobuf.CodedOutputStream
.computeBytesSize(2, qualifier_);
}
- size += getUnknownFields().getSerializedSize();
- memoizedSerializedSize = size;
+ size += unknownFields.getSerializedSize();
+ memoizedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
- protected java.lang.Object writeReplace()
- throws java.io.ObjectStreamException {
- return super.writeReplace();
- }
-
- @java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
@@ -507,12 +493,10 @@ public final class HBaseProtos {
result = result && getQualifier()
.equals(other.getQualifier());
}
- result = result &&
- getUnknownFields().equals(other.getUnknownFields());
+ result = result && unknownFields.equals(other.unknownFields);
return result;
}
- private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
@@ -528,7 +512,7 @@ public final class HBaseProtos {
hash = (37 * hash) + QUALIFIER_FIELD_NUMBER;
hash = (53 * hash) + getQualifier().hashCode();
}
- hash = (29 * hash) + getUnknownFields().hashCode();
+ hash = (29 * hash) + unknownFields.hashCode();
memoizedHashCode = hash;
return hash;
}
@@ -556,66 +540,78 @@ public final class HBaseProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
- public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder() {
+ return DEFAULT_INSTANCE.toBuilder();
+ }
public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName prototype) {
- return newBuilder().mergeFrom(prototype);
+ return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+ }
+ public Builder toBuilder() {
+ return this == DEFAULT_INSTANCE
+ ? new Builder() : new Builder().mergeFrom(this);
}
- public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
- * Protobuf type {@code hbase.pb.TableName}
- *
* <pre>
**
* Table Name
* </pre>
+ *
+ * Protobuf type {@code hbase.pb.TableName}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder {
+ com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ // @@protoc_insertion_point(builder_implements:hbase.pb.TableName)
+ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableName_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableName_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -628,18 +624,15 @@ public final class HBaseProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ if (com.google.protobuf.GeneratedMessageV3
+ .alwaysUseFieldBuilders) {
}
}
- private static Builder create() {
- return new Builder();
- }
-
public Builder clear() {
super.clear();
namespace_ = com.google.protobuf.ByteString.EMPTY;
@@ -649,10 +642,6 @@ public final class HBaseProtos {
return this;
}
- public Builder clone() {
- return create().mergeFrom(buildPartial());
- }
-
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableName_descriptor;
@@ -687,6 +676,32 @@ public final class HBaseProtos {
return result;
}
+ public Builder clone() {
+ return (Builder) super.clone();
+ }
+ public Builder setField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ Object value) {
+ return (Builder) super.setField(field, value);
+ }
+ public Builder clearField(
+ com.google.protobuf.Descriptors.FieldDescriptor field) {
+ return (Builder) super.clearField(field);
+ }
+ public Builder clearOneof(
+ com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ return (Builder) super.clearOneof(oneof);
+ }
+ public Builder setRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ int index, Object value) {
+ return (Builder) super.setRepeatedField(field, index, value);
+ }
+ public Builder addRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ Object value) {
+ return (Builder) super.addRepeatedField(field, value);
+ }
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName)other);
@@ -704,17 +719,16 @@ public final class HBaseProtos {
if (other.hasQualifier()) {
setQualifier(other.getQualifier());
}
- this.mergeUnknownFields(other.getUnknownFields());
+ this.mergeUnknownFields(other.unknownFields);
+ onChanged();
return this;
}
public final boolean isInitialized() {
if (!hasNamespace()) {
-
return false;
}
if (!hasQualifier()) {
-
return false;
}
return true;
@@ -729,7 +743,7 @@ public final class HBaseProtos {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName) e.getUnfinishedMessage();
- throw e;
+ throw e.unwrapIOException();
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
@@ -739,7 +753,6 @@ public final class HBaseProtos {
}
private int bitField0_;
- // required bytes namespace = 1;
private com.google.protobuf.ByteString namespace_ = com.google.protobuf.ByteString.EMPTY;
/**
* <code>required bytes namespace = 1;</code>
@@ -775,7 +788,6 @@ public final class HBaseProtos {
return this;
}
- // required bytes qualifier = 2;
private com.google.protobuf.ByteString qualifier_ = com.google.protobuf.ByteString.EMPTY;
/**
* <code>required bytes qualifier = 2;</code>
@@ -810,22 +822,59 @@ public final class HBaseProtos {
onChanged();
return this;
}
+ public final Builder setUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.setUnknownFields(unknownFields);
+ }
+
+ public final Builder mergeUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.mergeUnknownFields(unknownFields);
+ }
+
// @@protoc_insertion_point(builder_scope:hbase.pb.TableName)
}
+ // @@protoc_insertion_point(class_scope:hbase.pb.TableName)
+ private static final org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName DEFAULT_INSTANCE;
static {
- defaultInstance = new TableName(true);
- defaultInstance.initFields();
+ DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName();
+ }
+
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getDefaultInstance() {
+ return DEFAULT_INSTANCE;
+ }
+
+ @java.lang.Deprecated public static final com.google.protobuf.Parser<TableName>
+ PARSER = new com.google.protobuf.AbstractParser<TableName>() {
+ public TableName parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new TableName(input, extensionRegistry);
+ }
+ };
+
+ public static com.google.protobuf.Parser<TableName> parser() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<TableName> getParserForType() {
+ return PARSER;
+ }
+
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getDefaultInstanceForType() {
+ return DEFAULT_INSTANCE;
}
- // @@protoc_insertion_point(class_scope:hbase.pb.TableName)
}
- public interface TableSchemaOrBuilder
- extends com.google.protobuf.MessageOrBuilder {
+ public interface TableSchemaOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:hbase.pb.TableSchema)
+ com.google.protobuf.MessageOrBuilder {
- // optional .hbase.pb.TableName table_name = 1;
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
@@ -839,7 +888,6 @@ public final class HBaseProtos {
*/
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder();
- // repeated .hbase.pb.BytesBytesPair attributes = 2;
/**
* <code>repeated .hbase.pb.BytesBytesPair attributes = 2;</code>
*/
@@ -864,7 +912,6 @@ public final class HBaseProtos {
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getAttributesOrBuilder(
int index);
- // repeated .hbase.pb.ColumnFamilySchema column_families = 3;
/**
* <code>repeated .hbase.pb.ColumnFamilySchema column_families = 3;</code>
*/
@@ -889,7 +936,6 @@ public final class HBaseProtos {
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder(
int index);
- // repeated .hbase.pb.NameStringPair configuration = 4;
/**
* <code>repeated .hbase.pb.NameStringPair configuration = 4;</code>
*/
@@ -915,44 +961,38 @@ public final class HBaseProtos {
int index);
}
/**
- * Protobuf type {@code hbase.pb.TableSchema}
- *
* <pre>
**
* Table Schema
* Inspired by the rest TableSchema
* </pre>
+ *
+ * Protobuf type {@code hbase.pb.TableSchema}
*/
- public static final class TableSchema extends
- com.google.protobuf.GeneratedMessage
- implements TableSchemaOrBuilder {
+ public static final class TableSchema extends
+ com.google.protobuf.GeneratedMessageV3 implements
+ // @@protoc_insertion_point(message_implements:hbase.pb.TableSchema)
+ TableSchemaOrBuilder {
// Use TableSchema.newBuilder() to construct.
- private TableSchema(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ private TableSchema(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
- this.unknownFields = builder.getUnknownFields();
}
- private TableSchema(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
- private static final TableSchema defaultInstance;
- public static TableSchema getDefaultInstance() {
- return defaultInstance;
- }
-
- public TableSchema getDefaultInstanceForType() {
- return defaultInstance;
+ private TableSchema() {
+ attributes_ = java.util.Collections.emptyList();
+ columnFamilies_ = java.util.Collections.emptyList();
+ configuration_ = java.util.Collections.emptyList();
}
- private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
+ getUnknownFields() {
return this.unknownFields;
}
private TableSchema(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- initFields();
+ this();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
@@ -989,7 +1029,8 @@ public final class HBaseProtos {
attributes_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair>();
mutable_bitField0_ |= 0x00000002;
}
- attributes_.add(input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair.PARSER, extensionRegistry));
+ attributes_.add(
+ input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair.PARSER, extensionRegistry));
break;
}
case 26: {
@@ -997,7 +1038,8 @@ public final class HBaseProtos {
columnFamilies_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema>();
mutable_bitField0_ |= 0x00000004;
}
- columnFamilies_.add(input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.PARSER, extensionRegistry));
+ columnFamilies_.add(
+ input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.PARSER, extensionRegistry));
break;
}
case 34: {
@@ -1005,7 +1047,8 @@ public final class HBaseProtos {
configuration_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair>();
mutable_bitField0_ |= 0x00000008;
}
- configuration_.add(input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair.PARSER, extensionRegistry));
+ configuration_.add(
+ input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair.PARSER, extensionRegistry));
break;
}
}
@@ -1014,7 +1057,7 @@ public final class HBaseProtos {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
- e.getMessage()).setUnfinishedMessage(this);
+ e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
attributes_ = java.util.Collections.unmodifiableList(attributes_);
@@ -1034,30 +1077,14 @@ public final class HBaseProtos {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableSchema_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableSchema_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.class, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder.class);
}
- public static com.google.protobuf.Parser<TableSchema> PARSER =
- new com.google.protobuf.AbstractParser<TableSchema>() {
- public TableSchema parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return new TableSchema(input, extensionRegistry);
- }
- };
-
- @java.lang.Override
- public com.google.protobuf.Parser<TableSchema> getParserForType() {
- return PARSER;
- }
-
private int bitField0_;
- // optional .hbase.pb.TableName table_name = 1;
public static final int TABLE_NAME_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_;
/**
@@ -1070,16 +1097,15 @@ public final class HBaseProtos {
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() {
- return tableName_;
+ return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
}
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
- return tableName_;
+ return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
}
- // repeated .hbase.pb.BytesBytesPair attributes = 2;
public static final int ATTRIBUTES_FIELD_NUMBER = 2;
private java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair> attributes_;
/**
@@ -1115,7 +1141,6 @@ public final class HBaseProtos {
return attributes_.get(index);
}
- // repeated .hbase.pb.ColumnFamilySchema column_families = 3;
public static final int COLUMN_FAMILIES_FIELD_NUMBER = 3;
private java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema> columnFamilies_;
/**
@@ -1151,7 +1176,6 @@ public final class HBaseProtos {
return columnFamilies_.get(index);
}
- // repeated .hbase.pb.NameStringPair configuration = 4;
public static final int CONFIGURATION_FIELD_NUMBER = 4;
private java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair> configuration_;
/**
@@ -1187,16 +1211,11 @@ public final class HBaseProtos {
return configuration_.get(index);
}
- private void initFields() {
- tableName_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
- attributes_ = java.util.Collections.emptyList();
- columnFamilies_ = java.util.Collections.emptyList();
- configuration_ = java.util.Collections.emptyList();
- }
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
- if (isInitialized != -1) return isInitialized == 1;
+ if (isInitialized == 1) return true;
+ if (isInitialized == 0) return false;
if (hasTableName()) {
if (!getTableName().isInitialized()) {
@@ -1228,9 +1247,8 @@ public final class HBaseProtos {
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
- getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- output.writeMessage(1, tableName_);
+ output.writeMessage(1, getTableName());
}
for (int i = 0; i < attributes_.size(); i++) {
output.writeMessage(2, attributes_.get(i));
@@ -1241,18 +1259,17 @@ public final class HBaseProtos {
for (int i = 0; i < configuration_.size(); i++) {
output.writeMessage(4, configuration_.get(i));
}
- getUnknownFields().writeTo(output);
+ unknownFields.writeTo(output);
}
- private int memoizedSerializedSize = -1;
public int getSerializedSize() {
- int size = memoizedSerializedSize;
+ int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
- .computeMessageSize(1, tableName_);
+ .computeMessageSize(1, getTableName());
}
for (int i = 0; i < attributes_.size(); i++) {
size += com.google.protobuf.CodedOutputStream
@@ -1266,19 +1283,13 @@ public final class HBaseProtos {
size += com.google.protobuf.CodedOutputStream
.computeMessageSize(4, configuration_.get(i));
}
- size += getUnknownFields().getSerializedSize();
- memoizedSerializedSize = size;
+ size += unknownFields.getSerializedSize();
+ memoizedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
- protected java.lang.Object writeReplace()
- throws java.io.ObjectStreamException {
- return super.writeReplace();
- }
-
- @java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
@@ -1300,12 +1311,10 @@ public final class HBaseProtos {
.equals(other.getColumnFamiliesList());
result = result && getConfigurationList()
.equals(other.getConfigurationList());
- result = result &&
- getUnknownFields().equals(other.getUnknownFields());
+ result = result && unknownFields.equals(other.unknownFields);
return result;
}
- private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
@@ -1329,7 +1338,7 @@ public final class HBaseProtos {
hash = (37 * hash) + CONFIGURATION_FIELD_NUMBER;
hash = (53 * hash) + getConfigurationList().hashCode();
}
- hash = (29 * hash) + getUnknownFields().hashCode();
+ hash = (29 * hash) + unknownFields.hashCode();
memoizedHashCode = hash;
return hash;
}
@@ -1357,67 +1366,79 @@ public final class HBaseProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
- public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder() {
+ return DEFAULT_INSTANCE.toBuilder();
+ }
public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema prototype) {
- return newBuilder().mergeFrom(prototype);
+ return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+ }
+ public Builder toBuilder() {
+ return this == DEFAULT_INSTANCE
+ ? new Builder() : new Builder().mergeFrom(this);
}
- public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
- * Protobuf type {@code hbase.pb.TableSchema}
- *
* <pre>
**
* Table Schema
* Inspired by the rest TableSchema
* </pre>
+ *
+ * Protobuf type {@code hbase.pb.TableSchema}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder {
+ com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ // @@protoc_insertion_point(builder_implements:hbase.pb.TableSchema)
+ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableSchema_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableSchema_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1430,26 +1451,23 @@ public final class HBaseProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ if (com.google.protobuf.GeneratedMessageV3
+ .alwaysUseFieldBuilders) {
getTableNameFieldBuilder();
getAttributesFieldBuilder();
getColumnFamiliesFieldBuilder();
getConfigurationFieldBuilder();
}
}
- private static Builder create() {
- return new Builder();
- }
-
public Builder clear() {
super.clear();
if (tableNameBuilder_ == null) {
- tableName_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = null;
} else {
tableNameBuilder_.clear();
}
@@ -1475,10 +1493,6 @@ public final class HBaseProtos {
return this;
}
- public Builder clone() {
- return create().mergeFrom(buildPartial());
- }
-
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableSchema_descriptor;
@@ -1540,6 +1554,32 @@ public final class HBaseProtos {
return result;
}
+ public Builder clone() {
+ return (Builder) super.clone();
+ }
+ public Builder setField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ Object value) {
+ return (Builder) super.setField(field, value);
+ }
+ public Builder clearField(
+ com.google.protobuf.Descriptors.FieldDescriptor field) {
+ return (Builder) super.clearField(field);
+ }
+ public Builder clearOneof(
+ com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ return (Builder) super.clearOneof(oneof);
+ }
+ public Builder setRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ int index, Object value) {
+ return (Builder) super.setRepeatedField(field, index, value);
+ }
+ public Builder addRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ Object value) {
+ return (Builder) super.addRepeatedField(field, value);
+ }
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema)other);
@@ -1573,7 +1613,7 @@ public final class HBaseProtos {
attributes_ = other.attributes_;
bitField0_ = (bitField0_ & ~0x00000002);
attributesBuilder_ =
- com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ?
+ com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getAttributesFieldBuilder() : null;
} else {
attributesBuilder_.addAllMessages(other.attributes_);
@@ -1599,7 +1639,7 @@ public final class HBaseProtos {
columnFamilies_ = other.columnFamilies_;
bitField0_ = (bitField0_ & ~0x00000004);
columnFamiliesBuilder_ =
- com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ?
+ com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getColumnFamiliesFieldBuilder() : null;
} else {
columnFamiliesBuilder_.addAllMessages(other.columnFamilies_);
@@ -1625,39 +1665,36 @@ public final class HBaseProtos {
configuration_ = other.configuration_;
bitField0_ = (bitField0_ & ~0x00000008);
configurationBuilder_ =
- com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ?
+ com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getConfigurationFieldBuilder() : null;
} else {
configurationBuilder_.addAllMessages(other.configuration_);
}
}
}
- this.mergeUnknownFields(other.getUnknownFields());
+ this.mergeUnknownFields(other.unknownFields);
+ onChanged();
return this;
}
public final boolean isInitialized() {
if (hasTableName()) {
if (!getTableName().isInitialized()) {
-
return false;
}
}
for (int i = 0; i < getAttributesCount(); i++) {
if (!getAttributes(i).isInitialized()) {
-
return false;
}
}
for (int i = 0; i < getColumnFamiliesCount(); i++) {
if (!getColumnFamilies(i).isInitialized()) {
-
return false;
}
}
for (int i = 0; i < getConfigurationCount(); i++) {
if (!getConfiguration(i).isInitialized()) {
-
return false;
}
}
@@ -1673,7 +1710,7 @@ public final class HBaseProtos {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema) e.getUnfinishedMessage();
- throw e;
+ throw e.unwrapIOException();
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
@@ -1683,9 +1720,8 @@ public final class HBaseProtos {
}
private int bitField0_;
- // optional .hbase.pb.TableName table_name = 1;
- private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
- private com.google.protobuf.SingleFieldBuilder<
+ private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_ = null;
+ private com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
@@ -1698,7 +1734,7 @@ public final class HBaseProtos {
*/
public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() {
if (tableNameBuilder_ == null) {
- return tableName_;
+ return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
} else {
return tableNameBuilder_.getMessage();
}
@@ -1739,6 +1775,7 @@ public final class HBaseProtos {
public Builder mergeTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (((bitField0_ & 0x00000001) == 0x00000001) &&
+ tableName_ != null &&
tableName_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) {
tableName_ =
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
@@ -1757,7 +1794,7 @@ public final class HBaseProtos {
*/
public Builder clearTableName() {
if (tableNameBuilder_ == null) {
- tableName_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = null;
onChanged();
} else {
tableNameBuilder_.clear();
@@ -1780,19 +1817,20 @@ public final class HBaseProtos {
if (tableNameBuilder_ != null) {
return tableNameBuilder_.getMessageOrBuilder();
} else {
- return tableName_;
+ return tableName_ == null ?
+ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
}
}
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- private com.google.protobuf.SingleFieldBuilder<
+ private com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>
getTableNameFieldBuilder() {
if (tableNameBuilder_ == null) {
- tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder<
+ tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
- tableName_,
+ getTableName(),
getParentForChildren(),
isClean());
tableName_ = null;
@@ -1800,7 +1838,6 @@ public final class HBaseProtos {
return tableNameBuilder_;
}
- // repeated .hbase.pb.BytesBytesPair attributes = 2;
private java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair> attributes_ =
java.util.Collections.emptyList();
private void ensureAttributesIsMutable() {
@@ -1810,7 +1847,7 @@ public final class HBaseProtos {
}
}
- private com.google.protobuf.RepeatedFieldBuilder<
+ private com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder> attributesBuilder_;
/**
@@ -1942,7 +1979,8 @@ public final class HBaseProtos {
java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair> values) {
if (attributesBuilder_ == null) {
ensureAttributesIsMutable();
- super.addAll(values, attributes_);
+ com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ values, attributes_);
onChanged();
} else {
attributesBuilder_.addAllMessages(values);
@@ -2025,11 +2063,11 @@ public final class HBaseProtos {
getAttributesBuilderList() {
return getAttributesFieldBuilder().getBuilderList();
}
- private com.google.protobuf.RepeatedFieldBuilder<
+ private com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder>
getAttributesFieldBuilder() {
if (attributesBuilder_ == null) {
- attributesBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
+ attributesBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder>(
attributes_,
((bitField0_ & 0x00000002) == 0x00000002),
@@ -2040,7 +2078,6 @@ public final class HBaseProtos {
return attributesBuilder_;
}
- // repeated .hbase.pb.ColumnFamilySchema column_families = 3;
private java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema> columnFamilies_ =
java.util.Collections.emptyList();
private void ensureColumnFamiliesIsMutable() {
@@ -2050,7 +2087,7 @@ public final class HBaseProtos {
}
}
- private com.google.protobuf.RepeatedFieldBuilder<
+ private com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> columnFamiliesBuilder_;
/**
@@ -2182,7 +2219,8 @@ public final class HBaseProtos {
java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema> values) {
if (columnFamiliesBuilder_ == null) {
ensureColumnFamiliesIsMutable();
- super.addAll(values, columnFamilies_);
+ com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ values, columnFamilies_);
onChanged();
} else {
columnFamiliesBuilder_.addAllMessages(values);
@@ -2265,11 +2303,11 @@ public final class HBaseProtos {
getColumnFamiliesBuilderList() {
return getColumnFamiliesFieldBuilder().getBuilderList();
}
- private com.google.protobuf.RepeatedFieldBuilder<
+ private com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder>
getColumnFamiliesFieldBuilder() {
if (columnFamiliesBuilder_ == null) {
- columnFamiliesBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
+ columnFamiliesBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder>(
columnFamilies_,
((bitField0_ & 0x00000004) == 0x00000004),
@@ -2280,7 +2318,6 @@ public final class HBaseProtos {
return columnFamiliesBuilder_;
}
- // repeated .hbase.pb.NameStringPair configuration = 4;
private java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair> configuration_ =
java.util.Collections.emptyList();
private void ensureConfigurationIsMutable() {
@@ -2290,7 +2327,7 @@ public final class HBaseProtos {
}
}
- private com.google.protobuf.RepeatedFieldBuilder<
+ private com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPairOrBuilder> configurationBuilder_;
/**
@@ -2422,7 +2459,8 @@ public final class HBaseProtos {
java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair> values) {
if (configurationBuilder_ == null) {
ensureConfigurationIsMutable();
- super.addAll(values, configuration_);
+ com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ values, configuration_);
onChanged();
} else {
configurationBuilder_.addAllMessages(values);
@@ -2505,11 +2543,11 @@ public final class HBaseProtos {
getConfigurationBuilderList() {
return getConfigurationFieldBuilder().getBuilderList();
}
- private com.google.protobuf.RepeatedFieldBuilder<
+ private com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPairOrBuilder>
getConfigurationFieldBuilder() {
if (configurationBuilder_ == null) {
- configurationBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
+ configurationBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPairOrBuilder>(
configuration_,
((bitField0_ & 0x00000008) == 0x00000008),
@@ -2519,76 +2557,105 @@ public final class HBaseProtos {
}
return configurationBuilder_;
}
+ public final Builder setUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.setUnknownFields(unknownFields);
+ }
+
+ public final Builder mergeUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.mergeUnknownFields(unknownFields);
+ }
+
// @@protoc_insertion_point(builder_scope:hbase.pb.TableSchema)
}
+ // @@protoc_insertion_point(class_scope:hbase.pb.TableSchema)
+ private static final org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema DEFAULT_INSTANCE;
static {
- defaultInstance = new TableSchema(true);
- defaultInstance.initFields();
+ DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema();
+ }
+
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema getDefaultInstance() {
+ return DEFAULT_INSTANCE;
+ }
+
+ @java.lang.Deprecated public static final com.google.protobuf.Parser<TableSchema>
+ PARSER = new com.google.protobuf.AbstractParser<TableSchema>() {
+ public TableSchema parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new TableSchema(input, extensionRegistry);
+ }
+ };
+
+ public static com.google.protobuf.Parser<TableSchema> parser() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<TableSchema> getParserForType() {
+ return PARSER;
+ }
+
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema getDefaultInstanceForType() {
+ return DEFAULT_INSTANCE;
}
- // @@protoc_insertion_point(class_scope:hbase.pb.TableSchema)
}
- public interface TableStateOrBuilder
- extends com.google.protobuf.MessageOrBuilder {
+ public interface TableStateOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:hbase.pb.TableState)
+ com.google.protobuf.MessageOrBuilder {
- // required .hbase.pb.TableState.State state = 1;
/**
- * <code>required .hbase.pb.TableState.State state = 1;</code>
- *
* <pre>
* This is the table's state.
* </pre>
+ *
+ * <code>required .hbase.pb.TableState.State state = 1;</code>
*/
boolean hasState();
/**
- * <code>required .hbase.pb.TableState.State state = 1;</code>
- *
* <pre>
* This is the table's state.
* </pre>
+ *
+ * <code>required .hbase.pb.TableState.State state = 1;</code>
*/
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.State getState();
}
/**
- * Protobuf type {@code hbase.pb.TableState}
- *
* <pre>
** Denotes state of the table
* </pre>
+ *
+ * Protobuf type {@code hbase.pb.TableState}
*/
- public static final class TableState extends
- com.google.protobuf.GeneratedMessage
- implements TableStateOrBuilder {
+ public static final class TableState extends
+ com.google.protobuf.GeneratedMessageV3 implements
+ // @@protoc_insertion_point(message_implements:hbase.pb.TableState)
+ TableStateOrBuilder {
// Use TableState.newBuilder() to construct.
- private TableState(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ private TableState(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
- this.unknownFields = builder.getUnknownFields();
- }
- private TableState(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
- private static final TableState defaultInstance;
- public static TableState getDefaultInstance() {
- return defaultInstance;
}
-
- public TableState getDefaultInstanceForType() {
- return defaultInstance;
+ private TableState() {
+ state_ = 0;
}
- private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
+ getUnknownFields() {
return this.unknownFields;
}
private TableState(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- initFields();
+ this();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
@@ -2614,7 +2681,7 @@ public final class HBaseProtos {
unknownFields.mergeVarintField(1, rawValue);
} else {
bitField0_ |= 0x00000001;
- state_ = value;
+ state_ = rawValue;
}
break;
}
@@ -2624,7 +2691,7 @@ public final class HBaseProtos {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
- e.getMessage()).setUnfinishedMessage(this);
+ e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
@@ -2635,53 +2702,38 @@ public final class HBaseProtos {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableState_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableState_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.class, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.Builder.class);
}
- public static com.google.protobuf.Parser<TableState> PARSER =
- new com.google.protobuf.AbstractParser<TableState>() {
- public TableState parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return new TableState(input, extensionRegistry);
- }
- };
-
- @java.lang.Override
- public com.google.protobuf.Parser<TableState> getParserForType() {
- return PARSER;
- }
-
/**
- * Protobuf enum {@code hbase.pb.TableState.State}
- *
* <pre>
* Table's current state
* </pre>
+ *
+ * Protobuf enum {@code hbase.pb.TableState.State}
*/
public enum State
implements com.google.protobuf.ProtocolMessageEnum {
/**
* <code>ENABLED = 0;</code>
*/
- ENABLED(0, 0),
+ ENABLED(0),
/**
* <code>DISABLED = 1;</code>
*/
- DISABLED(1, 1),
+ DISABLED(1),
/**
* <code>DISABLING = 2;</code>
*/
- DISABLING(2, 2),
+ DISABLING(2),
/**
* <code>ENABLING = 3;</code>
*/
- ENABLING(3, 3),
+ ENABLING(3),
;
/**
@@ -2702,9 +2754,19 @@ public final class HBaseProtos {
public static final int ENABLING_VALUE = 3;
- public final int getNumber() { return value; }
+ public final int getNumber() {
+ return value;
+ }
+ /**
+ * @deprecated Use {@link #forNumber(int)} instead.
+ */
+ @java.lang.Deprecated
public static State valueOf(int value) {
+ return forNumber(value);
+ }
+
+ public static State forNumber(int value) {
switch (value) {
case 0: return ENABLED;
case 1: return DISABLED;
@@ -2718,17 +2780,17 @@ public final class HBaseProtos {
internalGetValueMap() {
return internalValueMap;
}
- private static com.google.protobuf.Internal.EnumLiteMap<State>
- internalValueMap =
+ private static final com.google.protobuf.Internal.EnumLiteMap<
+ State> internalValueMap =
new com.google.protobuf.Internal.EnumLiteMap<State>() {
public State findValueByNumber(int number) {
- return State.valueOf(number);
+ return State.forNumber(number);
}
};
public final com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
- return getDescriptor().getValues().get(index);
+ return getDescriptor().getValues().get(ordinal());
}
public final com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
@@ -2750,11 +2812,9 @@ public final class HBaseProtos {
return VALUES[desc.getIndex()];
}
- private final int index;
private final int value;
- private State(int index, int value) {
- this.index = index;
+ private State(int value) {
this.value = value;
}
@@ -2762,37 +2822,35 @@ public final class HBaseProtos {
}
private int bitField0_;
- // required .hbase.pb.TableState.State state = 1;
public static final int STATE_FIELD_NUMBER = 1;
- private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.State state_;
+ private int state_;
/**
- * <code>required .hbase.pb.TableState.State state = 1;</code>
- *
* <pre>
* This is the table's state.
* </pre>
+ *
+ * <code>required .hbase.pb.TableState.State state = 1;</code>
*/
public boolean hasState() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
- * <code>required .hbase.pb.TableState.State state = 1;</code>
- *
* <pre>
* This is the table's state.
* </pre>
+ *
+ * <code>required .hbase.pb.TableState.State state = 1;</code>
*/
public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.State getState() {
- return state_;
+ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.State result = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.State.valueOf(state_);
+ return result == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.State.ENABLED : result;
}
- private void initFields() {
- state_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.State.ENABLED;
- }
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
- if (isInitialized != -1) return isInitialized == 1;
+ if (isInitialized == 1) return true;
+ if (isInitialized == 0) return false;
if (!hasState()) {
memoizedIsInitialized = 0;
@@ -2804,36 +2862,28 @@ public final class HBaseProtos {
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
- getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- output.writeEnum(1, state_.getNumber());
+ output.writeEnum(1, state_);
}
- getUnknownFields().writeTo(output);
+ unknownFields.writeTo(output);
}
- private int memoizedSerializedSize = -1;
public int getSerializedSize() {
- int size = memoizedSerializedSize;
+ int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
- .computeEnumSize(1, state_.getNumber());
+ .computeEnumSize(1, state_);
}
- size += getUnknownFields().getSerializedSize();
- memoizedSerializedSize = size;
+ size += unknownFields.getSerializedSize();
+ memoizedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
- protected java.lang.Object writeReplace()
- throws java.io.ObjectStreamException {
- return super.writeReplace();
- }
-
- @java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
@@ -2846,15 +2896,12 @@ public final class HBaseProtos {
boolean result = true;
result = result && (hasState() == other.hasState());
if (hasState()) {
- result = result &&
- (getState() == other.getState());
+ result = result && state_ == other.state_;
}
- result = result &&
- getUnknownFields().equals(other.getUnknownFields());
+ result = result && unknownFields.equals(other.unknownFields);
return result;
}
- private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
@@ -2864,9 +2911,9 @@ public final class HBaseProtos {
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasState()) {
hash = (37 * hash) + STATE_FIELD_NUMBER;
- hash = (53 * hash) + hashEnum(getState());
+ hash = (53 * hash) + state_;
}
- hash = (29 * hash) + getUnknownFields().hashCode();
+ hash = (29 * hash) + unknownFields.hashCode();
memoizedHashCode = hash;
return hash;
}
@@ -2894,65 +2941,77 @@ public final class HBaseProtos {
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
- public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder() {
+ return DEFAULT_INSTANCE.toBuilder();
+ }
public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState prototype) {
- return newBuilder().mergeFrom(prototype);
+ return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+ }
+ public Builder toBuilder() {
+ return this == DEFAULT_INSTANCE
+ ? new Builder() : new Builder().mergeFrom(this);
}
- public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
- * Protobuf type {@code hbase.pb.TableState}
- *
* <pre>
** Denotes state of the table
* </pre>
+ *
+ * Protobuf type {@code hbase.pb.TableState}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableStateOrBuilder {
+ com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ // @@protoc_insertion_point(builder_implements:hbase.pb.TableState)
+ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableStateOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableState_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableState_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2965,29 +3024,22 @@ public final class HBaseProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ if (com.google.protobuf.GeneratedMessageV3
+ .alwaysUseFieldBuilders) {
}
}
- private static Builder create() {
- return new Builder();
- }
-
public Builder clear() {
super.clear();
- state_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.State.ENABLED;
+ state_ = 0;
bitField0_ = (bitField0_ & ~0x00000001);
return this;
}
- public Builder clone() {
- return create().mergeFrom(buildPartial());
- }
-
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableState_descriptor;
@@ -3018,6 +3070,32 @@ public final class HBaseProtos {
return result;
}
+ public Builder clone() {
+ return (Builder) super.clone();
+ }
+ public Builder setField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ Object value) {
+ return (Builder) super.setField(field, value);
+ }
+ public Builder clearField(
+ com.google.protobuf.Descriptors.FieldDescriptor field) {
+ return (Builder) super.clearField(field);
+ }
+ public Builder clearOneof(
+ com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ return (Builder) super.clearOneof(oneof);
+ }
+ public Builder setRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ int index, Object value) {
+ return (Builder) super.setRepeatedField(field, index, value);
+ }
+ public Builder addRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ Object value) {
+ return (Builder) super.addRepeatedField(field, value);
+ }
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState) {
return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState)other);
@@ -3032,13 +3110,13 @@ public final class HBaseProtos {
if (other.hasState()) {
setState(other.getState());
}
- this.mergeUnknownFields(other.getUnknownFields());
+ this.mergeUnknownFields(other.unknownFields);
+ onChanged();
return this;
}
public final boolean isInitialized() {
if (!hasState()) {
-
return false;
}
return true;
@@ -3053,7 +3131,7 @@ public final class HBaseProtos {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState) e.getUnfinishedMessage();
- throw e;
+ throw e.unwrapIOException();
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
@@ -3063,74 +3141,111 @@ public final class HBaseProtos {
}
private int bitField0_;
- // required .hbase.pb.TableState.State state = 1;
- private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.State state_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.State.ENABLED;
+ private int state_ = 0;
/**
- * <code>required .hbase.pb.TableState.State state = 1;</code>
- *
* <pre>
* This is the table's state.
* </pre>
+ *
+ * <code>required .hbase.pb.TableState.State state = 1;</code>
*/
public boolean hasState() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
- * <code>required .hbase.pb.TableState.State state = 1;</code>
- *
* <pre>
* This is the table's state.
* </pre>
+ *
+ * <code>required .hbase.pb.TableState.State state = 1;</code>
*/
public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.State getState() {
- return state_;
+ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.State result = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.State.valueOf(state_);
+ return result == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.State.ENABLED : result;
}
/**
- * <code>required .hbase.pb.TableState.State state = 1;</code>
- *
* <pre>
* This is the table's state.
* </pre>
+ *
+ * <code>required .hbase.pb.TableState.State state = 1;</code>
*/
public Builder setState(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.State value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000001;
- state_ = value;
+ state_ = value.getNumber();
onChanged();
return this;
}
/**
- * <code>required .hbase.pb.TableState.State state = 1;</code>
- *
* <pre>
* This is the table's state.
* </pre>
+ *
+ * <code>required .hbase.pb.TableState.State state = 1;</code>
*/
public Builder clearState() {
bitField0_ = (bitField0_ & ~0x00000001);
- state_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.State.ENABLED;
+ state_ = 0;
onChanged();
return this;
}
+ public final Builder setUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.setUnknownFields(unknownFields);
+ }
+
+ public final Builder mergeUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.mergeUnknownFields(unknownFields);
+ }
+
// @@protoc_insertion_point(builder_scope:hbase.pb.TableState)
}
+ // @@protoc_insertion_point(class_scope:hbase.pb.TableState)
+ private static final org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState DEFAULT_INSTANCE;
static {
- defaultInstance = new TableState(true);
- defaultInstance.initFields();
+ DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState();
}
- // @@protoc_insertion_point(class_scope:hbase.pb.TableState)
- }
-
- public interface ColumnFamilySchemaOrBuilder
- extends com.google.protobuf.MessageOrBuilder {
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState getDefaultInstance() {
+ return DEFAULT_INSTANCE;
+ }
- // required bytes name = 1;
- /**
+ @java.lang.Deprecated public static final com.google.protobuf.Parser<TableState>
+ PARSER = new com.google.protobuf.AbstractParser<TableState>() {
+ public TableState parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new TableState(input, extensionRegistry);
+ }
+ };
+
+ public static com.google.protobuf.Parser<TableState> parser() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<TableState> getParserForType() {
+ return PARSER;
+ }
+
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState getDefaultInstanceForType() {
+ return DEFAULT_INSTANCE;
+ }
+
+ }
+
+ public interface ColumnFamilySchemaOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:hbase.pb.ColumnFamilySchema)
+ com.google.protobuf.MessageOrBuilder {
+
+ /**
* <code>required bytes name = 1;</code>
*/
boolean hasName();
@@ -3139,7 +3254,6 @@ public final class HBaseProtos {
*/
com.google.protobuf.ByteString getName();
- // repeated .hbase.pb.BytesBytesPair attributes = 2;
/**
* <code>repeated .hbase.pb.BytesBytesPair attributes = 2;</code>
*/
@@ -3164,7 +3278,6 @@ public final class HBaseProtos {
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getAttributesOrBuilder(
int index);
- // repeated .hbase.pb.NameStringPair configuration = 3;
/**
* <code>repeated .hbase.pb.NameStringPair configuration = 3;</code>
*/
@@ -3190,44 +3303,38 @@ public final class HBaseProtos {
int index);
}
/**
- * Protobuf type {@code hbase.pb.ColumnFamilySchema}
- *
* <pre>
**
* Column Family Schema
* Inspired by the rest ColumSchemaMessage
* </pre>
+ *
+ * Protobuf type {@code hbase.pb.ColumnFamilySchema}
*/
- public static final class ColumnFamilySchema extends
- com.google.protobuf.GeneratedMessage
- implements ColumnFamilySchemaOrBuilder {
+ public static final class ColumnFamilySchema extends
+ com.google.protobuf.GeneratedMessageV3 implements
+ // @@protoc_insertion_point(message_implements:hbase.pb.ColumnFamilySchema)
+ ColumnFamilySchemaOrBuilder {
// Use ColumnFamilySchema.newBuilder() to construct.
- private ColumnFamilySchema(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ private ColumnFamilySchema(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
- this.unknownFields = builder.getUnknownFields();
- }
- private ColumnFamilySchema(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
- private static final ColumnFamilySchema defaultInstance;
- public static ColumnFamilySchema getDefaultInstance() {
- return defaultInstance;
}
-
- public ColumnFamilySchema getDefaultInstanceForType() {
- return defaultInstance;
+ private ColumnFamilySchema() {
+ name_ = com.google.protobuf.ByteString.EMPTY;
+ attributes_ = java.util.Collections.emptyL
<TRUNCATED>