You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2016/09/29 19:37:37 UTC
[23/51] [partial] hbase git commit: HBASE-16264 Figure how to deal
with endpoints and shaded pb Shade our protobufs. Do it in a manner that
makes it so we can still have in our API references to com.google.protobuf
(and in REST). The c.g.p in API is for
http://git-wip-us.apache.org/repos/asf/hbase/blob/17d4b70d/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ClientProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ClientProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ClientProtos.java
new file mode 100644
index 0000000..a43b5a0
--- /dev/null
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/ClientProtos.java
@@ -0,0 +1,39725 @@
+// Generated by the protocol buffer compiler. DO NOT EDIT!
+// source: Client.proto
+
+package org.apache.hadoop.hbase.shaded.protobuf.generated;
+
+public final class ClientProtos {
+ private ClientProtos() {}
+ public static void registerAllExtensions(
+ com.google.protobuf.ExtensionRegistry registry) {
+ }
+ /**
+ * Protobuf enum {@code hbase.pb.Consistency}
+ *
+ * <pre>
+ **
+ * Consistency defines the expected consistency level for an operation.
+ * </pre>
+ */
+ public enum Consistency
+ implements com.google.protobuf.ProtocolMessageEnum {
+ /**
+ * <code>STRONG = 0;</code>
+ */
+ STRONG(0, 0),
+ /**
+ * <code>TIMELINE = 1;</code>
+ */
+ TIMELINE(1, 1),
+ ;
+
+ /**
+ * <code>STRONG = 0;</code>
+ */
+ public static final int STRONG_VALUE = 0;
+ /**
+ * <code>TIMELINE = 1;</code>
+ */
+ public static final int TIMELINE_VALUE = 1;
+
+
+ public final int getNumber() { return value; }
+
+ public static Consistency valueOf(int value) {
+ switch (value) {
+ case 0: return STRONG;
+ case 1: return TIMELINE;
+ default: return null;
+ }
+ }
+
+ public static com.google.protobuf.Internal.EnumLiteMap<Consistency>
+ internalGetValueMap() {
+ return internalValueMap;
+ }
+ private static com.google.protobuf.Internal.EnumLiteMap<Consistency>
+ internalValueMap =
+ new com.google.protobuf.Internal.EnumLiteMap<Consistency>() {
+ public Consistency findValueByNumber(int number) {
+ return Consistency.valueOf(number);
+ }
+ };
+
+ public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ getValueDescriptor() {
+ return getDescriptor().getValues().get(index);
+ }
+ public final com.google.protobuf.Descriptors.EnumDescriptor
+ getDescriptorForType() {
+ return getDescriptor();
+ }
+ public static final com.google.protobuf.Descriptors.EnumDescriptor
+ getDescriptor() {
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.getDescriptor().getEnumTypes().get(0);
+ }
+
+ private static final Consistency[] VALUES = values();
+
+ public static Consistency valueOf(
+ com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ if (desc.getType() != getDescriptor()) {
+ throw new java.lang.IllegalArgumentException(
+ "EnumValueDescriptor is not for this type.");
+ }
+ return VALUES[desc.getIndex()];
+ }
+
+ private final int index;
+ private final int value;
+
+ private Consistency(int index, int value) {
+ this.index = index;
+ this.value = value;
+ }
+
+ // @@protoc_insertion_point(enum_scope:hbase.pb.Consistency)
+ }
+
+ public interface AuthorizationsOrBuilder
+ extends com.google.protobuf.MessageOrBuilder {
+
+ // repeated string label = 1;
+ /**
+ * <code>repeated string label = 1;</code>
+ */
+ java.util.List<java.lang.String>
+ getLabelList();
+ /**
+ * <code>repeated string label = 1;</code>
+ */
+ int getLabelCount();
+ /**
+ * <code>repeated string label = 1;</code>
+ */
+ java.lang.String getLabel(int index);
+ /**
+ * <code>repeated string label = 1;</code>
+ */
+ com.google.protobuf.ByteString
+ getLabelBytes(int index);
+ }
+ /**
+ * Protobuf type {@code hbase.pb.Authorizations}
+ *
+ * <pre>
+ **
+ * The protocol buffer version of Authorizations.
+ * </pre>
+ */
+ public static final class Authorizations extends
+ com.google.protobuf.GeneratedMessage
+ implements AuthorizationsOrBuilder {
+ // Use Authorizations.newBuilder() to construct.
+ private Authorizations(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ super(builder);
+ this.unknownFields = builder.getUnknownFields();
+ }
+ private Authorizations(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
+ private static final Authorizations defaultInstance;
+ public static Authorizations getDefaultInstance() {
+ return defaultInstance;
+ }
+
+ public Authorizations getDefaultInstanceForType() {
+ return defaultInstance;
+ }
+
+ private final com.google.protobuf.UnknownFieldSet unknownFields;
+ @java.lang.Override
+ public final com.google.protobuf.UnknownFieldSet
+ getUnknownFields() {
+ return this.unknownFields;
+ }
+ private Authorizations(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ initFields();
+ int mutable_bitField0_ = 0;
+ com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ com.google.protobuf.UnknownFieldSet.newBuilder();
+ try {
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ default: {
+ if (!parseUnknownField(input, unknownFields,
+ extensionRegistry, tag)) {
+ done = true;
+ }
+ break;
+ }
+ case 10: {
+ if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
+ label_ = new com.google.protobuf.LazyStringArrayList();
+ mutable_bitField0_ |= 0x00000001;
+ }
+ label_.add(input.readBytes());
+ break;
+ }
+ }
+ }
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(this);
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(
+ e.getMessage()).setUnfinishedMessage(this);
+ } finally {
+ if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
+ label_ = new com.google.protobuf.UnmodifiableLazyStringList(label_);
+ }
+ this.unknownFields = unknownFields.build();
+ makeExtensionsImmutable();
+ }
+ }
+ public static final com.google.protobuf.Descriptors.Descriptor
+ getDescriptor() {
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Authorizations_descriptor;
+ }
+
+ protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ internalGetFieldAccessorTable() {
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Authorizations_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations.Builder.class);
+ }
+
+ public static com.google.protobuf.Parser<Authorizations> PARSER =
+ new com.google.protobuf.AbstractParser<Authorizations>() {
+ public Authorizations parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new Authorizations(input, extensionRegistry);
+ }
+ };
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<Authorizations> getParserForType() {
+ return PARSER;
+ }
+
+ // repeated string label = 1;
+ public static final int LABEL_FIELD_NUMBER = 1;
+ private com.google.protobuf.LazyStringList label_;
+ /**
+ * <code>repeated string label = 1;</code>
+ */
+ public java.util.List<java.lang.String>
+ getLabelList() {
+ return label_;
+ }
+ /**
+ * <code>repeated string label = 1;</code>
+ */
+ public int getLabelCount() {
+ return label_.size();
+ }
+ /**
+ * <code>repeated string label = 1;</code>
+ */
+ public java.lang.String getLabel(int index) {
+ return label_.get(index);
+ }
+ /**
+ * <code>repeated string label = 1;</code>
+ */
+ public com.google.protobuf.ByteString
+ getLabelBytes(int index) {
+ return label_.getByteString(index);
+ }
+
+ private void initFields() {
+ label_ = com.google.protobuf.LazyStringArrayList.EMPTY;
+ }
+ private byte memoizedIsInitialized = -1;
+ public final boolean isInitialized() {
+ byte isInitialized = memoizedIsInitialized;
+ if (isInitialized != -1) return isInitialized == 1;
+
+ memoizedIsInitialized = 1;
+ return true;
+ }
+
+ public void writeTo(com.google.protobuf.CodedOutputStream output)
+ throws java.io.IOException {
+ getSerializedSize();
+ for (int i = 0; i < label_.size(); i++) {
+ output.writeBytes(1, label_.getByteString(i));
+ }
+ getUnknownFields().writeTo(output);
+ }
+
+ private int memoizedSerializedSize = -1;
+ public int getSerializedSize() {
+ int size = memoizedSerializedSize;
+ if (size != -1) return size;
+
+ size = 0;
+ {
+ int dataSize = 0;
+ for (int i = 0; i < label_.size(); i++) {
+ dataSize += com.google.protobuf.CodedOutputStream
+ .computeBytesSizeNoTag(label_.getByteString(i));
+ }
+ size += dataSize;
+ size += 1 * getLabelList().size();
+ }
+ size += getUnknownFields().getSerializedSize();
+ memoizedSerializedSize = size;
+ return size;
+ }
+
+ private static final long serialVersionUID = 0L;
+ @java.lang.Override
+ protected java.lang.Object writeReplace()
+ throws java.io.ObjectStreamException {
+ return super.writeReplace();
+ }
+
+ @java.lang.Override
+ public boolean equals(final java.lang.Object obj) {
+ if (obj == this) {
+ return true;
+ }
+ if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations)) {
+ return super.equals(obj);
+ }
+ org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations other = (org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations) obj;
+
+ boolean result = true;
+ result = result && getLabelList()
+ .equals(other.getLabelList());
+ result = result &&
+ getUnknownFields().equals(other.getUnknownFields());
+ return result;
+ }
+
+ private int memoizedHashCode = 0;
+ @java.lang.Override
+ public int hashCode() {
+ if (memoizedHashCode != 0) {
+ return memoizedHashCode;
+ }
+ int hash = 41;
+ hash = (19 * hash) + getDescriptorForType().hashCode();
+ if (getLabelCount() > 0) {
+ hash = (37 * hash) + LABEL_FIELD_NUMBER;
+ hash = (53 * hash) + getLabelList().hashCode();
+ }
+ hash = (29 * hash) + getUnknownFields().hashCode();
+ memoizedHashCode = hash;
+ return hash;
+ }
+
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations parseFrom(
+ com.google.protobuf.ByteString data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations parseFrom(
+ com.google.protobuf.ByteString data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations parseFrom(byte[] data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations parseFrom(
+ byte[] data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations parseFrom(java.io.InputStream input)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations parseFrom(
+ java.io.InputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations parseDelimitedFrom(java.io.InputStream input)
+ throws java.io.IOException {
+ return PARSER.parseDelimitedFrom(input);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations parseDelimitedFrom(
+ java.io.InputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations parseFrom(
+ com.google.protobuf.CodedInputStream input)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations parseFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input, extensionRegistry);
+ }
+
+ public static Builder newBuilder() { return Builder.create(); }
+ public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations prototype) {
+ return newBuilder().mergeFrom(prototype);
+ }
+ public Builder toBuilder() { return newBuilder(this); }
+
+ @java.lang.Override
+ protected Builder newBuilderForType(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ Builder builder = new Builder(parent);
+ return builder;
+ }
+ /**
+ * Protobuf type {@code hbase.pb.Authorizations}
+ *
+ * <pre>
+ **
+ * The protocol buffer version of Authorizations.
+ * </pre>
+ */
+ public static final class Builder extends
+ com.google.protobuf.GeneratedMessage.Builder<Builder>
+ implements org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.AuthorizationsOrBuilder {
+ public static final com.google.protobuf.Descriptors.Descriptor
+ getDescriptor() {
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Authorizations_descriptor;
+ }
+
+ protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ internalGetFieldAccessorTable() {
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Authorizations_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations.Builder.class);
+ }
+
+ // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations.newBuilder()
+ private Builder() {
+ maybeForceBuilderInitialization();
+ }
+
+ private Builder(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ super(parent);
+ maybeForceBuilderInitialization();
+ }
+ private void maybeForceBuilderInitialization() {
+ if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ }
+ }
+ private static Builder create() {
+ return new Builder();
+ }
+
+ public Builder clear() {
+ super.clear();
+ label_ = com.google.protobuf.LazyStringArrayList.EMPTY;
+ bitField0_ = (bitField0_ & ~0x00000001);
+ return this;
+ }
+
+ public Builder clone() {
+ return create().mergeFrom(buildPartial());
+ }
+
+ public com.google.protobuf.Descriptors.Descriptor
+ getDescriptorForType() {
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Authorizations_descriptor;
+ }
+
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations getDefaultInstanceForType() {
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations.getDefaultInstance();
+ }
+
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations build() {
+ org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations result = buildPartial();
+ if (!result.isInitialized()) {
+ throw newUninitializedMessageException(result);
+ }
+ return result;
+ }
+
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations buildPartial() {
+ org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations result = new org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations(this);
+ int from_bitField0_ = bitField0_;
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ label_ = new com.google.protobuf.UnmodifiableLazyStringList(
+ label_);
+ bitField0_ = (bitField0_ & ~0x00000001);
+ }
+ result.label_ = label_;
+ onBuilt();
+ return result;
+ }
+
+ public Builder mergeFrom(com.google.protobuf.Message other) {
+ if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations) {
+ return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations)other);
+ } else {
+ super.mergeFrom(other);
+ return this;
+ }
+ }
+
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations other) {
+ if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations.getDefaultInstance()) return this;
+ if (!other.label_.isEmpty()) {
+ if (label_.isEmpty()) {
+ label_ = other.label_;
+ bitField0_ = (bitField0_ & ~0x00000001);
+ } else {
+ ensureLabelIsMutable();
+ label_.addAll(other.label_);
+ }
+ onChanged();
+ }
+ this.mergeUnknownFields(other.getUnknownFields());
+ return this;
+ }
+
+ public final boolean isInitialized() {
+ return true;
+ }
+
+ public Builder mergeFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations parsedMessage = null;
+ try {
+ parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Authorizations) e.getUnfinishedMessage();
+ throw e;
+ } finally {
+ if (parsedMessage != null) {
+ mergeFrom(parsedMessage);
+ }
+ }
+ return this;
+ }
+ private int bitField0_;
+
+ // repeated string label = 1;
+ private com.google.protobuf.LazyStringList label_ = com.google.protobuf.LazyStringArrayList.EMPTY;
+ private void ensureLabelIsMutable() {
+ if (!((bitField0_ & 0x00000001) == 0x00000001)) {
+ label_ = new com.google.protobuf.LazyStringArrayList(label_);
+ bitField0_ |= 0x00000001;
+ }
+ }
+ /**
+ * <code>repeated string label = 1;</code>
+ */
+ public java.util.List<java.lang.String>
+ getLabelList() {
+ return java.util.Collections.unmodifiableList(label_);
+ }
+ /**
+ * <code>repeated string label = 1;</code>
+ */
+ public int getLabelCount() {
+ return label_.size();
+ }
+ /**
+ * <code>repeated string label = 1;</code>
+ */
+ public java.lang.String getLabel(int index) {
+ return label_.get(index);
+ }
+ /**
+ * <code>repeated string label = 1;</code>
+ */
+ public com.google.protobuf.ByteString
+ getLabelBytes(int index) {
+ return label_.getByteString(index);
+ }
+ /**
+ * <code>repeated string label = 1;</code>
+ */
+ public Builder setLabel(
+ int index, java.lang.String value) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ ensureLabelIsMutable();
+ label_.set(index, value);
+ onChanged();
+ return this;
+ }
+ /**
+ * <code>repeated string label = 1;</code>
+ */
+ public Builder addLabel(
+ java.lang.String value) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ ensureLabelIsMutable();
+ label_.add(value);
+ onChanged();
+ return this;
+ }
+ /**
+ * <code>repeated string label = 1;</code>
+ */
+ public Builder addAllLabel(
+ java.lang.Iterable<java.lang.String> values) {
+ ensureLabelIsMutable();
+ super.addAll(values, label_);
+ onChanged();
+ return this;
+ }
+ /**
+ * <code>repeated string label = 1;</code>
+ */
+ public Builder clearLabel() {
+ label_ = com.google.protobuf.LazyStringArrayList.EMPTY;
+ bitField0_ = (bitField0_ & ~0x00000001);
+ onChanged();
+ return this;
+ }
+ /**
+ * <code>repeated string label = 1;</code>
+ */
+ public Builder addLabelBytes(
+ com.google.protobuf.ByteString value) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ ensureLabelIsMutable();
+ label_.add(value);
+ onChanged();
+ return this;
+ }
+
+ // @@protoc_insertion_point(builder_scope:hbase.pb.Authorizations)
+ }
+
+ static {
+ defaultInstance = new Authorizations(true);
+ defaultInstance.initFields();
+ }
+
+ // @@protoc_insertion_point(class_scope:hbase.pb.Authorizations)
+ }
+
+ public interface CellVisibilityOrBuilder
+ extends com.google.protobuf.MessageOrBuilder {
+
+ // required string expression = 1;
+ /**
+ * <code>required string expression = 1;</code>
+ */
+ boolean hasExpression();
+ /**
+ * <code>required string expression = 1;</code>
+ */
+ java.lang.String getExpression();
+ /**
+ * <code>required string expression = 1;</code>
+ */
+ com.google.protobuf.ByteString
+ getExpressionBytes();
+ }
+ /**
+ * Protobuf type {@code hbase.pb.CellVisibility}
+ *
+ * <pre>
+ **
+ * The protocol buffer version of CellVisibility.
+ * </pre>
+ */
+ public static final class CellVisibility extends
+ com.google.protobuf.GeneratedMessage
+ implements CellVisibilityOrBuilder {
+ // Use CellVisibility.newBuilder() to construct.
+ private CellVisibility(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ super(builder);
+ this.unknownFields = builder.getUnknownFields();
+ }
+ private CellVisibility(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
+ private static final CellVisibility defaultInstance;
+ public static CellVisibility getDefaultInstance() {
+ return defaultInstance;
+ }
+
+ public CellVisibility getDefaultInstanceForType() {
+ return defaultInstance;
+ }
+
+ private final com.google.protobuf.UnknownFieldSet unknownFields;
+ @java.lang.Override
+ public final com.google.protobuf.UnknownFieldSet
+ getUnknownFields() {
+ return this.unknownFields;
+ }
+ private CellVisibility(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ initFields();
+ int mutable_bitField0_ = 0;
+ com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ com.google.protobuf.UnknownFieldSet.newBuilder();
+ try {
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ default: {
+ if (!parseUnknownField(input, unknownFields,
+ extensionRegistry, tag)) {
+ done = true;
+ }
+ break;
+ }
+ case 10: {
+ bitField0_ |= 0x00000001;
+ expression_ = input.readBytes();
+ break;
+ }
+ }
+ }
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(this);
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(
+ e.getMessage()).setUnfinishedMessage(this);
+ } finally {
+ this.unknownFields = unknownFields.build();
+ makeExtensionsImmutable();
+ }
+ }
+ public static final com.google.protobuf.Descriptors.Descriptor
+ getDescriptor() {
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_CellVisibility_descriptor;
+ }
+
+ protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ internalGetFieldAccessorTable() {
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_CellVisibility_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility.Builder.class);
+ }
+
+ public static com.google.protobuf.Parser<CellVisibility> PARSER =
+ new com.google.protobuf.AbstractParser<CellVisibility>() {
+ public CellVisibility parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new CellVisibility(input, extensionRegistry);
+ }
+ };
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<CellVisibility> getParserForType() {
+ return PARSER;
+ }
+
+ private int bitField0_;
+ // required string expression = 1;
+ public static final int EXPRESSION_FIELD_NUMBER = 1;
+ private java.lang.Object expression_;
+ /**
+ * <code>required string expression = 1;</code>
+ */
+ public boolean hasExpression() {
+ return ((bitField0_ & 0x00000001) == 0x00000001);
+ }
+ /**
+ * <code>required string expression = 1;</code>
+ */
+ public java.lang.String getExpression() {
+ java.lang.Object ref = expression_;
+ if (ref instanceof java.lang.String) {
+ return (java.lang.String) ref;
+ } else {
+ com.google.protobuf.ByteString bs =
+ (com.google.protobuf.ByteString) ref;
+ java.lang.String s = bs.toStringUtf8();
+ if (bs.isValidUtf8()) {
+ expression_ = s;
+ }
+ return s;
+ }
+ }
+ /**
+ * <code>required string expression = 1;</code>
+ */
+ public com.google.protobuf.ByteString
+ getExpressionBytes() {
+ java.lang.Object ref = expression_;
+ if (ref instanceof java.lang.String) {
+ com.google.protobuf.ByteString b =
+ com.google.protobuf.ByteString.copyFromUtf8(
+ (java.lang.String) ref);
+ expression_ = b;
+ return b;
+ } else {
+ return (com.google.protobuf.ByteString) ref;
+ }
+ }
+
+ private void initFields() {
+ expression_ = "";
+ }
+ private byte memoizedIsInitialized = -1;
+ public final boolean isInitialized() {
+ byte isInitialized = memoizedIsInitialized;
+ if (isInitialized != -1) return isInitialized == 1;
+
+ if (!hasExpression()) {
+ memoizedIsInitialized = 0;
+ return false;
+ }
+ memoizedIsInitialized = 1;
+ return true;
+ }
+
+ public void writeTo(com.google.protobuf.CodedOutputStream output)
+ throws java.io.IOException {
+ getSerializedSize();
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ output.writeBytes(1, getExpressionBytes());
+ }
+ getUnknownFields().writeTo(output);
+ }
+
+ private int memoizedSerializedSize = -1;
+ public int getSerializedSize() {
+ int size = memoizedSerializedSize;
+ if (size != -1) return size;
+
+ size = 0;
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ size += com.google.protobuf.CodedOutputStream
+ .computeBytesSize(1, getExpressionBytes());
+ }
+ size += getUnknownFields().getSerializedSize();
+ memoizedSerializedSize = size;
+ return size;
+ }
+
+ private static final long serialVersionUID = 0L;
+ @java.lang.Override
+ protected java.lang.Object writeReplace()
+ throws java.io.ObjectStreamException {
+ return super.writeReplace();
+ }
+
+ @java.lang.Override
+ public boolean equals(final java.lang.Object obj) {
+ if (obj == this) {
+ return true;
+ }
+ if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility)) {
+ return super.equals(obj);
+ }
+ org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility other = (org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility) obj;
+
+ boolean result = true;
+ result = result && (hasExpression() == other.hasExpression());
+ if (hasExpression()) {
+ result = result && getExpression()
+ .equals(other.getExpression());
+ }
+ result = result &&
+ getUnknownFields().equals(other.getUnknownFields());
+ return result;
+ }
+
+ private int memoizedHashCode = 0;
+ @java.lang.Override
+ public int hashCode() {
+ if (memoizedHashCode != 0) {
+ return memoizedHashCode;
+ }
+ int hash = 41;
+ hash = (19 * hash) + getDescriptorForType().hashCode();
+ if (hasExpression()) {
+ hash = (37 * hash) + EXPRESSION_FIELD_NUMBER;
+ hash = (53 * hash) + getExpression().hashCode();
+ }
+ hash = (29 * hash) + getUnknownFields().hashCode();
+ memoizedHashCode = hash;
+ return hash;
+ }
+
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility parseFrom(
+ com.google.protobuf.ByteString data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility parseFrom(
+ com.google.protobuf.ByteString data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility parseFrom(byte[] data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility parseFrom(
+ byte[] data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility parseFrom(java.io.InputStream input)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility parseFrom(
+ java.io.InputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility parseDelimitedFrom(java.io.InputStream input)
+ throws java.io.IOException {
+ return PARSER.parseDelimitedFrom(input);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility parseDelimitedFrom(
+ java.io.InputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility parseFrom(
+ com.google.protobuf.CodedInputStream input)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility parseFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input, extensionRegistry);
+ }
+
+ public static Builder newBuilder() { return Builder.create(); }
+ public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility prototype) {
+ return newBuilder().mergeFrom(prototype);
+ }
+ public Builder toBuilder() { return newBuilder(this); }
+
+ @java.lang.Override
+ protected Builder newBuilderForType(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ Builder builder = new Builder(parent);
+ return builder;
+ }
+ /**
+ * Protobuf type {@code hbase.pb.CellVisibility}
+ *
+ * <pre>
+ **
+ * The protocol buffer version of CellVisibility.
+ * </pre>
+ */
+ public static final class Builder extends
+ com.google.protobuf.GeneratedMessage.Builder<Builder>
+ implements org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibilityOrBuilder {
+ public static final com.google.protobuf.Descriptors.Descriptor
+ getDescriptor() {
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_CellVisibility_descriptor;
+ }
+
+ protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ internalGetFieldAccessorTable() {
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_CellVisibility_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility.Builder.class);
+ }
+
+ // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility.newBuilder()
+ private Builder() {
+ maybeForceBuilderInitialization();
+ }
+
+ private Builder(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ super(parent);
+ maybeForceBuilderInitialization();
+ }
+ private void maybeForceBuilderInitialization() {
+ if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ }
+ }
+ private static Builder create() {
+ return new Builder();
+ }
+
+ public Builder clear() {
+ super.clear();
+ expression_ = "";
+ bitField0_ = (bitField0_ & ~0x00000001);
+ return this;
+ }
+
+ public Builder clone() {
+ return create().mergeFrom(buildPartial());
+ }
+
+ public com.google.protobuf.Descriptors.Descriptor
+ getDescriptorForType() {
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_CellVisibility_descriptor;
+ }
+
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility getDefaultInstanceForType() {
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility.getDefaultInstance();
+ }
+
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility build() {
+ org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility result = buildPartial();
+ if (!result.isInitialized()) {
+ throw newUninitializedMessageException(result);
+ }
+ return result;
+ }
+
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility buildPartial() {
+ org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility result = new org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility(this);
+ int from_bitField0_ = bitField0_;
+ int to_bitField0_ = 0;
+ if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
+ to_bitField0_ |= 0x00000001;
+ }
+ result.expression_ = expression_;
+ result.bitField0_ = to_bitField0_;
+ onBuilt();
+ return result;
+ }
+
+ public Builder mergeFrom(com.google.protobuf.Message other) {
+ if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility) {
+ return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility)other);
+ } else {
+ super.mergeFrom(other);
+ return this;
+ }
+ }
+
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility other) {
+ if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility.getDefaultInstance()) return this;
+ if (other.hasExpression()) {
+ bitField0_ |= 0x00000001;
+ expression_ = other.expression_;
+ onChanged();
+ }
+ this.mergeUnknownFields(other.getUnknownFields());
+ return this;
+ }
+
+ public final boolean isInitialized() {
+ if (!hasExpression()) {
+
+ return false;
+ }
+ return true;
+ }
+
+ public Builder mergeFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility parsedMessage = null;
+ try {
+ parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CellVisibility) e.getUnfinishedMessage();
+ throw e;
+ } finally {
+ if (parsedMessage != null) {
+ mergeFrom(parsedMessage);
+ }
+ }
+ return this;
+ }
+ private int bitField0_;
+
+ // required string expression = 1;
+ private java.lang.Object expression_ = "";
+ /**
+ * <code>required string expression = 1;</code>
+ */
+ public boolean hasExpression() {
+ return ((bitField0_ & 0x00000001) == 0x00000001);
+ }
+ /**
+ * <code>required string expression = 1;</code>
+ */
+ public java.lang.String getExpression() {
+ java.lang.Object ref = expression_;
+ if (!(ref instanceof java.lang.String)) {
+ java.lang.String s = ((com.google.protobuf.ByteString) ref)
+ .toStringUtf8();
+ expression_ = s;
+ return s;
+ } else {
+ return (java.lang.String) ref;
+ }
+ }
+ /**
+ * <code>required string expression = 1;</code>
+ */
+ public com.google.protobuf.ByteString
+ getExpressionBytes() {
+ java.lang.Object ref = expression_;
+ if (ref instanceof String) {
+ com.google.protobuf.ByteString b =
+ com.google.protobuf.ByteString.copyFromUtf8(
+ (java.lang.String) ref);
+ expression_ = b;
+ return b;
+ } else {
+ return (com.google.protobuf.ByteString) ref;
+ }
+ }
+ /**
+ * <code>required string expression = 1;</code>
+ */
+ public Builder setExpression(
+ java.lang.String value) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ bitField0_ |= 0x00000001;
+ expression_ = value;
+ onChanged();
+ return this;
+ }
+ /**
+ * <code>required string expression = 1;</code>
+ */
+ public Builder clearExpression() {
+ bitField0_ = (bitField0_ & ~0x00000001);
+ expression_ = getDefaultInstance().getExpression();
+ onChanged();
+ return this;
+ }
+ /**
+ * <code>required string expression = 1;</code>
+ */
+ public Builder setExpressionBytes(
+ com.google.protobuf.ByteString value) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ bitField0_ |= 0x00000001;
+ expression_ = value;
+ onChanged();
+ return this;
+ }
+
+ // @@protoc_insertion_point(builder_scope:hbase.pb.CellVisibility)
+ }
+
+ static {
+ defaultInstance = new CellVisibility(true);
+ defaultInstance.initFields();
+ }
+
+ // @@protoc_insertion_point(class_scope:hbase.pb.CellVisibility)
+ }
+
+ public interface ColumnOrBuilder
+ extends com.google.protobuf.MessageOrBuilder {
+
+ // required bytes family = 1;
+ /**
+ * <code>required bytes family = 1;</code>
+ */
+ boolean hasFamily();
+ /**
+ * <code>required bytes family = 1;</code>
+ */
+ com.google.protobuf.ByteString getFamily();
+
+ // repeated bytes qualifier = 2;
+ /**
+ * <code>repeated bytes qualifier = 2;</code>
+ */
+ java.util.List<com.google.protobuf.ByteString> getQualifierList();
+ /**
+ * <code>repeated bytes qualifier = 2;</code>
+ */
+ int getQualifierCount();
+ /**
+ * <code>repeated bytes qualifier = 2;</code>
+ */
+ com.google.protobuf.ByteString getQualifier(int index);
+ }
+ /**
+ * Protobuf type {@code hbase.pb.Column}
+ *
+ * <pre>
+ **
+ * Container for a list of column qualifier names of a family.
+ * </pre>
+ */
+ public static final class Column extends
+ com.google.protobuf.GeneratedMessage
+ implements ColumnOrBuilder {
+ // Use Column.newBuilder() to construct.
+ private Column(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ super(builder);
+ this.unknownFields = builder.getUnknownFields();
+ }
+ private Column(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
+ private static final Column defaultInstance;
+ public static Column getDefaultInstance() {
+ return defaultInstance;
+ }
+
+ public Column getDefaultInstanceForType() {
+ return defaultInstance;
+ }
+
+ private final com.google.protobuf.UnknownFieldSet unknownFields;
+ @java.lang.Override
+ public final com.google.protobuf.UnknownFieldSet
+ getUnknownFields() {
+ return this.unknownFields;
+ }
+ private Column(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ initFields();
+ int mutable_bitField0_ = 0;
+ com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ com.google.protobuf.UnknownFieldSet.newBuilder();
+ try {
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ default: {
+ if (!parseUnknownField(input, unknownFields,
+ extensionRegistry, tag)) {
+ done = true;
+ }
+ break;
+ }
+ case 10: {
+ bitField0_ |= 0x00000001;
+ family_ = input.readBytes();
+ break;
+ }
+ case 18: {
+ if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
+ qualifier_ = new java.util.ArrayList<com.google.protobuf.ByteString>();
+ mutable_bitField0_ |= 0x00000002;
+ }
+ qualifier_.add(input.readBytes());
+ break;
+ }
+ }
+ }
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(this);
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(
+ e.getMessage()).setUnfinishedMessage(this);
+ } finally {
+ if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
+ qualifier_ = java.util.Collections.unmodifiableList(qualifier_);
+ }
+ this.unknownFields = unknownFields.build();
+ makeExtensionsImmutable();
+ }
+ }
+ public static final com.google.protobuf.Descriptors.Descriptor
+ getDescriptor() {
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Column_descriptor;
+ }
+
+ protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ internalGetFieldAccessorTable() {
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Column_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column.Builder.class);
+ }
+
+ public static com.google.protobuf.Parser<Column> PARSER =
+ new com.google.protobuf.AbstractParser<Column>() {
+ public Column parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new Column(input, extensionRegistry);
+ }
+ };
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<Column> getParserForType() {
+ return PARSER;
+ }
+
+ private int bitField0_;
+ // required bytes family = 1;
+ public static final int FAMILY_FIELD_NUMBER = 1;
+ private com.google.protobuf.ByteString family_;
+ /**
+ * <code>required bytes family = 1;</code>
+ */
+ public boolean hasFamily() {
+ return ((bitField0_ & 0x00000001) == 0x00000001);
+ }
+ /**
+ * <code>required bytes family = 1;</code>
+ */
+ public com.google.protobuf.ByteString getFamily() {
+ return family_;
+ }
+
+ // repeated bytes qualifier = 2;
+ public static final int QUALIFIER_FIELD_NUMBER = 2;
+ private java.util.List<com.google.protobuf.ByteString> qualifier_;
+ /**
+ * <code>repeated bytes qualifier = 2;</code>
+ */
+ public java.util.List<com.google.protobuf.ByteString>
+ getQualifierList() {
+ return qualifier_;
+ }
+ /**
+ * <code>repeated bytes qualifier = 2;</code>
+ */
+ public int getQualifierCount() {
+ return qualifier_.size();
+ }
+ /**
+ * <code>repeated bytes qualifier = 2;</code>
+ */
+ public com.google.protobuf.ByteString getQualifier(int index) {
+ return qualifier_.get(index);
+ }
+
+ private void initFields() {
+ family_ = com.google.protobuf.ByteString.EMPTY;
+ qualifier_ = java.util.Collections.emptyList();
+ }
+ private byte memoizedIsInitialized = -1;
+ public final boolean isInitialized() {
+ byte isInitialized = memoizedIsInitialized;
+ if (isInitialized != -1) return isInitialized == 1;
+
+ if (!hasFamily()) {
+ memoizedIsInitialized = 0;
+ return false;
+ }
+ memoizedIsInitialized = 1;
+ return true;
+ }
+
+ public void writeTo(com.google.protobuf.CodedOutputStream output)
+ throws java.io.IOException {
+ getSerializedSize();
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ output.writeBytes(1, family_);
+ }
+ for (int i = 0; i < qualifier_.size(); i++) {
+ output.writeBytes(2, qualifier_.get(i));
+ }
+ getUnknownFields().writeTo(output);
+ }
+
+ private int memoizedSerializedSize = -1;
+ public int getSerializedSize() {
+ int size = memoizedSerializedSize;
+ if (size != -1) return size;
+
+ size = 0;
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ size += com.google.protobuf.CodedOutputStream
+ .computeBytesSize(1, family_);
+ }
+ {
+ int dataSize = 0;
+ for (int i = 0; i < qualifier_.size(); i++) {
+ dataSize += com.google.protobuf.CodedOutputStream
+ .computeBytesSizeNoTag(qualifier_.get(i));
+ }
+ size += dataSize;
+ size += 1 * getQualifierList().size();
+ }
+ size += getUnknownFields().getSerializedSize();
+ memoizedSerializedSize = size;
+ return size;
+ }
+
+ private static final long serialVersionUID = 0L;
+ @java.lang.Override
+ protected java.lang.Object writeReplace()
+ throws java.io.ObjectStreamException {
+ return super.writeReplace();
+ }
+
+ @java.lang.Override
+ public boolean equals(final java.lang.Object obj) {
+ if (obj == this) {
+ return true;
+ }
+ if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column)) {
+ return super.equals(obj);
+ }
+ org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column other = (org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column) obj;
+
+ boolean result = true;
+ result = result && (hasFamily() == other.hasFamily());
+ if (hasFamily()) {
+ result = result && getFamily()
+ .equals(other.getFamily());
+ }
+ result = result && getQualifierList()
+ .equals(other.getQualifierList());
+ result = result &&
+ getUnknownFields().equals(other.getUnknownFields());
+ return result;
+ }
+
+ private int memoizedHashCode = 0;
+ @java.lang.Override
+ public int hashCode() {
+ if (memoizedHashCode != 0) {
+ return memoizedHashCode;
+ }
+ int hash = 41;
+ hash = (19 * hash) + getDescriptorForType().hashCode();
+ if (hasFamily()) {
+ hash = (37 * hash) + FAMILY_FIELD_NUMBER;
+ hash = (53 * hash) + getFamily().hashCode();
+ }
+ if (getQualifierCount() > 0) {
+ hash = (37 * hash) + QUALIFIER_FIELD_NUMBER;
+ hash = (53 * hash) + getQualifierList().hashCode();
+ }
+ hash = (29 * hash) + getUnknownFields().hashCode();
+ memoizedHashCode = hash;
+ return hash;
+ }
+
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column parseFrom(
+ com.google.protobuf.ByteString data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column parseFrom(
+ com.google.protobuf.ByteString data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column parseFrom(byte[] data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column parseFrom(
+ byte[] data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column parseFrom(java.io.InputStream input)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column parseFrom(
+ java.io.InputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column parseDelimitedFrom(java.io.InputStream input)
+ throws java.io.IOException {
+ return PARSER.parseDelimitedFrom(input);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column parseDelimitedFrom(
+ java.io.InputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column parseFrom(
+ com.google.protobuf.CodedInputStream input)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input);
+ }
+ public static org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column parseFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input, extensionRegistry);
+ }
+
+ public static Builder newBuilder() { return Builder.create(); }
+ public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column prototype) {
+ return newBuilder().mergeFrom(prototype);
+ }
+ public Builder toBuilder() { return newBuilder(this); }
+
+ @java.lang.Override
+ protected Builder newBuilderForType(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ Builder builder = new Builder(parent);
+ return builder;
+ }
+ /**
+ * Protobuf type {@code hbase.pb.Column}
+ *
+ * <pre>
+ **
+ * Container for a list of column qualifier names of a family.
+ * </pre>
+ */
+ public static final class Builder extends
+ com.google.protobuf.GeneratedMessage.Builder<Builder>
+ implements org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ColumnOrBuilder {
+ public static final com.google.protobuf.Descriptors.Descriptor
+ getDescriptor() {
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Column_descriptor;
+ }
+
+ protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ internalGetFieldAccessorTable() {
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Column_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column.Builder.class);
+ }
+
+ // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column.newBuilder()
+ private Builder() {
+ maybeForceBuilderInitialization();
+ }
+
+ private Builder(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ super(parent);
+ maybeForceBuilderInitialization();
+ }
+ private void maybeForceBuilderInitialization() {
+ if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ }
+ }
+ private static Builder create() {
+ return new Builder();
+ }
+
+ public Builder clear() {
+ super.clear();
+ family_ = com.google.protobuf.ByteString.EMPTY;
+ bitField0_ = (bitField0_ & ~0x00000001);
+ qualifier_ = java.util.Collections.emptyList();
+ bitField0_ = (bitField0_ & ~0x00000002);
+ return this;
+ }
+
+ public Builder clone() {
+ return create().mergeFrom(buildPartial());
+ }
+
+ public com.google.protobuf.Descriptors.Descriptor
+ getDescriptorForType() {
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Column_descriptor;
+ }
+
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column getDefaultInstanceForType() {
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column.getDefaultInstance();
+ }
+
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column build() {
+ org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column result = buildPartial();
+ if (!result.isInitialized()) {
+ throw newUninitializedMessageException(result);
+ }
+ return result;
+ }
+
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column buildPartial() {
+ org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column result = new org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column(this);
+ int from_bitField0_ = bitField0_;
+ int to_bitField0_ = 0;
+ if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
+ to_bitField0_ |= 0x00000001;
+ }
+ result.family_ = family_;
+ if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ qualifier_ = java.util.Collections.unmodifiableList(qualifier_);
+ bitField0_ = (bitField0_ & ~0x00000002);
+ }
+ result.qualifier_ = qualifier_;
+ result.bitField0_ = to_bitField0_;
+ onBuilt();
+ return result;
+ }
+
+ public Builder mergeFrom(com.google.protobuf.Message other) {
+ if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column) {
+ return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column)other);
+ } else {
+ super.mergeFrom(other);
+ return this;
+ }
+ }
+
+ public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column other) {
+ if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column.getDefaultInstance()) return this;
+ if (other.hasFamily()) {
+ setFamily(other.getFamily());
+ }
+ if (!other.qualifier_.isEmpty()) {
+ if (qualifier_.isEmpty()) {
+ qualifier_ = other.qualifier_;
+ bitField0_ = (bitField0_ & ~0x00000002);
+ } else {
+ ensureQualifierIsMutable();
+ qualifier_.addAll(other.qualifier_);
+ }
+ onChanged();
+ }
+ this.mergeUnknownFields(other.getUnknownFields());
+ return this;
+ }
+
+ public final boolean isInitialized() {
+ if (!hasFamily()) {
+
+ return false;
+ }
+ return true;
+ }
+
+ public Builder mergeFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column parsedMessage = null;
+ try {
+ parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column) e.getUnfinishedMessage();
+ throw e;
+ } finally {
+ if (parsedMessage != null) {
+ mergeFrom(parsedMessage);
+ }
+ }
+ return this;
+ }
+ private int bitField0_;
+
+ // required bytes family = 1;
+ private com.google.protobuf.ByteString family_ = com.google.protobuf.ByteString.EMPTY;
+ /**
+ * <code>required bytes family = 1;</code>
+ */
+ public boolean hasFamily() {
+ return ((bitField0_ & 0x00000001) == 0x00000001);
+ }
+ /**
+ * <code>required bytes family = 1;</code>
+ */
+ public com.google.protobuf.ByteString getFamily() {
+ return family_;
+ }
+ /**
+ * <code>required bytes family = 1;</code>
+ */
+ public Builder setFamily(com.google.protobuf.ByteString value) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ bitField0_ |= 0x00000001;
+ family_ = value;
+ onChanged();
+ return this;
+ }
+ /**
+ * <code>required bytes family = 1;</code>
+ */
+ public Builder clearFamily() {
+ bitField0_ = (bitField0_ & ~0x00000001);
+ family_ = getDefaultInstance().getFamily();
+ onChanged();
+ return this;
+ }
+
+ // repeated bytes qualifier = 2;
+ private java.util.List<com.google.protobuf.ByteString> qualifier_ = java.util.Collections.emptyList();
+ private void ensureQualifierIsMutable() {
+ if (!((bitField0_ & 0x00000002) == 0x00000002)) {
+ qualifier_ = new java.util.ArrayList<com.google.protobuf.ByteString>(qualifier_);
+ bitField0_ |= 0x00000002;
+ }
+ }
+ /**
+ * <code>repeated bytes qualifier = 2;</code>
+ */
+ public java.util.List<com.google.protobuf.ByteString>
+ getQualifierList() {
+ return java.util.Collections.unmodifiableList(qualifier_);
+ }
+ /**
+ * <code>repeated bytes qualifier = 2;</code>
+ */
+ public int getQualifierCount() {
+ return qualifier_.size();
+ }
+ /**
+ * <code>repeated bytes qualifier = 2;</code>
+ */
+ public com.google.protobuf.ByteString getQualifier(int index) {
+ return qualifier_.get(index);
+ }
+ /**
+ * <code>repeated bytes qualifier = 2;</code>
+ */
+ public Builder setQualifier(
+ int index, com.google.protobuf.ByteString value) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ ensureQualifierIsMutable();
+ qualifier_.set(index, value);
+ onChanged();
+ return this;
+ }
+ /**
+ * <code>repeated bytes qualifier = 2;</code>
+ */
+ public Builder addQualifier(com.google.protobuf.ByteString value) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ ensureQualifierIsMutable();
+ qualifier_.add(value);
+ onChanged();
+ return this;
+ }
+ /**
+ * <code>repeated bytes qualifier = 2;</code>
+ */
+ public Builder addAllQualifier(
+ java.lang.Iterable<? extends com.google.protobuf.ByteString> values) {
+ ensureQualifierIsMutable();
+ super.addAll(values, qualifier_);
+ onChanged();
+ return this;
+ }
+ /**
+ * <code>repeated bytes qualifier = 2;</code>
+ */
+ public Builder clearQualifier() {
+ qualifier_ = java.util.Collections.emptyList();
+ bitField0_ = (bitField0_ & ~0x00000002);
+ onChanged();
+ return this;
+ }
+
+ // @@protoc_insertion_point(builder_scope:hbase.pb.Column)
+ }
+
+ static {
+ defaultInstance = new Column(true);
+ defaultInstance.initFields();
+ }
+
+ // @@protoc_insertion_point(class_scope:hbase.pb.Column)
+ }
+
+ public interface GetOrBuilder
+ extends com.google.protobuf.MessageOrBuilder {
+
+ // required bytes row = 1;
+ /**
+ * <code>required bytes row = 1;</code>
+ */
+ boolean hasRow();
+ /**
+ * <code>required bytes row = 1;</code>
+ */
+ com.google.protobuf.ByteString getRow();
+
+ // repeated .hbase.pb.Column column = 2;
+ /**
+ * <code>repeated .hbase.pb.Column column = 2;</code>
+ */
+ java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column>
+ getColumnList();
+ /**
+ * <code>repeated .hbase.pb.Column column = 2;</code>
+ */
+ org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column getColumn(int index);
+ /**
+ * <code>repeated .hbase.pb.Column column = 2;</code>
+ */
+ int getColumnCount();
+ /**
+ * <code>repeated .hbase.pb.Column column = 2;</code>
+ */
+ java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ColumnOrBuilder>
+ getColumnOrBuilderList();
+ /**
+ * <code>repeated .hbase.pb.Column column = 2;</code>
+ */
+ org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ColumnOrBuilder getColumnOrBuilder(
+ int index);
+
+ // repeated .hbase.pb.NameBytesPair attribute = 3;
+ /**
+ * <code>repeated .hbase.pb.NameBytesPair attribute = 3;</code>
+ */
+ java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameBytesPair>
+ getAttributeList();
+ /**
+ * <code>repeated .hbase.pb.NameBytesPair attribute = 3;</code>
+ */
+ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameBytesPair getAttribute(int index);
+ /**
+ * <code>repeated .hbase.pb.NameBytesPair attribute = 3;</code>
+ */
+ int getAttributeCount();
+ /**
+ * <code>repeated .hbase.pb.NameBytesPair attribute = 3;</code>
+ */
+ java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameBytesPairOrBuilder>
+ getAttributeOrBuilderList();
+ /**
+ * <code>repeated .hbase.pb.NameBytesPair attribute = 3;</code>
+ */
+ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameBytesPairOrBuilder getAttributeOrBuilder(
+ int index);
+
+ // optional .hbase.pb.Filter filter = 4;
+ /**
+ * <code>optional .hbase.pb.Filter filter = 4;</code>
+ */
+ boolean hasFilter();
+ /**
+ * <code>optional .hbase.pb.Filter filter = 4;</code>
+ */
+ org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter getFilter();
+ /**
+ * <code>optional .hbase.pb.Filter filter = 4;</code>
+ */
+ org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.FilterOrBuilder getFilterOrBuilder();
+
+ // optional .hbase.pb.TimeRange time_range = 5;
+ /**
+ * <code>optional .hbase.pb.TimeRange time_range = 5;</code>
+ */
+ boolean hasTimeRange();
+ /**
+ * <code>optional .hbase.pb.TimeRange time_range = 5;</code>
+ */
+ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TimeRange getTimeRange();
+ /**
+ * <code>optional .hbase.pb.TimeRange time_range = 5;</code>
+ */
+ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TimeRangeOrBuilder getTimeRangeOrBuilder();
+
+ // optional uint32 max_versions = 6 [default = 1];
+ /**
+ * <code>optional uint32 max_versions = 6 [default = 1];</code>
+ */
+ boolean hasMaxVersions();
+ /**
+ * <code>optional uint32 max_versions = 6 [default = 1];</code>
+ */
+ int getMaxVersions();
+
+ // optional bool cache_blocks = 7 [default = true];
+ /**
+ * <code>optional bool cache_blocks = 7 [default = true];</code>
+ */
+ boolean hasCacheBlocks();
+ /**
+ * <code>optional bool cache_blocks = 7 [default = true];</code>
+ */
+ boolean getCacheBlocks();
+
+ // optional uint32 store_limit = 8;
+ /**
+ * <code>optional uint32 store_limit = 8;</code>
+ */
+ boolean hasStoreLimit();
+ /**
+ * <code>optional uint32 store_limit = 8;</code>
+ */
+ int getStoreLimit();
+
+ // optional uint32 store_offset = 9;
+ /**
+ * <code>optional uint32 store_offset = 9;</code>
+ */
+ boolean hasStoreOffset();
+ /**
+ * <code>optional uint32 store_offset = 9;</code>
+ */
+ int getStoreOffset();
+
+ // optional bool existence_only = 10 [default = false];
+ /**
+ * <code>optional bool existence_only = 10 [default = false];</code>
+ *
+ * <pre>
+ * The result isn't asked for, just check for
+ * the existence.
+ * </pre>
+ */
+ boolean hasExistenceOnly();
+ /**
+ * <code>optional bool existence_only = 10 [default = false];</code>
+ *
+ * <pre>
+ * The result isn't asked for, just check for
+ * the existence.
+ * </pre>
+ */
+ boolean getExistenceOnly();
+
+ // optional .hbase.pb.Consistency consistency = 12 [default = STRONG];
+ /**
+ * <code>optional .hbase.pb.Consistency consistency = 12 [default = STRONG];</code>
+ */
+ boolean hasConsistency();
+ /**
+ * <code>optional .hbase.pb.Consistency consistency = 12 [default = STRONG];</code>
+ */
+ org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Consistency getConsistency();
+
+ // repeated .hbase.pb.ColumnFamilyTimeRange cf_time_range = 13;
+ /**
+ * <code>repeated .hbase.pb.ColumnFamilyTimeRange cf_time_range = 13;</code>
+ */
+ java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilyTimeRange>
+ getCfTimeRangeList();
+ /**
+ * <code>repeated .hbase.pb.ColumnFamilyTimeRange cf_time_range = 13;</code>
+ */
+ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilyTimeRange getCfTimeRange(int index);
+ /**
+ * <code>repeated .hbase.pb.ColumnFamilyTimeRange cf_time_range = 13;</code>
+ */
+ int getCfTimeRangeCount();
+ /**
+ * <code>repeated .hbase.pb.ColumnFamilyTimeRange cf_time_range = 13;</code>
+ */
+ java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilyTimeRangeOrBuilder>
+ getCfTimeRangeOrBuilderList();
+ /**
+ * <code>repeated .hbase.pb.ColumnFamilyTimeRange cf_time_range = 13;</code>
+ */
+ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilyTimeRangeOrBuilder getCfTimeRangeOrBuilder(
+ int index);
+ }
+ /**
+ * Protobuf type {@code hbase.pb.Get}
+ *
+ * <pre>
+ **
+ * The protocol buffer version of Get.
+ * Unless existence_only is specified, return all the requested data
+ * for the row that matches exactly.
+ * </pre>
+ */
+ public static final class Get extends
+ com.google.protobuf.GeneratedMessage
+ implements GetOrBuilder {
+ // Use Get.newBuilder() to construct.
+ private Get(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ super(builder);
+ this.unknownFields = builder.getUnknownFields();
+ }
+ private Get(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
+ private static final Get defaultInstance;
+ public static Get getDefaultInstance() {
+ return defaultInstance;
+ }
+
+ public Get getDefaultInstanceForType() {
+ return defaultInstance;
+ }
+
+ private final com.google.protobuf.UnknownFieldSet unknownFields;
+ @java.lang.Override
+ public final com.google.protobuf.UnknownFieldSet
+ getUnknownFields() {
+ return this.unknownFields;
+ }
+ private Get(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ initFields();
+ int mutable_bitField0_ = 0;
+ com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ com.google.protobuf.UnknownFieldSet.newBuilder();
+ try {
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ default: {
+ if (!parseUnknownField(input, unknownFields,
+ extensionRegistry, tag)) {
+ done = true;
+ }
+ break;
+ }
+ case 10: {
+ bitField0_ |= 0x00000001;
+ row_ = input.readBytes();
+ break;
+ }
+ case 18: {
+ if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
+ column_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column>();
+ mutable_bitField0_ |= 0x00000002;
+ }
+ column_.add(input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column.PARSER, extensionRegistry));
+ break;
+ }
+ case 26: {
+ if (!((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
+ attribute_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameBytesPair>();
+ mutable_bitField0_ |= 0x00000004;
+ }
+ attribute_.add(input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameBytesPair.PARSER, extensionRegistry));
+ break;
+ }
+ case 34: {
+ org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter.Builder subBuilder = null;
+ if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ subBuilder = filter_.toBuilder();
+ }
+ filter_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter.PARSER, extensionRegistry);
+ if (subBuilder != null) {
+ subBuilder.mergeFrom(filter_);
+ filter_ = subBuilder.buildPartial();
+ }
+ bitField0_ |= 0x00000002;
+ break;
+ }
+ case 42: {
+ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TimeRange.Builder subBuilder = null;
+ if (((bitField0_ & 0x00000004) == 0x00000004)) {
+ subBuilder = timeRange_.toBuilder();
+ }
+ timeRange_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TimeRange.PARSER, extensionRegistry);
+ if (subBuilder != null) {
+ subBuilder.mergeFrom(timeRange_);
+ timeRange_ = subBuilder.buildPartial();
+ }
+ bitField0_ |= 0x00000004;
+ break;
+ }
+ case 48: {
+ bitField0_ |= 0x00000008;
+ maxVersions_ = input.readUInt32();
+ break;
+ }
+ case 56: {
+ bitField0_ |= 0x00000010;
+ cacheBlocks_ = input.readBool();
+ break;
+ }
+ case 64: {
+ bitField0_ |= 0x00000020;
+ storeLimit_ = input.readUInt32();
+ break;
+ }
+ case 72: {
+ bitField0_ |= 0x00000040;
+ storeOffset_ = input.readUInt32();
+ break;
+ }
+ case 80: {
+ bitField0_ |= 0x00000080;
+ existenceOnly_ = input.readBool();
+ break;
+ }
+ case 96: {
+ int rawValue = input.readEnum();
+ org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Consistency value = org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Consistency.valueOf(rawValue);
+ if (value == null) {
+ unknownFields.mergeVarintField(12, rawValue);
+ } else {
+ bitField0_ |= 0x00000100;
+ consistency_ = value;
+ }
+ break;
+ }
+ case 106: {
+ if (!((mutable_bitField0_ & 0x00000800) == 0x00000800)) {
+ cfTimeRange_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilyTimeRange>();
+ mutable_bitField0_ |= 0x00000800;
+ }
+ cfTimeRange_.add(input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilyTimeRange.PARSER, extensionRegistry));
+ break;
+ }
+ }
+ }
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(this);
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(
+ e.getMessage()).setUnfinishedMessage(this);
+ } finally {
+ if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
+ column_ = java.util.Collections.unmodifiableList(column_);
+ }
+ if (((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
+ attribute_ = java.util.Collections.unmodifiableList(attribute_);
+ }
+ if (((mutable_bitField0_ & 0x00000800) == 0x00000800)) {
+ cfTimeRange_ = java.util.Collections.unmodifiableList(cfTimeRange_);
+ }
+ this.unknownFields = unknownFields.build();
+ makeExtensionsImmutable();
+ }
+ }
+ public static final com.google.protobuf.Descriptors.Descriptor
+ getDescriptor() {
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Get_descriptor;
+ }
+
+ protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ internalGetFieldAccessorTable() {
+ return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.internal_static_hbase_pb_Get_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Get.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Get.Builder.class);
+ }
+
+ public static com.google.protobuf.Parser<Get> PARSER =
+ new com.google.protobuf.AbstractParser<Get>() {
+ public Get parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new Get(input, extensionRegistry);
+ }
+ };
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<Get> getParserForType() {
+ return PARSER;
+ }
+
+ private int bitField0_;
+ // required bytes row = 1;
+ public static final int ROW_FIELD_NUMBER = 1;
+ private com.google.protobuf.ByteString row_;
+ /**
+ * <code>required bytes row = 1;</code>
+ */
+ public boolean hasRow() {
+ return ((bitField0_ & 0x00000001) == 0x00000001);
+ }
+ /**
+ * <code>required bytes row = 1;</code>
+ */
+ public com.google.protobuf.ByteString getRow() {
+ return row_;
+ }
+
+ // repeated .hbase.pb.Column column = 2;
+ public static final int COLUMN_FIELD_NUMBER = 2;
+ private java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column> column_;
+ /**
+ * <code>repeated .hbase.pb.Column column = 2;</code>
+ */
+ public java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column> getColumnList() {
+ return column_;
+ }
+ /**
+ * <code>repeated .hbase.pb.Column column = 2;</code>
+ */
+ public java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ColumnOrBuilder>
+ getColumnOrBuilderList() {
+ return column_;
+ }
+ /**
+ * <code>repeated .hbase.pb.Column column = 2;</code>
+ */
+ public int getColumnCount() {
+ return column_.size();
+ }
+ /**
+ * <code>repeated .hbase.pb.Column column = 2;</code>
+ */
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Column getColumn(int index) {
+ return column_.get(index);
+ }
+ /**
+ * <code>repeated .hbase.pb.Column column = 2;</code>
+ */
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ColumnOrBuilder getColumnOrBuilder(
+ int index) {
+ return column_.get(index);
+ }
+
+ // repeated .hbase.pb.NameBytesPair attribute = 3;
+ public static final int ATTRIBUTE_FIELD_NUMBER = 3;
+ private java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameBytesPair> attribute_;
+ /**
+ * <code>repeated .hbase.pb.NameBytesPair attribute = 3;</code>
+ */
+ public java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameBytesPair> getAttributeList() {
+ return attribute_;
+ }
+ /**
+ * <code>repeated .hbase.pb.NameBytesPair attribute = 3;</code>
+ */
+ public java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameBytesPairOrBuilder>
+ getAttributeOrBuilderList() {
+ return attribute_;
+ }
+ /**
+ * <code>repeated .hbase.pb.NameBytesPair attribute = 3;</code>
+ */
+ public int getAttributeCount() {
+ return attribute_.size();
+ }
+ /**
+ * <code>repeated .hbase.pb.NameBytesPair attribute = 3;</code>
+ */
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameBytesPair getAttribute(int index) {
+ return attribute_.get(index);
+ }
+ /**
+ * <code>repeated .hbase.pb.NameBytesPair attribute = 3;</code>
+ */
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameBytesPairOrBuilder getAttributeOrBuilder(
+ int index) {
+ return attribute_.get(index);
+ }
+
+ // optional .hbase.pb.Filter filter = 4;
+ public static final int FILTER_FIELD_NUMBER = 4;
+ private org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter filter_;
+ /**
+ * <code>optional .hbase.pb.Filter filter = 4;</code>
+ */
+ public boolean hasFilter() {
+ return ((bitField0_ & 0x00000002) == 0x00000002);
+ }
+ /**
+ * <code>optional .hbase.pb.Filter filter = 4;</code>
+ */
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter getFilter() {
+ return filter_;
+ }
+ /**
+ * <code>optional .hbase.pb.Filter filter = 4;</code>
+ */
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.FilterOrBuilder getFilterOrBuilder() {
+ return filter_;
+ }
+
+ // optional .hbase.pb.TimeRange time_range = 5;
+ public static final int TIME_RANGE_FIELD_NUMBER = 5;
+ private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TimeRange timeRange_;
+ /**
+ * <code>optional .hbase.pb.TimeRange time_range = 5;</code>
+ */
+ public boolean hasTimeRange() {
+ return ((bitField0_ & 0x00000004) == 0x00000004);
+ }
+ /**
+ * <code>optional .hbase.pb.TimeRange time_range = 5;</code>
+ */
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TimeRange getTimeRange() {
+ return timeRange_;
+ }
+ /**
+ * <code>optional .hbase.pb.TimeRange time_range = 5;</code>
+ */
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TimeRangeOrBuilder getTimeRangeOrBuilder() {
+ return timeRange_;
+ }
+
+ // optional uint32 max_versions = 6 [default = 1];
+ public static final int MAX_VERSIONS_FIELD_NUMBER = 6;
+ private int maxVersions_;
+ /**
+ * <code>optional uint32 max_versions = 6 [default = 1];</code>
+ */
+ public boolean hasMaxVersions() {
+ return ((bitField0_ & 0x00000008) == 0x00000008);
+ }
+ /**
+ * <code>optional uint32 max_versions = 6 [default = 1];</code>
+ */
+ public int getMaxVersions() {
+ return maxVersions_;
+ }
+
+ // optional bool cache_blocks = 7 [default = true];
+ public static final int CACHE_BLOCKS_FIELD_NUMBER = 7;
+ private boolean cacheBlocks_;
+ /**
+ * <code>optional bool cache_blocks = 7 [default = true];</code>
+ */
+ public boolean hasCacheBlocks() {
+ return ((bitField0_ & 0x00000010) == 0x00000010);
+ }
+ /**
+ * <code>optional bool cache_blocks = 7 [default = true];</code>
+ */
+ public boolean getCacheBlocks() {
+ return cacheBlocks_;
+ }
+
+ // optional uint32 store_limit = 8;
+ public static final int STORE_LIMIT_FIELD_NUMBER = 8;
+ private int storeLimit_;
+ /**
+ * <code>optional uint32 store_limit = 8;</code>
+ */
+ public boolean hasStoreLimit() {
+ return ((bitField0_ & 0x00000020) == 0x00000020);
+ }
+ /**
+ * <code>optional uint32 store_limit = 8;</code>
+ */
+ public int getStoreLimit() {
+ return storeLimit_;
+ }
+
+ // optional uint32 store_offset = 9;
+ public static final int STORE_OFFSET_FIELD_NUMBER = 9;
+ private int storeOffset_;
+ /**
+ * <code>optional uint32 store_offset = 9;</code>
+ */
+ public boolean hasStoreOffset() {
+ return ((bitField0_ & 0x00000040) == 0x00000040);
+ }
+ /**
+ * <code>optional uint32 store_offset = 9;</code>
+ */
+ public int getStoreOffset() {
+ return storeOffset_;
+ }
+
+ // optional bool existence_only = 10 [default = false];
+ public static final int EXISTENCE_ONLY_FIELD_NUMBER = 10;
+ private boolean existenceOnly_;
+ /**
+ * <code>optional bool existence_only = 10 [default = false];</code>
+ *
+ * <pre>
+ * The result isn't asked for, just check for
+ * the existence.
+ * </pre>
+ */
+ public boolean hasExistenceOnly() {
+ return ((bitField0_ & 0x00000080) == 0x00000080);
+ }
+ /**
+ * <code>optional bool existence_only = 10 [default = false];</code>
+ *
+ * <pre>
+ * The result isn't asked for, just check for
+ * the existence.
+ * </pre>
+ */
+ public boolean getExistenceOnly() {
+ return existenceOnly_;
+ }
+
+ // optional .hbase.pb.Consistency consistency = 12 [default = STRONG];
+ public static final int CONSISTENCY_FIELD_NUMBER = 12;
+ private org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Consistency consistency_;
+ /**
+ * <code>optional .hbase.pb.Consistency consistency = 12 [default = STRONG];</code>
+ */
+ public boolean hasConsistency() {
+ return ((bitField0_ & 0x00000100) == 0x00000100);
+ }
+ /**
+ * <code>optional .hbase.pb.Consistency consistency = 12 [default = STRONG];</code>
+ */
+ public org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Consistency getConsistency() {
+ return consistency_;
+ }
+
+ // repeated .hbase.pb.ColumnFamilyTimeRange cf_time_range = 13;
+ public static final int CF_TIME_RANGE_FIELD_NUMBER = 13;
+ private java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilyTimeRange> cfTimeRange_;
+ /**
+ * <code>repeated .hbase.pb.ColumnFamilyTimeRange cf_time_range = 13;</c
<TRUNCATED>