You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2013/04/10 18:42:29 UTC
svn commit: r1466556 [2/41] - in /hbase/trunk: ./
hbase-client/src/main/java/org/apache/hadoop/hbase/
hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/
hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/
hbase-protocol/s...
Modified: hbase/trunk/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AccessControlProtos.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AccessControlProtos.java?rev=1466556&r1=1466555&r2=1466556&view=diff
==============================================================================
--- hbase/trunk/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AccessControlProtos.java (original)
+++ hbase/trunk/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AccessControlProtos.java Wed Apr 10 16:42:28 2013
@@ -10,70 +10,243 @@ public final class AccessControlProtos {
}
public interface PermissionOrBuilder
extends com.google.protobuf.MessageOrBuilder {
-
+
// repeated .Permission.Action action = 1;
+ /**
+ * <code>repeated .Permission.Action action = 1;</code>
+ */
java.util.List<org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Action> getActionList();
+ /**
+ * <code>repeated .Permission.Action action = 1;</code>
+ */
int getActionCount();
+ /**
+ * <code>repeated .Permission.Action action = 1;</code>
+ */
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Action getAction(int index);
-
+
// optional bytes table = 2;
+ /**
+ * <code>optional bytes table = 2;</code>
+ */
boolean hasTable();
+ /**
+ * <code>optional bytes table = 2;</code>
+ */
com.google.protobuf.ByteString getTable();
-
+
// optional bytes family = 3;
+ /**
+ * <code>optional bytes family = 3;</code>
+ */
boolean hasFamily();
+ /**
+ * <code>optional bytes family = 3;</code>
+ */
com.google.protobuf.ByteString getFamily();
-
+
// optional bytes qualifier = 4;
+ /**
+ * <code>optional bytes qualifier = 4;</code>
+ */
boolean hasQualifier();
+ /**
+ * <code>optional bytes qualifier = 4;</code>
+ */
com.google.protobuf.ByteString getQualifier();
}
+ /**
+ * Protobuf type {@code Permission}
+ */
public static final class Permission extends
com.google.protobuf.GeneratedMessage
implements PermissionOrBuilder {
// Use Permission.newBuilder() to construct.
- private Permission(Builder builder) {
+ private Permission(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
super(builder);
+ this.unknownFields = builder.getUnknownFields();
}
- private Permission(boolean noInit) {}
-
+ private Permission(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
private static final Permission defaultInstance;
public static Permission getDefaultInstance() {
return defaultInstance;
}
-
+
public Permission getDefaultInstanceForType() {
return defaultInstance;
}
-
+
+ private final com.google.protobuf.UnknownFieldSet unknownFields;
+ @java.lang.Override
+ public final com.google.protobuf.UnknownFieldSet
+ getUnknownFields() {
+ return this.unknownFields;
+ }
+ private Permission(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ initFields();
+ int mutable_bitField0_ = 0;
+ com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ com.google.protobuf.UnknownFieldSet.newBuilder();
+ try {
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ default: {
+ if (!parseUnknownField(input, unknownFields,
+ extensionRegistry, tag)) {
+ done = true;
+ }
+ break;
+ }
+ case 8: {
+ int rawValue = input.readEnum();
+ org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Action value = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Action.valueOf(rawValue);
+ if (value == null) {
+ unknownFields.mergeVarintField(1, rawValue);
+ } else {
+ if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
+ action_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Action>();
+ mutable_bitField0_ |= 0x00000001;
+ }
+ action_.add(value);
+ }
+ break;
+ }
+ case 10: {
+ int length = input.readRawVarint32();
+ int oldLimit = input.pushLimit(length);
+ while(input.getBytesUntilLimit() > 0) {
+ int rawValue = input.readEnum();
+ org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Action value = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Action.valueOf(rawValue);
+ if (value == null) {
+ unknownFields.mergeVarintField(1, rawValue);
+ } else {
+ if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
+ action_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Action>();
+ mutable_bitField0_ |= 0x00000001;
+ }
+ action_.add(value);
+ }
+ }
+ input.popLimit(oldLimit);
+ break;
+ }
+ case 18: {
+ bitField0_ |= 0x00000001;
+ table_ = input.readBytes();
+ break;
+ }
+ case 26: {
+ bitField0_ |= 0x00000002;
+ family_ = input.readBytes();
+ break;
+ }
+ case 34: {
+ bitField0_ |= 0x00000004;
+ qualifier_ = input.readBytes();
+ break;
+ }
+ }
+ }
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(this);
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(
+ e.getMessage()).setUnfinishedMessage(this);
+ } finally {
+ if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
+ action_ = java.util.Collections.unmodifiableList(action_);
+ }
+ this.unknownFields = unknownFields.build();
+ makeExtensionsImmutable();
+ }
+ }
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.internal_static_Permission_descriptor;
}
-
+
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.internal_static_Permission_fieldAccessorTable;
+ return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.internal_static_Permission_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.class, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Builder.class);
+ }
+
+ public static com.google.protobuf.Parser<Permission> PARSER =
+ new com.google.protobuf.AbstractParser<Permission>() {
+ public Permission parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new Permission(input, extensionRegistry);
+ }
+ };
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<Permission> getParserForType() {
+ return PARSER;
}
-
+
+ /**
+ * Protobuf enum {@code Permission.Action}
+ */
public enum Action
implements com.google.protobuf.ProtocolMessageEnum {
+ /**
+ * <code>READ = 0;</code>
+ */
READ(0, 0),
+ /**
+ * <code>WRITE = 1;</code>
+ */
WRITE(1, 1),
+ /**
+ * <code>EXEC = 2;</code>
+ */
EXEC(2, 2),
+ /**
+ * <code>CREATE = 3;</code>
+ */
CREATE(3, 3),
+ /**
+ * <code>ADMIN = 4;</code>
+ */
ADMIN(4, 4),
;
-
+
+ /**
+ * <code>READ = 0;</code>
+ */
public static final int READ_VALUE = 0;
+ /**
+ * <code>WRITE = 1;</code>
+ */
public static final int WRITE_VALUE = 1;
+ /**
+ * <code>EXEC = 2;</code>
+ */
public static final int EXEC_VALUE = 2;
+ /**
+ * <code>CREATE = 3;</code>
+ */
public static final int CREATE_VALUE = 3;
+ /**
+ * <code>ADMIN = 4;</code>
+ */
public static final int ADMIN_VALUE = 4;
-
-
+
+
public final int getNumber() { return value; }
-
+
public static Action valueOf(int value) {
switch (value) {
case 0: return READ;
@@ -84,7 +257,7 @@ public final class AccessControlProtos {
default: return null;
}
}
-
+
public static com.google.protobuf.Internal.EnumLiteMap<Action>
internalGetValueMap() {
return internalValueMap;
@@ -96,7 +269,7 @@ public final class AccessControlProtos {
return Action.valueOf(number);
}
};
-
+
public final com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(index);
@@ -109,11 +282,9 @@ public final class AccessControlProtos {
getDescriptor() {
return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.getDescriptor().getEnumTypes().get(0);
}
-
- private static final Action[] VALUES = {
- READ, WRITE, EXEC, CREATE, ADMIN,
- };
-
+
+ private static final Action[] VALUES = values();
+
public static Action valueOf(
com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
@@ -122,62 +293,89 @@ public final class AccessControlProtos {
}
return VALUES[desc.getIndex()];
}
-
+
private final int index;
private final int value;
-
+
private Action(int index, int value) {
this.index = index;
this.value = value;
}
-
+
// @@protoc_insertion_point(enum_scope:Permission.Action)
}
-
+
private int bitField0_;
// repeated .Permission.Action action = 1;
public static final int ACTION_FIELD_NUMBER = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Action> action_;
+ /**
+ * <code>repeated .Permission.Action action = 1;</code>
+ */
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Action> getActionList() {
return action_;
}
+ /**
+ * <code>repeated .Permission.Action action = 1;</code>
+ */
public int getActionCount() {
return action_.size();
}
+ /**
+ * <code>repeated .Permission.Action action = 1;</code>
+ */
public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Action getAction(int index) {
return action_.get(index);
}
-
+
// optional bytes table = 2;
public static final int TABLE_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString table_;
+ /**
+ * <code>optional bytes table = 2;</code>
+ */
public boolean hasTable() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
+ /**
+ * <code>optional bytes table = 2;</code>
+ */
public com.google.protobuf.ByteString getTable() {
return table_;
}
-
+
// optional bytes family = 3;
public static final int FAMILY_FIELD_NUMBER = 3;
private com.google.protobuf.ByteString family_;
+ /**
+ * <code>optional bytes family = 3;</code>
+ */
public boolean hasFamily() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
+ /**
+ * <code>optional bytes family = 3;</code>
+ */
public com.google.protobuf.ByteString getFamily() {
return family_;
}
-
+
// optional bytes qualifier = 4;
public static final int QUALIFIER_FIELD_NUMBER = 4;
private com.google.protobuf.ByteString qualifier_;
+ /**
+ * <code>optional bytes qualifier = 4;</code>
+ */
public boolean hasQualifier() {
return ((bitField0_ & 0x00000004) == 0x00000004);
}
+ /**
+ * <code>optional bytes qualifier = 4;</code>
+ */
public com.google.protobuf.ByteString getQualifier() {
return qualifier_;
}
-
+
private void initFields() {
action_ = java.util.Collections.emptyList();
table_ = com.google.protobuf.ByteString.EMPTY;
@@ -188,11 +386,11 @@ public final class AccessControlProtos {
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
-
+
memoizedIsInitialized = 1;
return true;
}
-
+
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
@@ -210,12 +408,12 @@ public final class AccessControlProtos {
}
getUnknownFields().writeTo(output);
}
-
+
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
-
+
size = 0;
{
int dataSize = 0;
@@ -242,14 +440,14 @@ public final class AccessControlProtos {
memoizedSerializedSize = size;
return size;
}
-
+
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
-
+
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
@@ -259,7 +457,7 @@ public final class AccessControlProtos {
return super.equals(obj);
}
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission other = (org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission) obj;
-
+
boolean result = true;
result = result && getActionList()
.equals(other.getActionList());
@@ -282,9 +480,13 @@ public final class AccessControlProtos {
getUnknownFields().equals(other.getUnknownFields());
return result;
}
-
+
+ private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
+ if (memoizedHashCode != 0) {
+ return memoizedHashCode;
+ }
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (getActionCount() > 0) {
@@ -304,89 +506,79 @@ public final class AccessControlProtos {
hash = (53 * hash) + getQualifier().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
+ memoizedHashCode = hash;
return hash;
}
-
+
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data).buildParsed();
+ return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data).buildParsed();
+ return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return newBuilder().mergeFrom(input).buildParsed();
+ return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return newBuilder().mergeFrom(input, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- Builder builder = newBuilder();
- if (builder.mergeDelimitedFrom(input)) {
- return builder.buildParsed();
- } else {
- return null;
- }
+ return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- Builder builder = newBuilder();
- if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
- return builder.buildParsed();
- } else {
- return null;
- }
+ return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return newBuilder().mergeFrom(input).buildParsed();
+ return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return newBuilder().mergeFrom(input, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(input, extensionRegistry);
}
-
+
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
-
+
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
+ /**
+ * Protobuf type {@code Permission}
+ */
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder<Builder>
implements org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.PermissionOrBuilder {
@@ -394,18 +586,21 @@ public final class AccessControlProtos {
getDescriptor() {
return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.internal_static_Permission_descriptor;
}
-
+
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.internal_static_Permission_fieldAccessorTable;
+ return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.internal_static_Permission_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.class, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Builder.class);
}
-
+
// Construct using org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
-
- private Builder(BuilderParent parent) {
+
+ private Builder(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
@@ -416,7 +611,7 @@ public final class AccessControlProtos {
private static Builder create() {
return new Builder();
}
-
+
public Builder clear() {
super.clear();
action_ = java.util.Collections.emptyList();
@@ -429,20 +624,20 @@ public final class AccessControlProtos {
bitField0_ = (bitField0_ & ~0x00000008);
return this;
}
-
+
public Builder clone() {
return create().mergeFrom(buildPartial());
}
-
+
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
- return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.getDescriptor();
+ return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.internal_static_Permission_descriptor;
}
-
+
public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission getDefaultInstanceForType() {
return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.getDefaultInstance();
}
-
+
public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission build() {
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission result = buildPartial();
if (!result.isInitialized()) {
@@ -450,17 +645,7 @@ public final class AccessControlProtos {
}
return result;
}
-
- private org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission buildParsed()
- throws com.google.protobuf.InvalidProtocolBufferException {
- org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission result = buildPartial();
- if (!result.isInitialized()) {
- throw newUninitializedMessageException(
- result).asInvalidProtocolBufferException();
- }
- return result;
- }
-
+
public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission buildPartial() {
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission result = new org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission(this);
int from_bitField0_ = bitField0_;
@@ -486,7 +671,7 @@ public final class AccessControlProtos {
onBuilt();
return result;
}
-
+
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission) {
return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission)other);
@@ -495,7 +680,7 @@ public final class AccessControlProtos {
return this;
}
}
-
+
public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission other) {
if (other == org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.getDefaultInstance()) return this;
if (!other.action_.isEmpty()) {
@@ -520,80 +705,30 @@ public final class AccessControlProtos {
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
-
+
public final boolean isInitialized() {
return true;
}
-
+
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder(
- this.getUnknownFields());
- while (true) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
- this.setUnknownFields(unknownFields.build());
- onChanged();
- return this;
- default: {
- if (!parseUnknownField(input, unknownFields,
- extensionRegistry, tag)) {
- this.setUnknownFields(unknownFields.build());
- onChanged();
- return this;
- }
- break;
- }
- case 8: {
- int rawValue = input.readEnum();
- org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Action value = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Action.valueOf(rawValue);
- if (value == null) {
- unknownFields.mergeVarintField(1, rawValue);
- } else {
- addAction(value);
- }
- break;
- }
- case 10: {
- int length = input.readRawVarint32();
- int oldLimit = input.pushLimit(length);
- while(input.getBytesUntilLimit() > 0) {
- int rawValue = input.readEnum();
- org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Action value = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Action.valueOf(rawValue);
- if (value == null) {
- unknownFields.mergeVarintField(1, rawValue);
- } else {
- addAction(value);
- }
- }
- input.popLimit(oldLimit);
- break;
- }
- case 18: {
- bitField0_ |= 0x00000002;
- table_ = input.readBytes();
- break;
- }
- case 26: {
- bitField0_ |= 0x00000004;
- family_ = input.readBytes();
- break;
- }
- case 34: {
- bitField0_ |= 0x00000008;
- qualifier_ = input.readBytes();
- break;
- }
+ org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission parsedMessage = null;
+ try {
+ parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission) e.getUnfinishedMessage();
+ throw e;
+ } finally {
+ if (parsedMessage != null) {
+ mergeFrom(parsedMessage);
}
}
+ return this;
}
-
private int bitField0_;
-
+
// repeated .Permission.Action action = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Action> action_ =
java.util.Collections.emptyList();
@@ -603,15 +738,27 @@ public final class AccessControlProtos {
bitField0_ |= 0x00000001;
}
}
+ /**
+ * <code>repeated .Permission.Action action = 1;</code>
+ */
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Action> getActionList() {
return java.util.Collections.unmodifiableList(action_);
}
+ /**
+ * <code>repeated .Permission.Action action = 1;</code>
+ */
public int getActionCount() {
return action_.size();
}
+ /**
+ * <code>repeated .Permission.Action action = 1;</code>
+ */
public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Action getAction(int index) {
return action_.get(index);
}
+ /**
+ * <code>repeated .Permission.Action action = 1;</code>
+ */
public Builder setAction(
int index, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Action value) {
if (value == null) {
@@ -622,6 +769,9 @@ public final class AccessControlProtos {
onChanged();
return this;
}
+ /**
+ * <code>repeated .Permission.Action action = 1;</code>
+ */
public Builder addAction(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Action value) {
if (value == null) {
throw new NullPointerException();
@@ -631,6 +781,9 @@ public final class AccessControlProtos {
onChanged();
return this;
}
+ /**
+ * <code>repeated .Permission.Action action = 1;</code>
+ */
public Builder addAllAction(
java.lang.Iterable<? extends org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Action> values) {
ensureActionIsMutable();
@@ -638,21 +791,33 @@ public final class AccessControlProtos {
onChanged();
return this;
}
+ /**
+ * <code>repeated .Permission.Action action = 1;</code>
+ */
public Builder clearAction() {
action_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
return this;
}
-
+
// optional bytes table = 2;
private com.google.protobuf.ByteString table_ = com.google.protobuf.ByteString.EMPTY;
+ /**
+ * <code>optional bytes table = 2;</code>
+ */
public boolean hasTable() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
+ /**
+ * <code>optional bytes table = 2;</code>
+ */
public com.google.protobuf.ByteString getTable() {
return table_;
}
+ /**
+ * <code>optional bytes table = 2;</code>
+ */
public Builder setTable(com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
@@ -662,21 +827,33 @@ public final class AccessControlProtos {
onChanged();
return this;
}
+ /**
+ * <code>optional bytes table = 2;</code>
+ */
public Builder clearTable() {
bitField0_ = (bitField0_ & ~0x00000002);
table_ = getDefaultInstance().getTable();
onChanged();
return this;
}
-
+
// optional bytes family = 3;
private com.google.protobuf.ByteString family_ = com.google.protobuf.ByteString.EMPTY;
+ /**
+ * <code>optional bytes family = 3;</code>
+ */
public boolean hasFamily() {
return ((bitField0_ & 0x00000004) == 0x00000004);
}
+ /**
+ * <code>optional bytes family = 3;</code>
+ */
public com.google.protobuf.ByteString getFamily() {
return family_;
}
+ /**
+ * <code>optional bytes family = 3;</code>
+ */
public Builder setFamily(com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
@@ -686,21 +863,33 @@ public final class AccessControlProtos {
onChanged();
return this;
}
+ /**
+ * <code>optional bytes family = 3;</code>
+ */
public Builder clearFamily() {
bitField0_ = (bitField0_ & ~0x00000004);
family_ = getDefaultInstance().getFamily();
onChanged();
return this;
}
-
+
// optional bytes qualifier = 4;
private com.google.protobuf.ByteString qualifier_ = com.google.protobuf.ByteString.EMPTY;
+ /**
+ * <code>optional bytes qualifier = 4;</code>
+ */
public boolean hasQualifier() {
return ((bitField0_ & 0x00000008) == 0x00000008);
}
+ /**
+ * <code>optional bytes qualifier = 4;</code>
+ */
public com.google.protobuf.ByteString getQualifier() {
return qualifier_;
}
+ /**
+ * <code>optional bytes qualifier = 4;</code>
+ */
public Builder setQualifier(com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
@@ -710,88 +899,201 @@ public final class AccessControlProtos {
onChanged();
return this;
}
+ /**
+ * <code>optional bytes qualifier = 4;</code>
+ */
public Builder clearQualifier() {
bitField0_ = (bitField0_ & ~0x00000008);
qualifier_ = getDefaultInstance().getQualifier();
onChanged();
return this;
}
-
+
// @@protoc_insertion_point(builder_scope:Permission)
}
-
+
static {
defaultInstance = new Permission(true);
defaultInstance.initFields();
}
-
+
// @@protoc_insertion_point(class_scope:Permission)
}
-
+
public interface UserPermissionOrBuilder
extends com.google.protobuf.MessageOrBuilder {
-
+
// required bytes user = 1;
+ /**
+ * <code>required bytes user = 1;</code>
+ */
boolean hasUser();
+ /**
+ * <code>required bytes user = 1;</code>
+ */
com.google.protobuf.ByteString getUser();
-
+
// required .Permission permission = 2;
+ /**
+ * <code>required .Permission permission = 2;</code>
+ */
boolean hasPermission();
+ /**
+ * <code>required .Permission permission = 2;</code>
+ */
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission getPermission();
+ /**
+ * <code>required .Permission permission = 2;</code>
+ */
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.PermissionOrBuilder getPermissionOrBuilder();
}
+ /**
+ * Protobuf type {@code UserPermission}
+ */
public static final class UserPermission extends
com.google.protobuf.GeneratedMessage
implements UserPermissionOrBuilder {
// Use UserPermission.newBuilder() to construct.
- private UserPermission(Builder builder) {
+ private UserPermission(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
super(builder);
+ this.unknownFields = builder.getUnknownFields();
}
- private UserPermission(boolean noInit) {}
-
+ private UserPermission(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
private static final UserPermission defaultInstance;
public static UserPermission getDefaultInstance() {
return defaultInstance;
}
-
+
public UserPermission getDefaultInstanceForType() {
return defaultInstance;
}
-
+
+ private final com.google.protobuf.UnknownFieldSet unknownFields;
+ @java.lang.Override
+ public final com.google.protobuf.UnknownFieldSet
+ getUnknownFields() {
+ return this.unknownFields;
+ }
+ private UserPermission(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ initFields();
+ int mutable_bitField0_ = 0;
+ com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ com.google.protobuf.UnknownFieldSet.newBuilder();
+ try {
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ default: {
+ if (!parseUnknownField(input, unknownFields,
+ extensionRegistry, tag)) {
+ done = true;
+ }
+ break;
+ }
+ case 10: {
+ bitField0_ |= 0x00000001;
+ user_ = input.readBytes();
+ break;
+ }
+ case 18: {
+ org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Builder subBuilder = null;
+ if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ subBuilder = permission_.toBuilder();
+ }
+ permission_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.PARSER, extensionRegistry);
+ if (subBuilder != null) {
+ subBuilder.mergeFrom(permission_);
+ permission_ = subBuilder.buildPartial();
+ }
+ bitField0_ |= 0x00000002;
+ break;
+ }
+ }
+ }
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(this);
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(
+ e.getMessage()).setUnfinishedMessage(this);
+ } finally {
+ this.unknownFields = unknownFields.build();
+ makeExtensionsImmutable();
+ }
+ }
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.internal_static_UserPermission_descriptor;
}
-
+
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.internal_static_UserPermission_fieldAccessorTable;
+ return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.internal_static_UserPermission_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission.class, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission.Builder.class);
+ }
+
+ public static com.google.protobuf.Parser<UserPermission> PARSER =
+ new com.google.protobuf.AbstractParser<UserPermission>() {
+ public UserPermission parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new UserPermission(input, extensionRegistry);
+ }
+ };
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<UserPermission> getParserForType() {
+ return PARSER;
}
-
+
private int bitField0_;
// required bytes user = 1;
public static final int USER_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString user_;
+ /**
+ * <code>required bytes user = 1;</code>
+ */
public boolean hasUser() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
+ /**
+ * <code>required bytes user = 1;</code>
+ */
public com.google.protobuf.ByteString getUser() {
return user_;
}
-
+
// required .Permission permission = 2;
public static final int PERMISSION_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission permission_;
+ /**
+ * <code>required .Permission permission = 2;</code>
+ */
public boolean hasPermission() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
+ /**
+ * <code>required .Permission permission = 2;</code>
+ */
public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission getPermission() {
return permission_;
}
+ /**
+ * <code>required .Permission permission = 2;</code>
+ */
public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.PermissionOrBuilder getPermissionOrBuilder() {
return permission_;
}
-
+
private void initFields() {
user_ = com.google.protobuf.ByteString.EMPTY;
permission_ = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.getDefaultInstance();
@@ -800,7 +1102,7 @@ public final class AccessControlProtos {
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
-
+
if (!hasUser()) {
memoizedIsInitialized = 0;
return false;
@@ -812,7 +1114,7 @@ public final class AccessControlProtos {
memoizedIsInitialized = 1;
return true;
}
-
+
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
@@ -824,12 +1126,12 @@ public final class AccessControlProtos {
}
getUnknownFields().writeTo(output);
}
-
+
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
-
+
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
@@ -843,14 +1145,14 @@ public final class AccessControlProtos {
memoizedSerializedSize = size;
return size;
}
-
+
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
-
+
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
@@ -860,7 +1162,7 @@ public final class AccessControlProtos {
return super.equals(obj);
}
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission other = (org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission) obj;
-
+
boolean result = true;
result = result && (hasUser() == other.hasUser());
if (hasUser()) {
@@ -876,9 +1178,13 @@ public final class AccessControlProtos {
getUnknownFields().equals(other.getUnknownFields());
return result;
}
-
+
+ private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
+ if (memoizedHashCode != 0) {
+ return memoizedHashCode;
+ }
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasUser()) {
@@ -890,89 +1196,79 @@ public final class AccessControlProtos {
hash = (53 * hash) + getPermission().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
+ memoizedHashCode = hash;
return hash;
}
-
+
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data).buildParsed();
+ return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data).buildParsed();
+ return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return newBuilder().mergeFrom(input).buildParsed();
+ return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return newBuilder().mergeFrom(input, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- Builder builder = newBuilder();
- if (builder.mergeDelimitedFrom(input)) {
- return builder.buildParsed();
- } else {
- return null;
- }
+ return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- Builder builder = newBuilder();
- if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
- return builder.buildParsed();
- } else {
- return null;
- }
+ return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return newBuilder().mergeFrom(input).buildParsed();
+ return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return newBuilder().mergeFrom(input, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(input, extensionRegistry);
}
-
+
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
-
+
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
+ /**
+ * Protobuf type {@code UserPermission}
+ */
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder<Builder>
implements org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermissionOrBuilder {
@@ -980,18 +1276,21 @@ public final class AccessControlProtos {
getDescriptor() {
return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.internal_static_UserPermission_descriptor;
}
-
+
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.internal_static_UserPermission_fieldAccessorTable;
+ return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.internal_static_UserPermission_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission.class, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission.Builder.class);
}
-
+
// Construct using org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
-
- private Builder(BuilderParent parent) {
+
+ private Builder(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
@@ -1003,7 +1302,7 @@ public final class AccessControlProtos {
private static Builder create() {
return new Builder();
}
-
+
public Builder clear() {
super.clear();
user_ = com.google.protobuf.ByteString.EMPTY;
@@ -1016,20 +1315,20 @@ public final class AccessControlProtos {
bitField0_ = (bitField0_ & ~0x00000002);
return this;
}
-
+
public Builder clone() {
return create().mergeFrom(buildPartial());
}
-
+
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
- return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission.getDescriptor();
+ return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.internal_static_UserPermission_descriptor;
}
-
+
public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission getDefaultInstanceForType() {
return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission.getDefaultInstance();
}
-
+
public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission build() {
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission result = buildPartial();
if (!result.isInitialized()) {
@@ -1037,17 +1336,7 @@ public final class AccessControlProtos {
}
return result;
}
-
- private org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission buildParsed()
- throws com.google.protobuf.InvalidProtocolBufferException {
- org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission result = buildPartial();
- if (!result.isInitialized()) {
- throw newUninitializedMessageException(
- result).asInvalidProtocolBufferException();
- }
- return result;
- }
-
+
public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission buildPartial() {
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission result = new org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission(this);
int from_bitField0_ = bitField0_;
@@ -1068,7 +1357,7 @@ public final class AccessControlProtos {
onBuilt();
return result;
}
-
+
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission) {
return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission)other);
@@ -1077,7 +1366,7 @@ public final class AccessControlProtos {
return this;
}
}
-
+
public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission other) {
if (other == org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission.getDefaultInstance()) return this;
if (other.hasUser()) {
@@ -1089,7 +1378,7 @@ public final class AccessControlProtos {
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
-
+
public final boolean isInitialized() {
if (!hasUser()) {
@@ -1101,58 +1390,43 @@ public final class AccessControlProtos {
}
return true;
}
-
+
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder(
- this.getUnknownFields());
- while (true) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
- this.setUnknownFields(unknownFields.build());
- onChanged();
- return this;
- default: {
- if (!parseUnknownField(input, unknownFields,
- extensionRegistry, tag)) {
- this.setUnknownFields(unknownFields.build());
- onChanged();
- return this;
- }
- break;
- }
- case 10: {
- bitField0_ |= 0x00000001;
- user_ = input.readBytes();
- break;
- }
- case 18: {
- org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.newBuilder();
- if (hasPermission()) {
- subBuilder.mergeFrom(getPermission());
- }
- input.readMessage(subBuilder, extensionRegistry);
- setPermission(subBuilder.buildPartial());
- break;
- }
+ org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission parsedMessage = null;
+ try {
+ parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission) e.getUnfinishedMessage();
+ throw e;
+ } finally {
+ if (parsedMessage != null) {
+ mergeFrom(parsedMessage);
}
}
+ return this;
}
-
private int bitField0_;
-
+
// required bytes user = 1;
private com.google.protobuf.ByteString user_ = com.google.protobuf.ByteString.EMPTY;
+ /**
+ * <code>required bytes user = 1;</code>
+ */
public boolean hasUser() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
+ /**
+ * <code>required bytes user = 1;</code>
+ */
public com.google.protobuf.ByteString getUser() {
return user_;
}
+ /**
+ * <code>required bytes user = 1;</code>
+ */
public Builder setUser(com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
@@ -1162,20 +1436,29 @@ public final class AccessControlProtos {
onChanged();
return this;
}
+ /**
+ * <code>required bytes user = 1;</code>
+ */
public Builder clearUser() {
bitField0_ = (bitField0_ & ~0x00000001);
user_ = getDefaultInstance().getUser();
onChanged();
return this;
}
-
+
// required .Permission permission = 2;
private org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission permission_ = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Builder, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.PermissionOrBuilder> permissionBuilder_;
+ /**
+ * <code>required .Permission permission = 2;</code>
+ */
public boolean hasPermission() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
+ /**
+ * <code>required .Permission permission = 2;</code>
+ */
public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission getPermission() {
if (permissionBuilder_ == null) {
return permission_;
@@ -1183,6 +1466,9 @@ public final class AccessControlProtos {
return permissionBuilder_.getMessage();
}
}
+ /**
+ * <code>required .Permission permission = 2;</code>
+ */
public Builder setPermission(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission value) {
if (permissionBuilder_ == null) {
if (value == null) {
@@ -1196,6 +1482,9 @@ public final class AccessControlProtos {
bitField0_ |= 0x00000002;
return this;
}
+ /**
+ * <code>required .Permission permission = 2;</code>
+ */
public Builder setPermission(
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Builder builderForValue) {
if (permissionBuilder_ == null) {
@@ -1207,6 +1496,9 @@ public final class AccessControlProtos {
bitField0_ |= 0x00000002;
return this;
}
+ /**
+ * <code>required .Permission permission = 2;</code>
+ */
public Builder mergePermission(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission value) {
if (permissionBuilder_ == null) {
if (((bitField0_ & 0x00000002) == 0x00000002) &&
@@ -1223,6 +1515,9 @@ public final class AccessControlProtos {
bitField0_ |= 0x00000002;
return this;
}
+ /**
+ * <code>required .Permission permission = 2;</code>
+ */
public Builder clearPermission() {
if (permissionBuilder_ == null) {
permission_ = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.getDefaultInstance();
@@ -1233,11 +1528,17 @@ public final class AccessControlProtos {
bitField0_ = (bitField0_ & ~0x00000002);
return this;
}
+ /**
+ * <code>required .Permission permission = 2;</code>
+ */
public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Builder getPermissionBuilder() {
bitField0_ |= 0x00000002;
onChanged();
return getPermissionFieldBuilder().getBuilder();
}
+ /**
+ * <code>required .Permission permission = 2;</code>
+ */
public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.PermissionOrBuilder getPermissionOrBuilder() {
if (permissionBuilder_ != null) {
return permissionBuilder_.getMessageOrBuilder();
@@ -1245,6 +1546,9 @@ public final class AccessControlProtos {
return permission_;
}
}
+ /**
+ * <code>required .Permission permission = 2;</code>
+ */
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Builder, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.PermissionOrBuilder>
getPermissionFieldBuilder() {
@@ -1258,136 +1562,349 @@ public final class AccessControlProtos {
}
return permissionBuilder_;
}
-
+
// @@protoc_insertion_point(builder_scope:UserPermission)
}
-
+
static {
defaultInstance = new UserPermission(true);
defaultInstance.initFields();
}
-
+
// @@protoc_insertion_point(class_scope:UserPermission)
}
-
+
public interface UserTablePermissionsOrBuilder
extends com.google.protobuf.MessageOrBuilder {
-
+
// repeated .UserTablePermissions.UserPermissions permissions = 1;
+ /**
+ * <code>repeated .UserTablePermissions.UserPermissions permissions = 1;</code>
+ */
java.util.List<org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions>
getPermissionsList();
+ /**
+ * <code>repeated .UserTablePermissions.UserPermissions permissions = 1;</code>
+ */
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions getPermissions(int index);
+ /**
+ * <code>repeated .UserTablePermissions.UserPermissions permissions = 1;</code>
+ */
int getPermissionsCount();
+ /**
+ * <code>repeated .UserTablePermissions.UserPermissions permissions = 1;</code>
+ */
java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissionsOrBuilder>
getPermissionsOrBuilderList();
+ /**
+ * <code>repeated .UserTablePermissions.UserPermissions permissions = 1;</code>
+ */
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissionsOrBuilder getPermissionsOrBuilder(
int index);
}
+ /**
+ * Protobuf type {@code UserTablePermissions}
+ *
+ * <pre>
+ **
+ * Content of the /hbase/acl/<table> znode.
+ * </pre>
+ */
public static final class UserTablePermissions extends
com.google.protobuf.GeneratedMessage
implements UserTablePermissionsOrBuilder {
// Use UserTablePermissions.newBuilder() to construct.
- private UserTablePermissions(Builder builder) {
+ private UserTablePermissions(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
super(builder);
+ this.unknownFields = builder.getUnknownFields();
}
- private UserTablePermissions(boolean noInit) {}
-
+ private UserTablePermissions(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
private static final UserTablePermissions defaultInstance;
public static UserTablePermissions getDefaultInstance() {
return defaultInstance;
}
-
+
public UserTablePermissions getDefaultInstanceForType() {
return defaultInstance;
}
-
+
+ private final com.google.protobuf.UnknownFieldSet unknownFields;
+ @java.lang.Override
+ public final com.google.protobuf.UnknownFieldSet
+ getUnknownFields() {
+ return this.unknownFields;
+ }
+ private UserTablePermissions(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ initFields();
+ int mutable_bitField0_ = 0;
+ com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ com.google.protobuf.UnknownFieldSet.newBuilder();
+ try {
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ default: {
+ if (!parseUnknownField(input, unknownFields,
+ extensionRegistry, tag)) {
+ done = true;
+ }
+ break;
+ }
+ case 10: {
+ if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
+ permissions_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions>();
+ mutable_bitField0_ |= 0x00000001;
+ }
+ permissions_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions.PARSER, extensionRegistry));
+ break;
+ }
+ }
+ }
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(this);
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(
+ e.getMessage()).setUnfinishedMessage(this);
+ } finally {
+ if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
+ permissions_ = java.util.Collections.unmodifiableList(permissions_);
+ }
+ this.unknownFields = unknownFields.build();
+ makeExtensionsImmutable();
+ }
+ }
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.internal_static_UserTablePermissions_descriptor;
}
-
+
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.internal_static_UserTablePermissions_fieldAccessorTable;
+ return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.internal_static_UserTablePermissions_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.class, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.Builder.class);
+ }
+
+ public static com.google.protobuf.Parser<UserTablePermissions> PARSER =
+ new com.google.protobuf.AbstractParser<UserTablePermissions>() {
+ public UserTablePermissions parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new UserTablePermissions(input, extensionRegistry);
+ }
+ };
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<UserTablePermissions> getParserForType() {
+ return PARSER;
}
-
+
public interface UserPermissionsOrBuilder
extends com.google.protobuf.MessageOrBuilder {
-
+
// required bytes user = 1;
+ /**
+ * <code>required bytes user = 1;</code>
+ */
boolean hasUser();
+ /**
+ * <code>required bytes user = 1;</code>
+ */
com.google.protobuf.ByteString getUser();
-
+
// repeated .Permission permissions = 2;
+ /**
+ * <code>repeated .Permission permissions = 2;</code>
+ */
java.util.List<org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission>
getPermissionsList();
+ /**
+ * <code>repeated .Permission permissions = 2;</code>
+ */
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission getPermissions(int index);
+ /**
+ * <code>repeated .Permission permissions = 2;</code>
+ */
int getPermissionsCount();
+ /**
+ * <code>repeated .Permission permissions = 2;</code>
+ */
java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.PermissionOrBuilder>
getPermissionsOrBuilderList();
+ /**
+ * <code>repeated .Permission permissions = 2;</code>
+ */
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.PermissionOrBuilder getPermissionsOrBuilder(
int index);
}
+ /**
+ * Protobuf type {@code UserTablePermissions.UserPermissions}
+ */
public static final class UserPermissions extends
com.google.protobuf.GeneratedMessage
implements UserPermissionsOrBuilder {
// Use UserPermissions.newBuilder() to construct.
- private UserPermissions(Builder builder) {
+ private UserPermissions(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
super(builder);
+ this.unknownFields = builder.getUnknownFields();
}
- private UserPermissions(boolean noInit) {}
-
+ private UserPermissions(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
private static final UserPermissions defaultInstance;
public static UserPermissions getDefaultInstance() {
return defaultInstance;
}
-
+
public UserPermissions getDefaultInstanceForType() {
return defaultInstance;
}
-
+
+ private final com.google.protobuf.UnknownFieldSet unknownFields;
+ @java.lang.Override
+ public final com.google.protobuf.UnknownFieldSet
+ getUnknownFields() {
+ return this.unknownFields;
+ }
+ private UserPermissions(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ initFields();
+ int mutable_bitField0_ = 0;
+ com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ com.google.protobuf.UnknownFieldSet.newBuilder();
+ try {
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ default: {
+ if (!parseUnknownField(input, unknownFields,
+ extensionRegistry, tag)) {
+ done = true;
+ }
+ break;
+ }
+ case 10: {
+ bitField0_ |= 0x00000001;
+ user_ = input.readBytes();
+ break;
+ }
+ case 18: {
+ if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
+ permissions_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission>();
+ mutable_bitField0_ |= 0x00000002;
+ }
+ permissions_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.PARSER, extensionRegistry));
+ break;
+ }
+ }
+ }
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(this);
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(
+ e.getMessage()).setUnfinishedMessage(this);
+ } finally {
+ if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
+ permissions_ = java.util.Collections.unmodifiableList(permissions_);
+ }
+ this.unknownFields = unknownFields.build();
+ makeExtensionsImmutable();
+ }
+ }
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.internal_static_UserTablePermissions_UserPermissions_descriptor;
}
-
+
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.internal_static_UserTablePermissions_UserPermissions_fieldAccessorTable;
+ return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.internal_static_UserTablePermissions_UserPermissions_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions.class, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions.Builder.class);
+ }
+
+ public static com.google.protobuf.Parser<UserPermissions> PARSER =
+ new com.google.protobuf.AbstractParser<UserPermissions>() {
+ public UserPermissions parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new UserPermissions(input, extensionRegistry);
+ }
+ };
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<UserPermissions> getParserForType() {
+ return PARSER;
}
-
+
private int bitField0_;
// required bytes user = 1;
public static final int USER_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString user_;
+ /**
+ * <code>required bytes user = 1;</code>
+ */
public boolean hasUser() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
+ /**
+ * <code>required bytes user = 1;</code>
+ */
public com.google.protobuf.ByteString getUser() {
return user_;
}
-
+
// repeated .Permission permissions = 2;
public static final int PERMISSIONS_FIELD_NUMBER = 2;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission> permissions_;
+ /**
+ * <code>repeated .Permission permissions = 2;</code>
+ */
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission> getPermissionsList() {
return permissions_;
}
+ /**
+ * <code>repeated .Permission permissions = 2;</code>
+ */
public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.PermissionOrBuilder>
getPermissionsOrBuilderList() {
return permissions_;
}
+ /**
+ * <code>repeated .Permission permissions = 2;</code>
+ */
public int getPermissionsCount() {
return permissions_.size();
}
+ /**
+ * <code>repeated .Permission permissions = 2;</code>
+ */
public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission getPermissions(int index) {
return permissions_.get(index);
}
+ /**
+ * <code>repeated .Permission permissions = 2;</code>
+ */
public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.PermissionOrBuilder getPermissionsOrBuilder(
int index) {
return permissions_.get(index);
}
-
+
private void initFields() {
user_ = com.google.protobuf.ByteString.EMPTY;
permissions_ = java.util.Collections.emptyList();
@@ -1396,7 +1913,7 @@ public final class AccessControlProtos {
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
-
+
if (!hasUser()) {
memoizedIsInitialized = 0;
return false;
@@ -1404,7 +1921,7 @@ public final class AccessControlProtos {
memoizedIsInitialized = 1;
return true;
}
-
+
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
@@ -1416,12 +1933,12 @@ public final class AccessControlProtos {
}
getUnknownFields().writeTo(output);
}
-
+
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
-
+
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
@@ -1435,14 +1952,14 @@ public final class AccessControlProtos {
memoizedSerializedSize = size;
return size;
}
-
+
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
-
+
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
@@ -1452,7 +1969,7 @@ public final class AccessControlProtos {
return super.equals(obj);
}
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions other = (org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions) obj;
-
+
boolean result = true;
result = result && (hasUser() == other.hasUser());
if (hasUser()) {
@@ -1465,9 +1982,13 @@ public final class AccessControlProtos {
getUnknownFields().equals(other.getUnknownFields());
return result;
}
-
+
+ private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
+ if (memoizedHashCode != 0) {
+ return memoizedHashCode;
+ }
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasUser()) {
@@ -1479,89 +2000,79 @@ public final class AccessControlProtos {
hash = (53 * hash) + getPermissionsList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
+ memoizedHashCode = hash;
return hash;
}
-
+
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data).buildParsed();
+ return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data).buildParsed();
+ return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return newBuilder().mergeFrom(input).buildParsed();
+ return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return newBuilder().mergeFrom(input, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- Builder builder = newBuilder();
- if (builder.mergeDelimitedFrom(input)) {
- return builder.buildParsed();
- } else {
- return null;
- }
+ return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- Builder builder = newBuilder();
- if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
- return builder.buildParsed();
- } else {
- return null;
- }
+ return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return newBuilder().mergeFrom(input).buildParsed();
+ return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return newBuilder().mergeFrom(input, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(input, extensionRegistry);
}
-
+
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
-
+
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
+ /**
+ * Protobuf type {@code UserTablePermissions.UserPermissions}
+ */
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder<Builder>
implements org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissionsOrBuilder {
@@ -1569,18 +2080,21 @@ public final class AccessControlProtos {
getDescriptor() {
return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.internal_static_UserTablePermissions_UserPermissions_descriptor;
}
-
+
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.internal_static_UserTablePermissions_UserPermissions_fieldAccessorTable;
+ return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.internal_static_UserTablePermissions_UserPermissions_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions.class, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions.Builder.class);
}
-
+
// Construct using org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
-
- private Builder(BuilderParent parent) {
+
+ private Builder(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
@@ -1592,7 +2106,7 @@ public final class AccessControlProtos {
private static Builder create() {
return new Builder();
}
-
+
public Builder clear() {
super.clear();
user_ = com.google.protobuf.ByteString.EMPTY;
@@ -1605,20 +2119,20 @@ public final class AccessControlProtos {
}
return this;
}
-
+
public Builder clone() {
return create().mergeFrom(buildPartial());
}
-
+
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
- return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions.getDescriptor();
+ return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.internal_static_UserTablePermissions_UserPermissions_descriptor;
}
-
+
public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions getDefaultInstanceForType() {
return org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions.getDefaultInstance();
}
-
+
public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions build() {
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions result = buildPartial();
if (!result.isInitialized()) {
@@ -1626,17 +2140,7 @@ public final class AccessControlProtos {
}
return result;
}
-
- private org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions buildParsed()
- throws com.google.protobuf.InvalidProtocolBufferException {
- org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions result = buildPartial();
- if (!result.isInitialized()) {
- throw newUninitializedMessageException(
- result).asInvalidProtocolBufferException();
- }
- return result;
- }
-
+
public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions buildPartial() {
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions result = new org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions(this);
int from_bitField0_ = bitField0_;
@@ -1658,7 +2162,7 @@ public final class AccessControlProtos {
onBuilt();
return result;
}
-
+
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions) {
return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions)other);
@@ -1667,7 +2171,7 @@ public final class AccessControlProtos {
return this;
}
}
-
+
public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions other) {
if (other == org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserTablePermissions.UserPermissions.getDefaultInstance()) return this;
if (other.hasUser()) {
@@ -1702,7 +2206,7 @@ public final class AccessControlProtos {
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
-
+
public final boolean isInitialized() {
if (!hasUser()) {
@@ -1710,55 +2214,43 @@ public final class AccessControlProtos {
}
return true;
}
-
+
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder(
- this.getUnknownFields());
- while (true) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
[... 5772 lines stripped ...]