You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2013/08/21 07:10:45 UTC
svn commit: r1516086 [39/43] - in /hbase/branches/0.95: ./
hbase-client/src/main/java/org/apache/hadoop/hbase/
hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/
hbase-common/src/test/java/org/apache/hadoop/hbase/
hbase-protocol/src/main/java...
Modified: hbase/branches/0.95/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/IncrementCounterProcessorTestProtos.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.95/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/IncrementCounterProcessorTestProtos.java?rev=1516086&r1=1516085&r2=1516086&view=diff
==============================================================================
--- hbase/branches/0.95/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/IncrementCounterProcessorTestProtos.java (original)
+++ hbase/branches/0.95/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/IncrementCounterProcessorTestProtos.java Wed Aug 21 05:10:43 2013
@@ -10,64 +10,160 @@ public final class IncrementCounterProce
}
public interface IncCounterProcessorRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder {
-
+
// required bytes row = 1;
+ /**
+ * <code>required bytes row = 1;</code>
+ */
boolean hasRow();
+ /**
+ * <code>required bytes row = 1;</code>
+ */
com.google.protobuf.ByteString getRow();
-
+
// required int32 counter = 2;
+ /**
+ * <code>required int32 counter = 2;</code>
+ */
boolean hasCounter();
+ /**
+ * <code>required int32 counter = 2;</code>
+ */
int getCounter();
}
+ /**
+ * Protobuf type {@code IncCounterProcessorRequest}
+ */
public static final class IncCounterProcessorRequest extends
com.google.protobuf.GeneratedMessage
implements IncCounterProcessorRequestOrBuilder {
// Use IncCounterProcessorRequest.newBuilder() to construct.
- private IncCounterProcessorRequest(Builder builder) {
+ private IncCounterProcessorRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
super(builder);
+ this.unknownFields = builder.getUnknownFields();
}
- private IncCounterProcessorRequest(boolean noInit) {}
-
+ private IncCounterProcessorRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
private static final IncCounterProcessorRequest defaultInstance;
public static IncCounterProcessorRequest getDefaultInstance() {
return defaultInstance;
}
-
+
public IncCounterProcessorRequest getDefaultInstanceForType() {
return defaultInstance;
}
-
+
+ private final com.google.protobuf.UnknownFieldSet unknownFields;
+ @java.lang.Override
+ public final com.google.protobuf.UnknownFieldSet
+ getUnknownFields() {
+ return this.unknownFields;
+ }
+ private IncCounterProcessorRequest(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ initFields();
+ int mutable_bitField0_ = 0;
+ com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ com.google.protobuf.UnknownFieldSet.newBuilder();
+ try {
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ default: {
+ if (!parseUnknownField(input, unknownFields,
+ extensionRegistry, tag)) {
+ done = true;
+ }
+ break;
+ }
+ case 10: {
+ bitField0_ |= 0x00000001;
+ row_ = input.readBytes();
+ break;
+ }
+ case 16: {
+ bitField0_ |= 0x00000002;
+ counter_ = input.readInt32();
+ break;
+ }
+ }
+ }
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(this);
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(
+ e.getMessage()).setUnfinishedMessage(this);
+ } finally {
+ this.unknownFields = unknownFields.build();
+ makeExtensionsImmutable();
+ }
+ }
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorRequest_descriptor;
}
-
+
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorRequest_fieldAccessorTable;
+ return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorRequest_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest.Builder.class);
+ }
+
+ public static com.google.protobuf.Parser<IncCounterProcessorRequest> PARSER =
+ new com.google.protobuf.AbstractParser<IncCounterProcessorRequest>() {
+ public IncCounterProcessorRequest parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new IncCounterProcessorRequest(input, extensionRegistry);
+ }
+ };
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<IncCounterProcessorRequest> getParserForType() {
+ return PARSER;
}
-
+
private int bitField0_;
// required bytes row = 1;
public static final int ROW_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString row_;
+ /**
+ * <code>required bytes row = 1;</code>
+ */
public boolean hasRow() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
+ /**
+ * <code>required bytes row = 1;</code>
+ */
public com.google.protobuf.ByteString getRow() {
return row_;
}
-
+
// required int32 counter = 2;
public static final int COUNTER_FIELD_NUMBER = 2;
private int counter_;
+ /**
+ * <code>required int32 counter = 2;</code>
+ */
public boolean hasCounter() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
+ /**
+ * <code>required int32 counter = 2;</code>
+ */
public int getCounter() {
return counter_;
}
-
+
private void initFields() {
row_ = com.google.protobuf.ByteString.EMPTY;
counter_ = 0;
@@ -76,7 +172,7 @@ public final class IncrementCounterProce
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
-
+
if (!hasRow()) {
memoizedIsInitialized = 0;
return false;
@@ -88,7 +184,7 @@ public final class IncrementCounterProce
memoizedIsInitialized = 1;
return true;
}
-
+
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
@@ -100,12 +196,12 @@ public final class IncrementCounterProce
}
getUnknownFields().writeTo(output);
}
-
+
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
-
+
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
@@ -119,14 +215,14 @@ public final class IncrementCounterProce
memoizedSerializedSize = size;
return size;
}
-
+
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
-
+
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
@@ -136,7 +232,7 @@ public final class IncrementCounterProce
return super.equals(obj);
}
org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest other = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest) obj;
-
+
boolean result = true;
result = result && (hasRow() == other.hasRow());
if (hasRow()) {
@@ -152,9 +248,13 @@ public final class IncrementCounterProce
getUnknownFields().equals(other.getUnknownFields());
return result;
}
-
+
+ private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
+ if (memoizedHashCode != 0) {
+ return memoizedHashCode;
+ }
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasRow()) {
@@ -166,89 +266,79 @@ public final class IncrementCounterProce
hash = (53 * hash) + getCounter();
}
hash = (29 * hash) + getUnknownFields().hashCode();
+ memoizedHashCode = hash;
return hash;
}
-
+
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data).buildParsed();
+ return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data).buildParsed();
+ return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return newBuilder().mergeFrom(input).buildParsed();
+ return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return newBuilder().mergeFrom(input, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- Builder builder = newBuilder();
- if (builder.mergeDelimitedFrom(input)) {
- return builder.buildParsed();
- } else {
- return null;
- }
+ return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- Builder builder = newBuilder();
- if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
- return builder.buildParsed();
- } else {
- return null;
- }
+ return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return newBuilder().mergeFrom(input).buildParsed();
+ return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return newBuilder().mergeFrom(input, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(input, extensionRegistry);
}
-
+
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
-
+
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
+ /**
+ * Protobuf type {@code IncCounterProcessorRequest}
+ */
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder<Builder>
implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequestOrBuilder {
@@ -256,18 +346,21 @@ public final class IncrementCounterProce
getDescriptor() {
return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorRequest_descriptor;
}
-
+
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorRequest_fieldAccessorTable;
+ return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorRequest_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest.Builder.class);
}
-
+
// Construct using org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
-
- private Builder(BuilderParent parent) {
+
+ private Builder(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
@@ -278,7 +371,7 @@ public final class IncrementCounterProce
private static Builder create() {
return new Builder();
}
-
+
public Builder clear() {
super.clear();
row_ = com.google.protobuf.ByteString.EMPTY;
@@ -287,20 +380,20 @@ public final class IncrementCounterProce
bitField0_ = (bitField0_ & ~0x00000002);
return this;
}
-
+
public Builder clone() {
return create().mergeFrom(buildPartial());
}
-
+
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
- return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest.getDescriptor();
+ return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorRequest_descriptor;
}
-
+
public org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest getDefaultInstanceForType() {
return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest.getDefaultInstance();
}
-
+
public org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest build() {
org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest result = buildPartial();
if (!result.isInitialized()) {
@@ -308,17 +401,7 @@ public final class IncrementCounterProce
}
return result;
}
-
- private org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest buildParsed()
- throws com.google.protobuf.InvalidProtocolBufferException {
- org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest result = buildPartial();
- if (!result.isInitialized()) {
- throw newUninitializedMessageException(
- result).asInvalidProtocolBufferException();
- }
- return result;
- }
-
+
public org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest buildPartial() {
org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest result = new org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest(this);
int from_bitField0_ = bitField0_;
@@ -335,7 +418,7 @@ public final class IncrementCounterProce
onBuilt();
return result;
}
-
+
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest) {
return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest)other);
@@ -344,7 +427,7 @@ public final class IncrementCounterProce
return this;
}
}
-
+
public Builder mergeFrom(org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest other) {
if (other == org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest.getDefaultInstance()) return this;
if (other.hasRow()) {
@@ -356,7 +439,7 @@ public final class IncrementCounterProce
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
-
+
public final boolean isInitialized() {
if (!hasRow()) {
@@ -368,54 +451,43 @@ public final class IncrementCounterProce
}
return true;
}
-
+
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder(
- this.getUnknownFields());
- while (true) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
- this.setUnknownFields(unknownFields.build());
- onChanged();
- return this;
- default: {
- if (!parseUnknownField(input, unknownFields,
- extensionRegistry, tag)) {
- this.setUnknownFields(unknownFields.build());
- onChanged();
- return this;
- }
- break;
- }
- case 10: {
- bitField0_ |= 0x00000001;
- row_ = input.readBytes();
- break;
- }
- case 16: {
- bitField0_ |= 0x00000002;
- counter_ = input.readInt32();
- break;
- }
+ org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parsedMessage = null;
+ try {
+ parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest) e.getUnfinishedMessage();
+ throw e;
+ } finally {
+ if (parsedMessage != null) {
+ mergeFrom(parsedMessage);
}
}
+ return this;
}
-
private int bitField0_;
-
+
// required bytes row = 1;
private com.google.protobuf.ByteString row_ = com.google.protobuf.ByteString.EMPTY;
+ /**
+ * <code>required bytes row = 1;</code>
+ */
public boolean hasRow() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
+ /**
+ * <code>required bytes row = 1;</code>
+ */
public com.google.protobuf.ByteString getRow() {
return row_;
}
+ /**
+ * <code>required bytes row = 1;</code>
+ */
public Builder setRow(com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
@@ -425,91 +497,185 @@ public final class IncrementCounterProce
onChanged();
return this;
}
+ /**
+ * <code>required bytes row = 1;</code>
+ */
public Builder clearRow() {
bitField0_ = (bitField0_ & ~0x00000001);
row_ = getDefaultInstance().getRow();
onChanged();
return this;
}
-
+
// required int32 counter = 2;
private int counter_ ;
+ /**
+ * <code>required int32 counter = 2;</code>
+ */
public boolean hasCounter() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
+ /**
+ * <code>required int32 counter = 2;</code>
+ */
public int getCounter() {
return counter_;
}
+ /**
+ * <code>required int32 counter = 2;</code>
+ */
public Builder setCounter(int value) {
bitField0_ |= 0x00000002;
counter_ = value;
onChanged();
return this;
}
+ /**
+ * <code>required int32 counter = 2;</code>
+ */
public Builder clearCounter() {
bitField0_ = (bitField0_ & ~0x00000002);
counter_ = 0;
onChanged();
return this;
}
-
+
// @@protoc_insertion_point(builder_scope:IncCounterProcessorRequest)
}
-
+
static {
defaultInstance = new IncCounterProcessorRequest(true);
defaultInstance.initFields();
}
-
+
// @@protoc_insertion_point(class_scope:IncCounterProcessorRequest)
}
-
+
public interface IncCounterProcessorResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder {
-
+
// required int32 response = 1;
+ /**
+ * <code>required int32 response = 1;</code>
+ */
boolean hasResponse();
+ /**
+ * <code>required int32 response = 1;</code>
+ */
int getResponse();
}
+ /**
+ * Protobuf type {@code IncCounterProcessorResponse}
+ */
public static final class IncCounterProcessorResponse extends
com.google.protobuf.GeneratedMessage
implements IncCounterProcessorResponseOrBuilder {
// Use IncCounterProcessorResponse.newBuilder() to construct.
- private IncCounterProcessorResponse(Builder builder) {
+ private IncCounterProcessorResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
super(builder);
+ this.unknownFields = builder.getUnknownFields();
}
- private IncCounterProcessorResponse(boolean noInit) {}
-
+ private IncCounterProcessorResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
private static final IncCounterProcessorResponse defaultInstance;
public static IncCounterProcessorResponse getDefaultInstance() {
return defaultInstance;
}
-
+
public IncCounterProcessorResponse getDefaultInstanceForType() {
return defaultInstance;
}
-
+
+ private final com.google.protobuf.UnknownFieldSet unknownFields;
+ @java.lang.Override
+ public final com.google.protobuf.UnknownFieldSet
+ getUnknownFields() {
+ return this.unknownFields;
+ }
+ private IncCounterProcessorResponse(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ initFields();
+ int mutable_bitField0_ = 0;
+ com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ com.google.protobuf.UnknownFieldSet.newBuilder();
+ try {
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ default: {
+ if (!parseUnknownField(input, unknownFields,
+ extensionRegistry, tag)) {
+ done = true;
+ }
+ break;
+ }
+ case 8: {
+ bitField0_ |= 0x00000001;
+ response_ = input.readInt32();
+ break;
+ }
+ }
+ }
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(this);
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(
+ e.getMessage()).setUnfinishedMessage(this);
+ } finally {
+ this.unknownFields = unknownFields.build();
+ makeExtensionsImmutable();
+ }
+ }
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorResponse_descriptor;
}
-
+
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorResponse_fieldAccessorTable;
+ return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorResponse_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse.Builder.class);
+ }
+
+ public static com.google.protobuf.Parser<IncCounterProcessorResponse> PARSER =
+ new com.google.protobuf.AbstractParser<IncCounterProcessorResponse>() {
+ public IncCounterProcessorResponse parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new IncCounterProcessorResponse(input, extensionRegistry);
+ }
+ };
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<IncCounterProcessorResponse> getParserForType() {
+ return PARSER;
}
-
+
private int bitField0_;
// required int32 response = 1;
public static final int RESPONSE_FIELD_NUMBER = 1;
private int response_;
+ /**
+ * <code>required int32 response = 1;</code>
+ */
public boolean hasResponse() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
+ /**
+ * <code>required int32 response = 1;</code>
+ */
public int getResponse() {
return response_;
}
-
+
private void initFields() {
response_ = 0;
}
@@ -517,7 +683,7 @@ public final class IncrementCounterProce
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
-
+
if (!hasResponse()) {
memoizedIsInitialized = 0;
return false;
@@ -525,7 +691,7 @@ public final class IncrementCounterProce
memoizedIsInitialized = 1;
return true;
}
-
+
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
@@ -534,12 +700,12 @@ public final class IncrementCounterProce
}
getUnknownFields().writeTo(output);
}
-
+
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
-
+
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
@@ -549,14 +715,14 @@ public final class IncrementCounterProce
memoizedSerializedSize = size;
return size;
}
-
+
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
-
+
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
@@ -566,7 +732,7 @@ public final class IncrementCounterProce
return super.equals(obj);
}
org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse other = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse) obj;
-
+
boolean result = true;
result = result && (hasResponse() == other.hasResponse());
if (hasResponse()) {
@@ -577,9 +743,13 @@ public final class IncrementCounterProce
getUnknownFields().equals(other.getUnknownFields());
return result;
}
-
+
+ private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
+ if (memoizedHashCode != 0) {
+ return memoizedHashCode;
+ }
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasResponse()) {
@@ -587,89 +757,79 @@ public final class IncrementCounterProce
hash = (53 * hash) + getResponse();
}
hash = (29 * hash) + getUnknownFields().hashCode();
+ memoizedHashCode = hash;
return hash;
}
-
+
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data).buildParsed();
+ return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data).buildParsed();
+ return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return newBuilder().mergeFrom(input).buildParsed();
+ return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return newBuilder().mergeFrom(input, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- Builder builder = newBuilder();
- if (builder.mergeDelimitedFrom(input)) {
- return builder.buildParsed();
- } else {
- return null;
- }
+ return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- Builder builder = newBuilder();
- if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
- return builder.buildParsed();
- } else {
- return null;
- }
+ return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return newBuilder().mergeFrom(input).buildParsed();
+ return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return newBuilder().mergeFrom(input, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(input, extensionRegistry);
}
-
+
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
-
+
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
+ /**
+ * Protobuf type {@code IncCounterProcessorResponse}
+ */
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder<Builder>
implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponseOrBuilder {
@@ -677,18 +837,21 @@ public final class IncrementCounterProce
getDescriptor() {
return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorResponse_descriptor;
}
-
+
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorResponse_fieldAccessorTable;
+ return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorResponse_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse.Builder.class);
}
-
+
// Construct using org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
-
- private Builder(BuilderParent parent) {
+
+ private Builder(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
@@ -699,27 +862,27 @@ public final class IncrementCounterProce
private static Builder create() {
return new Builder();
}
-
+
public Builder clear() {
super.clear();
response_ = 0;
bitField0_ = (bitField0_ & ~0x00000001);
return this;
}
-
+
public Builder clone() {
return create().mergeFrom(buildPartial());
}
-
+
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
- return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse.getDescriptor();
+ return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorResponse_descriptor;
}
-
+
public org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse getDefaultInstanceForType() {
return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse.getDefaultInstance();
}
-
+
public org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse build() {
org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse result = buildPartial();
if (!result.isInitialized()) {
@@ -727,17 +890,7 @@ public final class IncrementCounterProce
}
return result;
}
-
- private org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse buildParsed()
- throws com.google.protobuf.InvalidProtocolBufferException {
- org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse result = buildPartial();
- if (!result.isInitialized()) {
- throw newUninitializedMessageException(
- result).asInvalidProtocolBufferException();
- }
- return result;
- }
-
+
public org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse buildPartial() {
org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse result = new org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse(this);
int from_bitField0_ = bitField0_;
@@ -750,7 +903,7 @@ public final class IncrementCounterProce
onBuilt();
return result;
}
-
+
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse) {
return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse)other);
@@ -759,7 +912,7 @@ public final class IncrementCounterProce
return this;
}
}
-
+
public Builder mergeFrom(org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse other) {
if (other == org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse.getDefaultInstance()) return this;
if (other.hasResponse()) {
@@ -768,7 +921,7 @@ public final class IncrementCounterProce
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
-
+
public final boolean isInitialized() {
if (!hasResponse()) {
@@ -776,152 +929,287 @@ public final class IncrementCounterProce
}
return true;
}
-
+
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder(
- this.getUnknownFields());
- while (true) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
- this.setUnknownFields(unknownFields.build());
- onChanged();
- return this;
- default: {
- if (!parseUnknownField(input, unknownFields,
- extensionRegistry, tag)) {
- this.setUnknownFields(unknownFields.build());
- onChanged();
- return this;
- }
- break;
- }
- case 8: {
- bitField0_ |= 0x00000001;
- response_ = input.readInt32();
- break;
- }
+ org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parsedMessage = null;
+ try {
+ parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse) e.getUnfinishedMessage();
+ throw e;
+ } finally {
+ if (parsedMessage != null) {
+ mergeFrom(parsedMessage);
}
}
+ return this;
}
-
private int bitField0_;
-
+
// required int32 response = 1;
private int response_ ;
+ /**
+ * <code>required int32 response = 1;</code>
+ */
public boolean hasResponse() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
+ /**
+ * <code>required int32 response = 1;</code>
+ */
public int getResponse() {
return response_;
}
+ /**
+ * <code>required int32 response = 1;</code>
+ */
public Builder setResponse(int value) {
bitField0_ |= 0x00000001;
response_ = value;
onChanged();
return this;
}
+ /**
+ * <code>required int32 response = 1;</code>
+ */
public Builder clearResponse() {
bitField0_ = (bitField0_ & ~0x00000001);
response_ = 0;
onChanged();
return this;
}
-
+
// @@protoc_insertion_point(builder_scope:IncCounterProcessorResponse)
}
-
+
static {
defaultInstance = new IncCounterProcessorResponse(true);
defaultInstance.initFields();
}
-
+
// @@protoc_insertion_point(class_scope:IncCounterProcessorResponse)
}
-
+
public interface FriendsOfFriendsProcessorRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder {
-
+
// required bytes person = 1;
+ /**
+ * <code>required bytes person = 1;</code>
+ */
boolean hasPerson();
+ /**
+ * <code>required bytes person = 1;</code>
+ */
com.google.protobuf.ByteString getPerson();
-
+
// required bytes row = 2;
+ /**
+ * <code>required bytes row = 2;</code>
+ */
boolean hasRow();
+ /**
+ * <code>required bytes row = 2;</code>
+ */
com.google.protobuf.ByteString getRow();
-
+
// repeated string result = 3;
- java.util.List<String> getResultList();
+ /**
+ * <code>repeated string result = 3;</code>
+ */
+ java.util.List<java.lang.String>
+ getResultList();
+ /**
+ * <code>repeated string result = 3;</code>
+ */
int getResultCount();
- String getResult(int index);
- }
+ /**
+ * <code>repeated string result = 3;</code>
+ */
+ java.lang.String getResult(int index);
+ /**
+ * <code>repeated string result = 3;</code>
+ */
+ com.google.protobuf.ByteString
+ getResultBytes(int index);
+ }
+ /**
+ * Protobuf type {@code FriendsOfFriendsProcessorRequest}
+ */
public static final class FriendsOfFriendsProcessorRequest extends
com.google.protobuf.GeneratedMessage
implements FriendsOfFriendsProcessorRequestOrBuilder {
// Use FriendsOfFriendsProcessorRequest.newBuilder() to construct.
- private FriendsOfFriendsProcessorRequest(Builder builder) {
+ private FriendsOfFriendsProcessorRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
super(builder);
+ this.unknownFields = builder.getUnknownFields();
}
- private FriendsOfFriendsProcessorRequest(boolean noInit) {}
-
+ private FriendsOfFriendsProcessorRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
private static final FriendsOfFriendsProcessorRequest defaultInstance;
public static FriendsOfFriendsProcessorRequest getDefaultInstance() {
return defaultInstance;
}
-
+
public FriendsOfFriendsProcessorRequest getDefaultInstanceForType() {
return defaultInstance;
}
-
+
+ private final com.google.protobuf.UnknownFieldSet unknownFields;
+ @java.lang.Override
+ public final com.google.protobuf.UnknownFieldSet
+ getUnknownFields() {
+ return this.unknownFields;
+ }
+ private FriendsOfFriendsProcessorRequest(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ initFields();
+ int mutable_bitField0_ = 0;
+ com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ com.google.protobuf.UnknownFieldSet.newBuilder();
+ try {
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ default: {
+ if (!parseUnknownField(input, unknownFields,
+ extensionRegistry, tag)) {
+ done = true;
+ }
+ break;
+ }
+ case 10: {
+ bitField0_ |= 0x00000001;
+ person_ = input.readBytes();
+ break;
+ }
+ case 18: {
+ bitField0_ |= 0x00000002;
+ row_ = input.readBytes();
+ break;
+ }
+ case 26: {
+ if (!((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
+ result_ = new com.google.protobuf.LazyStringArrayList();
+ mutable_bitField0_ |= 0x00000004;
+ }
+ result_.add(input.readBytes());
+ break;
+ }
+ }
+ }
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(this);
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(
+ e.getMessage()).setUnfinishedMessage(this);
+ } finally {
+ if (((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
+ result_ = new com.google.protobuf.UnmodifiableLazyStringList(result_);
+ }
+ this.unknownFields = unknownFields.build();
+ makeExtensionsImmutable();
+ }
+ }
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorRequest_descriptor;
}
-
+
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorRequest_fieldAccessorTable;
+ return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorRequest_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest.Builder.class);
+ }
+
+ public static com.google.protobuf.Parser<FriendsOfFriendsProcessorRequest> PARSER =
+ new com.google.protobuf.AbstractParser<FriendsOfFriendsProcessorRequest>() {
+ public FriendsOfFriendsProcessorRequest parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new FriendsOfFriendsProcessorRequest(input, extensionRegistry);
+ }
+ };
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<FriendsOfFriendsProcessorRequest> getParserForType() {
+ return PARSER;
}
-
+
private int bitField0_;
// required bytes person = 1;
public static final int PERSON_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString person_;
+ /**
+ * <code>required bytes person = 1;</code>
+ */
public boolean hasPerson() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
+ /**
+ * <code>required bytes person = 1;</code>
+ */
public com.google.protobuf.ByteString getPerson() {
return person_;
}
-
+
// required bytes row = 2;
public static final int ROW_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString row_;
+ /**
+ * <code>required bytes row = 2;</code>
+ */
public boolean hasRow() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
+ /**
+ * <code>required bytes row = 2;</code>
+ */
public com.google.protobuf.ByteString getRow() {
return row_;
}
-
+
// repeated string result = 3;
public static final int RESULT_FIELD_NUMBER = 3;
private com.google.protobuf.LazyStringList result_;
- public java.util.List<String>
+ /**
+ * <code>repeated string result = 3;</code>
+ */
+ public java.util.List<java.lang.String>
getResultList() {
return result_;
}
+ /**
+ * <code>repeated string result = 3;</code>
+ */
public int getResultCount() {
return result_.size();
}
- public String getResult(int index) {
+ /**
+ * <code>repeated string result = 3;</code>
+ */
+ public java.lang.String getResult(int index) {
return result_.get(index);
}
-
+ /**
+ * <code>repeated string result = 3;</code>
+ */
+ public com.google.protobuf.ByteString
+ getResultBytes(int index) {
+ return result_.getByteString(index);
+ }
+
private void initFields() {
person_ = com.google.protobuf.ByteString.EMPTY;
row_ = com.google.protobuf.ByteString.EMPTY;
@@ -931,7 +1219,7 @@ public final class IncrementCounterProce
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
-
+
if (!hasPerson()) {
memoizedIsInitialized = 0;
return false;
@@ -943,7 +1231,7 @@ public final class IncrementCounterProce
memoizedIsInitialized = 1;
return true;
}
-
+
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
@@ -958,12 +1246,12 @@ public final class IncrementCounterProce
}
getUnknownFields().writeTo(output);
}
-
+
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
-
+
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
@@ -986,14 +1274,14 @@ public final class IncrementCounterProce
memoizedSerializedSize = size;
return size;
}
-
+
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
-
+
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
@@ -1003,7 +1291,7 @@ public final class IncrementCounterProce
return super.equals(obj);
}
org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest other = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest) obj;
-
+
boolean result = true;
result = result && (hasPerson() == other.hasPerson());
if (hasPerson()) {
@@ -1021,9 +1309,13 @@ public final class IncrementCounterProce
getUnknownFields().equals(other.getUnknownFields());
return result;
}
-
+
+ private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
+ if (memoizedHashCode != 0) {
+ return memoizedHashCode;
+ }
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasPerson()) {
@@ -1039,89 +1331,79 @@ public final class IncrementCounterProce
hash = (53 * hash) + getResultList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
+ memoizedHashCode = hash;
return hash;
}
-
+
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data).buildParsed();
+ return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data).buildParsed();
+ return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return newBuilder().mergeFrom(input).buildParsed();
+ return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return newBuilder().mergeFrom(input, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- Builder builder = newBuilder();
- if (builder.mergeDelimitedFrom(input)) {
- return builder.buildParsed();
- } else {
- return null;
- }
+ return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- Builder builder = newBuilder();
- if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
- return builder.buildParsed();
- } else {
- return null;
- }
+ return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return newBuilder().mergeFrom(input).buildParsed();
+ return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return newBuilder().mergeFrom(input, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(input, extensionRegistry);
}
-
+
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
-
+
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
+ /**
+ * Protobuf type {@code FriendsOfFriendsProcessorRequest}
+ */
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder<Builder>
implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequestOrBuilder {
@@ -1129,18 +1411,21 @@ public final class IncrementCounterProce
getDescriptor() {
return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorRequest_descriptor;
}
-
+
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorRequest_fieldAccessorTable;
+ return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorRequest_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest.Builder.class);
}
-
+
// Construct using org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
-
- private Builder(BuilderParent parent) {
+
+ private Builder(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
@@ -1151,7 +1436,7 @@ public final class IncrementCounterProce
private static Builder create() {
return new Builder();
}
-
+
public Builder clear() {
super.clear();
person_ = com.google.protobuf.ByteString.EMPTY;
@@ -1162,20 +1447,20 @@ public final class IncrementCounterProce
bitField0_ = (bitField0_ & ~0x00000004);
return this;
}
-
+
public Builder clone() {
return create().mergeFrom(buildPartial());
}
-
+
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
- return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest.getDescriptor();
+ return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorRequest_descriptor;
}
-
+
public org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest getDefaultInstanceForType() {
return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest.getDefaultInstance();
}
-
+
public org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest build() {
org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest result = buildPartial();
if (!result.isInitialized()) {
@@ -1183,17 +1468,7 @@ public final class IncrementCounterProce
}
return result;
}
-
- private org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest buildParsed()
- throws com.google.protobuf.InvalidProtocolBufferException {
- org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest result = buildPartial();
- if (!result.isInitialized()) {
- throw newUninitializedMessageException(
- result).asInvalidProtocolBufferException();
- }
- return result;
- }
-
+
public org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest buildPartial() {
org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest result = new org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest(this);
int from_bitField0_ = bitField0_;
@@ -1216,7 +1491,7 @@ public final class IncrementCounterProce
onBuilt();
return result;
}
-
+
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest) {
return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest)other);
@@ -1225,7 +1500,7 @@ public final class IncrementCounterProce
return this;
}
}
-
+
public Builder mergeFrom(org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest other) {
if (other == org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest.getDefaultInstance()) return this;
if (other.hasPerson()) {
@@ -1247,7 +1522,7 @@ public final class IncrementCounterProce
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
-
+
public final boolean isInitialized() {
if (!hasPerson()) {
@@ -1259,59 +1534,43 @@ public final class IncrementCounterProce
}
return true;
}
-
+
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder(
- this.getUnknownFields());
- while (true) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
- this.setUnknownFields(unknownFields.build());
- onChanged();
- return this;
- default: {
- if (!parseUnknownField(input, unknownFields,
- extensionRegistry, tag)) {
- this.setUnknownFields(unknownFields.build());
- onChanged();
- return this;
- }
- break;
- }
- case 10: {
- bitField0_ |= 0x00000001;
- person_ = input.readBytes();
- break;
- }
- case 18: {
- bitField0_ |= 0x00000002;
- row_ = input.readBytes();
- break;
- }
- case 26: {
- ensureResultIsMutable();
- result_.add(input.readBytes());
- break;
- }
+ org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parsedMessage = null;
+ try {
+ parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest) e.getUnfinishedMessage();
+ throw e;
+ } finally {
+ if (parsedMessage != null) {
+ mergeFrom(parsedMessage);
}
}
+ return this;
}
-
private int bitField0_;
-
+
// required bytes person = 1;
private com.google.protobuf.ByteString person_ = com.google.protobuf.ByteString.EMPTY;
+ /**
+ * <code>required bytes person = 1;</code>
+ */
public boolean hasPerson() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
+ /**
+ * <code>required bytes person = 1;</code>
+ */
public com.google.protobuf.ByteString getPerson() {
return person_;
}
+ /**
+ * <code>required bytes person = 1;</code>
+ */
public Builder setPerson(com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
@@ -1321,21 +1580,33 @@ public final class IncrementCounterProce
onChanged();
return this;
}
+ /**
+ * <code>required bytes person = 1;</code>
+ */
public Builder clearPerson() {
bitField0_ = (bitField0_ & ~0x00000001);
person_ = getDefaultInstance().getPerson();
onChanged();
return this;
}
-
+
// required bytes row = 2;
private com.google.protobuf.ByteString row_ = com.google.protobuf.ByteString.EMPTY;
+ /**
+ * <code>required bytes row = 2;</code>
+ */
public boolean hasRow() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
+ /**
+ * <code>required bytes row = 2;</code>
+ */
public com.google.protobuf.ByteString getRow() {
return row_;
}
+ /**
+ * <code>required bytes row = 2;</code>
+ */
public Builder setRow(com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
@@ -1345,13 +1616,16 @@ public final class IncrementCounterProce
onChanged();
return this;
}
+ /**
+ * <code>required bytes row = 2;</code>
+ */
public Builder clearRow() {
bitField0_ = (bitField0_ & ~0x00000002);
row_ = getDefaultInstance().getRow();
onChanged();
return this;
}
-
+
// repeated string result = 3;
private com.google.protobuf.LazyStringList result_ = com.google.protobuf.LazyStringArrayList.EMPTY;
private void ensureResultIsMutable() {
@@ -1360,18 +1634,37 @@ public final class IncrementCounterProce
bitField0_ |= 0x00000004;
}
}
- public java.util.List<String>
+ /**
+ * <code>repeated string result = 3;</code>
+ */
+ public java.util.List<java.lang.String>
getResultList() {
return java.util.Collections.unmodifiableList(result_);
}
+ /**
+ * <code>repeated string result = 3;</code>
+ */
public int getResultCount() {
return result_.size();
}
- public String getResult(int index) {
+ /**
+ * <code>repeated string result = 3;</code>
+ */
+ public java.lang.String getResult(int index) {
return result_.get(index);
}
+ /**
+ * <code>repeated string result = 3;</code>
+ */
+ public com.google.protobuf.ByteString
+ getResultBytes(int index) {
+ return result_.getByteString(index);
+ }
+ /**
+ * <code>repeated string result = 3;</code>
+ */
public Builder setResult(
- int index, String value) {
+ int index, java.lang.String value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1380,7 +1673,11 @@ public final class IncrementCounterProce
onChanged();
return this;
}
- public Builder addResult(String value) {
+ /**
+ * <code>repeated string result = 3;</code>
+ */
+ public Builder addResult(
+ java.lang.String value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1389,86 +1686,204 @@ public final class IncrementCounterProce
onChanged();
return this;
}
+ /**
+ * <code>repeated string result = 3;</code>
+ */
public Builder addAllResult(
- java.lang.Iterable<String> values) {
+ java.lang.Iterable<java.lang.String> values) {
ensureResultIsMutable();
super.addAll(values, result_);
onChanged();
return this;
}
+ /**
+ * <code>repeated string result = 3;</code>
+ */
public Builder clearResult() {
result_ = com.google.protobuf.LazyStringArrayList.EMPTY;
bitField0_ = (bitField0_ & ~0x00000004);
onChanged();
return this;
}
- void addResult(com.google.protobuf.ByteString value) {
- ensureResultIsMutable();
+ /**
+ * <code>repeated string result = 3;</code>
+ */
+ public Builder addResultBytes(
+ com.google.protobuf.ByteString value) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ ensureResultIsMutable();
result_.add(value);
onChanged();
+ return this;
}
-
+
// @@protoc_insertion_point(builder_scope:FriendsOfFriendsProcessorRequest)
}
-
+
static {
defaultInstance = new FriendsOfFriendsProcessorRequest(true);
defaultInstance.initFields();
}
-
+
// @@protoc_insertion_point(class_scope:FriendsOfFriendsProcessorRequest)
}
-
+
public interface FriendsOfFriendsProcessorResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder {
-
+
// repeated string result = 1;
- java.util.List<String> getResultList();
+ /**
+ * <code>repeated string result = 1;</code>
+ */
+ java.util.List<java.lang.String>
+ getResultList();
+ /**
+ * <code>repeated string result = 1;</code>
+ */
int getResultCount();
- String getResult(int index);
- }
+ /**
+ * <code>repeated string result = 1;</code>
+ */
+ java.lang.String getResult(int index);
+ /**
+ * <code>repeated string result = 1;</code>
+ */
+ com.google.protobuf.ByteString
+ getResultBytes(int index);
+ }
+ /**
+ * Protobuf type {@code FriendsOfFriendsProcessorResponse}
+ */
public static final class FriendsOfFriendsProcessorResponse extends
com.google.protobuf.GeneratedMessage
implements FriendsOfFriendsProcessorResponseOrBuilder {
// Use FriendsOfFriendsProcessorResponse.newBuilder() to construct.
- private FriendsOfFriendsProcessorResponse(Builder builder) {
+ private FriendsOfFriendsProcessorResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
super(builder);
+ this.unknownFields = builder.getUnknownFields();
}
- private FriendsOfFriendsProcessorResponse(boolean noInit) {}
-
+ private FriendsOfFriendsProcessorResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
private static final FriendsOfFriendsProcessorResponse defaultInstance;
public static FriendsOfFriendsProcessorResponse getDefaultInstance() {
return defaultInstance;
}
-
+
public FriendsOfFriendsProcessorResponse getDefaultInstanceForType() {
return defaultInstance;
}
-
+
+ private final com.google.protobuf.UnknownFieldSet unknownFields;
+ @java.lang.Override
+ public final com.google.protobuf.UnknownFieldSet
+ getUnknownFields() {
+ return this.unknownFields;
+ }
+ private FriendsOfFriendsProcessorResponse(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ initFields();
+ int mutable_bitField0_ = 0;
+ com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ com.google.protobuf.UnknownFieldSet.newBuilder();
+ try {
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ default: {
+ if (!parseUnknownField(input, unknownFields,
+ extensionRegistry, tag)) {
+ done = true;
+ }
+ break;
+ }
+ case 10: {
+ if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
+ result_ = new com.google.protobuf.LazyStringArrayList();
+ mutable_bitField0_ |= 0x00000001;
+ }
+ result_.add(input.readBytes());
+ break;
+ }
+ }
+ }
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(this);
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(
+ e.getMessage()).setUnfinishedMessage(this);
+ } finally {
+ if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
+ result_ = new com.google.protobuf.UnmodifiableLazyStringList(result_);
+ }
+ this.unknownFields = unknownFields.build();
+ makeExtensionsImmutable();
+ }
+ }
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorResponse_descriptor;
}
-
+
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorResponse_fieldAccessorTable;
+ return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorResponse_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse.Builder.class);
+ }
+
+ public static com.google.protobuf.Parser<FriendsOfFriendsProcessorResponse> PARSER =
+ new com.google.protobuf.AbstractParser<FriendsOfFriendsProcessorResponse>() {
+ public FriendsOfFriendsProcessorResponse parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new FriendsOfFriendsProcessorResponse(input, extensionRegistry);
+ }
+ };
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<FriendsOfFriendsProcessorResponse> getParserForType() {
+ return PARSER;
}
-
+
// repeated string result = 1;
public static final int RESULT_FIELD_NUMBER = 1;
private com.google.protobuf.LazyStringList result_;
- public java.util.List<String>
+ /**
+ * <code>repeated string result = 1;</code>
+ */
+ public java.util.List<java.lang.String>
getResultList() {
return result_;
}
+ /**
+ * <code>repeated string result = 1;</code>
+ */
public int getResultCount() {
return result_.size();
}
- public String getResult(int index) {
+ /**
+ * <code>repeated string result = 1;</code>
+ */
+ public java.lang.String getResult(int index) {
return result_.get(index);
}
-
+ /**
+ * <code>repeated string result = 1;</code>
+ */
+ public com.google.protobuf.ByteString
+ getResultBytes(int index) {
+ return result_.getByteString(index);
+ }
+
private void initFields() {
result_ = com.google.protobuf.LazyStringArrayList.EMPTY;
}
@@ -1476,11 +1891,11 @@ public final class IncrementCounterProce
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
-
+
memoizedIsInitialized = 1;
return true;
}
-
+
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
@@ -1489,12 +1904,12 @@ public final class IncrementCounterProce
}
getUnknownFields().writeTo(output);
}
-
+
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
-
+
size = 0;
{
int dataSize = 0;
@@ -1509,14 +1924,14 @@ public final class IncrementCounterProce
memoizedSerializedSize = size;
return size;
}
-
+
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
-
+
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
@@ -1526,7 +1941,7 @@ public final class IncrementCounterProce
return super.equals(obj);
}
org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse other = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse) obj;
-
+
boolean result = true;
result = result && getResultList()
.equals(other.getResultList());
@@ -1534,9 +1949,13 @@ public final class IncrementCounterProce
getUnknownFields().equals(other.getUnknownFields());
return result;
}
-
+
+ private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
+ if (memoizedHashCode != 0) {
+ return memoizedHashCode;
+ }
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (getResultCount() > 0) {
@@ -1544,89 +1963,79 @@ public final class IncrementCounterProce
hash = (53 * hash) + getResultList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
+ memoizedHashCode = hash;
return hash;
}
-
+
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data).buildParsed();
+ return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data).buildParsed();
+ return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return newBuilder().mergeFrom(input).buildParsed();
+ return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return newBuilder().mergeFrom(input, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- Builder builder = newBuilder();
- if (builder.mergeDelimitedFrom(input)) {
- return builder.buildParsed();
- } else {
- return null;
- }
+ return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- Builder builder = newBuilder();
- if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
- return builder.buildParsed();
- } else {
- return null;
- }
+ return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return newBuilder().mergeFrom(input).buildParsed();
+ return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return newBuilder().mergeFrom(input, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(input, extensionRegistry);
}
-
+
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
-
+
[... 2349 lines stripped ...]