You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2012/07/30 09:07:18 UTC
svn commit: r1367009 [2/4] - in /hbase/trunk/hbase-server/src:
main/java/org/apache/hadoop/hbase/ipc/
main/java/org/apache/hadoop/hbase/master/
main/java/org/apache/hadoop/hbase/monitoring/
main/java/org/apache/hadoop/hbase/protobuf/generated/ main/jav...
Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RPCProtos.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RPCProtos.java?rev=1367009&r1=1367008&r2=1367009&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RPCProtos.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RPCProtos.java Mon Jul 30 07:07:17 2012
@@ -1099,47 +1099,43 @@ public final class RPCProtos {
// @@protoc_insertion_point(class_scope:ConnectionHeader)
}
- public interface RpcRequestOrBuilder
+ public interface RpcRequestHeaderOrBuilder
extends com.google.protobuf.MessageOrBuilder {
- // required int32 callId = 1;
+ // required uint32 callId = 1;
boolean hasCallId();
int getCallId();
-
- // optional bytes request = 2;
- boolean hasRequest();
- com.google.protobuf.ByteString getRequest();
}
- public static final class RpcRequest extends
+ public static final class RpcRequestHeader extends
com.google.protobuf.GeneratedMessage
- implements RpcRequestOrBuilder {
- // Use RpcRequest.newBuilder() to construct.
- private RpcRequest(Builder builder) {
+ implements RpcRequestHeaderOrBuilder {
+ // Use RpcRequestHeader.newBuilder() to construct.
+ private RpcRequestHeader(Builder builder) {
super(builder);
}
- private RpcRequest(boolean noInit) {}
+ private RpcRequestHeader(boolean noInit) {}
- private static final RpcRequest defaultInstance;
- public static RpcRequest getDefaultInstance() {
+ private static final RpcRequestHeader defaultInstance;
+ public static RpcRequestHeader getDefaultInstance() {
return defaultInstance;
}
- public RpcRequest getDefaultInstanceForType() {
+ public RpcRequestHeader getDefaultInstanceForType() {
return defaultInstance;
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
- return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcRequest_descriptor;
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcRequestHeader_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcRequest_fieldAccessorTable;
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcRequestHeader_fieldAccessorTable;
}
private int bitField0_;
- // required int32 callId = 1;
+ // required uint32 callId = 1;
public static final int CALLID_FIELD_NUMBER = 1;
private int callId_;
public boolean hasCallId() {
@@ -1149,19 +1145,8 @@ public final class RPCProtos {
return callId_;
}
- // optional bytes request = 2;
- public static final int REQUEST_FIELD_NUMBER = 2;
- private com.google.protobuf.ByteString request_;
- public boolean hasRequest() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
- }
- public com.google.protobuf.ByteString getRequest() {
- return request_;
- }
-
private void initFields() {
callId_ = 0;
- request_ = com.google.protobuf.ByteString.EMPTY;
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
@@ -1180,10 +1165,7 @@ public final class RPCProtos {
throws java.io.IOException {
getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- output.writeInt32(1, callId_);
- }
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
- output.writeBytes(2, request_);
+ output.writeUInt32(1, callId_);
}
getUnknownFields().writeTo(output);
}
@@ -1196,11 +1178,7 @@ public final class RPCProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
- .computeInt32Size(1, callId_);
- }
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
- .computeBytesSize(2, request_);
+ .computeUInt32Size(1, callId_);
}
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
@@ -1219,10 +1197,10 @@ public final class RPCProtos {
if (obj == this) {
return true;
}
- if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest)) {
+ if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader)) {
return super.equals(obj);
}
- org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest other = (org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest) obj;
+ org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader other = (org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader) obj;
boolean result = true;
result = result && (hasCallId() == other.hasCallId());
@@ -1230,11 +1208,6 @@ public final class RPCProtos {
result = result && (getCallId()
== other.getCallId());
}
- result = result && (hasRequest() == other.hasRequest());
- if (hasRequest()) {
- result = result && getRequest()
- .equals(other.getRequest());
- }
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
@@ -1248,49 +1221,45 @@ public final class RPCProtos {
hash = (37 * hash) + CALLID_FIELD_NUMBER;
hash = (53 * hash) + getCallId();
}
- if (hasRequest()) {
- hash = (37 * hash) + REQUEST_FIELD_NUMBER;
- hash = (53 * hash) + getRequest().hashCode();
- }
hash = (29 * hash) + getUnknownFields().hashCode();
return hash;
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest parseFrom(
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data).buildParsed();
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest parseFrom(
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data, extensionRegistry)
.buildParsed();
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest parseFrom(byte[] data)
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data).buildParsed();
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest parseFrom(
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data, extensionRegistry)
.buildParsed();
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest parseFrom(java.io.InputStream input)
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader parseFrom(java.io.InputStream input)
throws java.io.IOException {
return newBuilder().mergeFrom(input).buildParsed();
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest parseFrom(
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return newBuilder().mergeFrom(input, extensionRegistry)
.buildParsed();
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest parseDelimitedFrom(java.io.InputStream input)
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
Builder builder = newBuilder();
if (builder.mergeDelimitedFrom(input)) {
@@ -1299,7 +1268,7 @@ public final class RPCProtos {
return null;
}
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest parseDelimitedFrom(
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
@@ -1310,12 +1279,12 @@ public final class RPCProtos {
return null;
}
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest parseFrom(
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return newBuilder().mergeFrom(input).buildParsed();
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest parseFrom(
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
@@ -1325,7 +1294,7 @@ public final class RPCProtos {
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
- public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest prototype) {
+ public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@@ -1338,18 +1307,18 @@ public final class RPCProtos {
}
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestOrBuilder {
+ implements org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeaderOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
- return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcRequest_descriptor;
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcRequestHeader_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcRequest_fieldAccessorTable;
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcRequestHeader_fieldAccessorTable;
}
- // Construct using org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest.newBuilder()
+ // Construct using org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
@@ -1370,8 +1339,6 @@ public final class RPCProtos {
super.clear();
callId_ = 0;
bitField0_ = (bitField0_ & ~0x00000001);
- request_ = com.google.protobuf.ByteString.EMPTY;
- bitField0_ = (bitField0_ & ~0x00000002);
return this;
}
@@ -1381,24 +1348,24 @@ public final class RPCProtos {
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
- return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest.getDescriptor();
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader.getDescriptor();
}
- public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest getDefaultInstanceForType() {
- return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest.getDefaultInstance();
+ public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader getDefaultInstanceForType() {
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader.getDefaultInstance();
}
- public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest build() {
- org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest result = buildPartial();
+ public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader build() {
+ org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
- private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest buildParsed()
+ private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader buildParsed()
throws com.google.protobuf.InvalidProtocolBufferException {
- org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest result = buildPartial();
+ org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(
result).asInvalidProtocolBufferException();
@@ -1406,40 +1373,33 @@ public final class RPCProtos {
return result;
}
- public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest buildPartial() {
- org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest result = new org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest(this);
+ public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader buildPartial() {
+ org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader result = new org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader(this);
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
to_bitField0_ |= 0x00000001;
}
result.callId_ = callId_;
- if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
- to_bitField0_ |= 0x00000002;
- }
- result.request_ = request_;
result.bitField0_ = to_bitField0_;
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
- if (other instanceof org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest) {
- return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest)other);
+ if (other instanceof org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader) {
+ return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader)other);
} else {
super.mergeFrom(other);
return this;
}
}
- public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest other) {
- if (other == org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequest.getDefaultInstance()) return this;
+ public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader other) {
+ if (other == org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestHeader.getDefaultInstance()) return this;
if (other.hasCallId()) {
setCallId(other.getCallId());
}
- if (other.hasRequest()) {
- setRequest(other.getRequest());
- }
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
@@ -1477,12 +1437,7 @@ public final class RPCProtos {
}
case 8: {
bitField0_ |= 0x00000001;
- callId_ = input.readInt32();
- break;
- }
- case 18: {
- bitField0_ |= 0x00000002;
- request_ = input.readBytes();
+ callId_ = input.readUInt32();
break;
}
}
@@ -1491,7 +1446,7 @@ public final class RPCProtos {
private int bitField0_;
- // required int32 callId = 1;
+ // required uint32 callId = 1;
private int callId_ ;
public boolean hasCallId() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@@ -1512,89 +1467,69 @@ public final class RPCProtos {
return this;
}
- // optional bytes request = 2;
- private com.google.protobuf.ByteString request_ = com.google.protobuf.ByteString.EMPTY;
- public boolean hasRequest() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
- }
- public com.google.protobuf.ByteString getRequest() {
- return request_;
- }
- public Builder setRequest(com.google.protobuf.ByteString value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000002;
- request_ = value;
- onChanged();
- return this;
- }
- public Builder clearRequest() {
- bitField0_ = (bitField0_ & ~0x00000002);
- request_ = getDefaultInstance().getRequest();
- onChanged();
- return this;
- }
-
- // @@protoc_insertion_point(builder_scope:RpcRequest)
+ // @@protoc_insertion_point(builder_scope:RpcRequestHeader)
}
static {
- defaultInstance = new RpcRequest(true);
+ defaultInstance = new RpcRequestHeader(true);
defaultInstance.initFields();
}
- // @@protoc_insertion_point(class_scope:RpcRequest)
+ // @@protoc_insertion_point(class_scope:RpcRequestHeader)
}
- public interface RpcExceptionOrBuilder
+ public interface RpcRequestBodyOrBuilder
extends com.google.protobuf.MessageOrBuilder {
- // required string exceptionName = 1;
- boolean hasExceptionName();
- String getExceptionName();
+ // required string methodName = 1;
+ boolean hasMethodName();
+ String getMethodName();
+
+ // optional uint64 clientProtocolVersion = 2;
+ boolean hasClientProtocolVersion();
+ long getClientProtocolVersion();
- // optional string stackTrace = 2;
- boolean hasStackTrace();
- String getStackTrace();
+ // optional bytes request = 3;
+ boolean hasRequest();
+ com.google.protobuf.ByteString getRequest();
}
- public static final class RpcException extends
+ public static final class RpcRequestBody extends
com.google.protobuf.GeneratedMessage
- implements RpcExceptionOrBuilder {
- // Use RpcException.newBuilder() to construct.
- private RpcException(Builder builder) {
+ implements RpcRequestBodyOrBuilder {
+ // Use RpcRequestBody.newBuilder() to construct.
+ private RpcRequestBody(Builder builder) {
super(builder);
}
- private RpcException(boolean noInit) {}
+ private RpcRequestBody(boolean noInit) {}
- private static final RpcException defaultInstance;
- public static RpcException getDefaultInstance() {
+ private static final RpcRequestBody defaultInstance;
+ public static RpcRequestBody getDefaultInstance() {
return defaultInstance;
}
- public RpcException getDefaultInstanceForType() {
+ public RpcRequestBody getDefaultInstanceForType() {
return defaultInstance;
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
- return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcException_descriptor;
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcRequestBody_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcException_fieldAccessorTable;
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcRequestBody_fieldAccessorTable;
}
private int bitField0_;
- // required string exceptionName = 1;
- public static final int EXCEPTIONNAME_FIELD_NUMBER = 1;
- private java.lang.Object exceptionName_;
- public boolean hasExceptionName() {
+ // required string methodName = 1;
+ public static final int METHODNAME_FIELD_NUMBER = 1;
+ private java.lang.Object methodName_;
+ public boolean hasMethodName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
- public String getExceptionName() {
- java.lang.Object ref = exceptionName_;
+ public String getMethodName() {
+ java.lang.Object ref = methodName_;
if (ref instanceof String) {
return (String) ref;
} else {
@@ -1602,65 +1537,54 @@ public final class RPCProtos {
(com.google.protobuf.ByteString) ref;
String s = bs.toStringUtf8();
if (com.google.protobuf.Internal.isValidUtf8(bs)) {
- exceptionName_ = s;
+ methodName_ = s;
}
return s;
}
}
- private com.google.protobuf.ByteString getExceptionNameBytes() {
- java.lang.Object ref = exceptionName_;
+ private com.google.protobuf.ByteString getMethodNameBytes() {
+ java.lang.Object ref = methodName_;
if (ref instanceof String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8((String) ref);
- exceptionName_ = b;
+ methodName_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
- // optional string stackTrace = 2;
- public static final int STACKTRACE_FIELD_NUMBER = 2;
- private java.lang.Object stackTrace_;
- public boolean hasStackTrace() {
+ // optional uint64 clientProtocolVersion = 2;
+ public static final int CLIENTPROTOCOLVERSION_FIELD_NUMBER = 2;
+ private long clientProtocolVersion_;
+ public boolean hasClientProtocolVersion() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
- public String getStackTrace() {
- java.lang.Object ref = stackTrace_;
- if (ref instanceof String) {
- return (String) ref;
- } else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
- String s = bs.toStringUtf8();
- if (com.google.protobuf.Internal.isValidUtf8(bs)) {
- stackTrace_ = s;
- }
- return s;
- }
+ public long getClientProtocolVersion() {
+ return clientProtocolVersion_;
}
- private com.google.protobuf.ByteString getStackTraceBytes() {
- java.lang.Object ref = stackTrace_;
- if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8((String) ref);
- stackTrace_ = b;
- return b;
- } else {
- return (com.google.protobuf.ByteString) ref;
- }
+
+ // optional bytes request = 3;
+ public static final int REQUEST_FIELD_NUMBER = 3;
+ private com.google.protobuf.ByteString request_;
+ public boolean hasRequest() {
+ return ((bitField0_ & 0x00000004) == 0x00000004);
+ }
+ public com.google.protobuf.ByteString getRequest() {
+ return request_;
}
private void initFields() {
- exceptionName_ = "";
- stackTrace_ = "";
+ methodName_ = "";
+ clientProtocolVersion_ = 0L;
+ request_ = com.google.protobuf.ByteString.EMPTY;
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
- if (!hasExceptionName()) {
+ if (!hasMethodName()) {
memoizedIsInitialized = 0;
return false;
}
@@ -1672,10 +1596,13 @@ public final class RPCProtos {
throws java.io.IOException {
getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- output.writeBytes(1, getExceptionNameBytes());
+ output.writeBytes(1, getMethodNameBytes());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- output.writeBytes(2, getStackTraceBytes());
+ output.writeUInt64(2, clientProtocolVersion_);
+ }
+ if (((bitField0_ & 0x00000004) == 0x00000004)) {
+ output.writeBytes(3, request_);
}
getUnknownFields().writeTo(output);
}
@@ -1688,11 +1615,15 @@ public final class RPCProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
- .computeBytesSize(1, getExceptionNameBytes());
+ .computeBytesSize(1, getMethodNameBytes());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
size += com.google.protobuf.CodedOutputStream
- .computeBytesSize(2, getStackTraceBytes());
+ .computeUInt64Size(2, clientProtocolVersion_);
+ }
+ if (((bitField0_ & 0x00000004) == 0x00000004)) {
+ size += com.google.protobuf.CodedOutputStream
+ .computeBytesSize(3, request_);
}
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
@@ -1711,21 +1642,26 @@ public final class RPCProtos {
if (obj == this) {
return true;
}
- if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException)) {
+ if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody)) {
return super.equals(obj);
}
- org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException other = (org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException) obj;
+ org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody other = (org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody) obj;
boolean result = true;
- result = result && (hasExceptionName() == other.hasExceptionName());
- if (hasExceptionName()) {
- result = result && getExceptionName()
- .equals(other.getExceptionName());
+ result = result && (hasMethodName() == other.hasMethodName());
+ if (hasMethodName()) {
+ result = result && getMethodName()
+ .equals(other.getMethodName());
+ }
+ result = result && (hasClientProtocolVersion() == other.hasClientProtocolVersion());
+ if (hasClientProtocolVersion()) {
+ result = result && (getClientProtocolVersion()
+ == other.getClientProtocolVersion());
}
- result = result && (hasStackTrace() == other.hasStackTrace());
- if (hasStackTrace()) {
- result = result && getStackTrace()
- .equals(other.getStackTrace());
+ result = result && (hasRequest() == other.hasRequest());
+ if (hasRequest()) {
+ result = result && getRequest()
+ .equals(other.getRequest());
}
result = result &&
getUnknownFields().equals(other.getUnknownFields());
@@ -1736,53 +1672,57 @@ public final class RPCProtos {
public int hashCode() {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
- if (hasExceptionName()) {
- hash = (37 * hash) + EXCEPTIONNAME_FIELD_NUMBER;
- hash = (53 * hash) + getExceptionName().hashCode();
+ if (hasMethodName()) {
+ hash = (37 * hash) + METHODNAME_FIELD_NUMBER;
+ hash = (53 * hash) + getMethodName().hashCode();
+ }
+ if (hasClientProtocolVersion()) {
+ hash = (37 * hash) + CLIENTPROTOCOLVERSION_FIELD_NUMBER;
+ hash = (53 * hash) + hashLong(getClientProtocolVersion());
}
- if (hasStackTrace()) {
- hash = (37 * hash) + STACKTRACE_FIELD_NUMBER;
- hash = (53 * hash) + getStackTrace().hashCode();
+ if (hasRequest()) {
+ hash = (37 * hash) + REQUEST_FIELD_NUMBER;
+ hash = (53 * hash) + getRequest().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
return hash;
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException parseFrom(
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data).buildParsed();
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException parseFrom(
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data, extensionRegistry)
.buildParsed();
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException parseFrom(byte[] data)
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data).buildParsed();
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException parseFrom(
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data, extensionRegistry)
.buildParsed();
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException parseFrom(java.io.InputStream input)
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody parseFrom(java.io.InputStream input)
throws java.io.IOException {
return newBuilder().mergeFrom(input).buildParsed();
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException parseFrom(
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return newBuilder().mergeFrom(input, extensionRegistry)
.buildParsed();
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException parseDelimitedFrom(java.io.InputStream input)
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
Builder builder = newBuilder();
if (builder.mergeDelimitedFrom(input)) {
@@ -1791,7 +1731,7 @@ public final class RPCProtos {
return null;
}
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException parseDelimitedFrom(
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
@@ -1802,12 +1742,12 @@ public final class RPCProtos {
return null;
}
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException parseFrom(
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return newBuilder().mergeFrom(input).buildParsed();
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException parseFrom(
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
@@ -1817,7 +1757,7 @@ public final class RPCProtos {
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
- public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException prototype) {
+ public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@@ -1830,18 +1770,18 @@ public final class RPCProtos {
}
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcExceptionOrBuilder {
+ implements org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBodyOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
- return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcException_descriptor;
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcRequestBody_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcException_fieldAccessorTable;
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcRequestBody_fieldAccessorTable;
}
- // Construct using org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException.newBuilder()
+ // Construct using org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
@@ -1860,10 +1800,12 @@ public final class RPCProtos {
public Builder clear() {
super.clear();
- exceptionName_ = "";
+ methodName_ = "";
bitField0_ = (bitField0_ & ~0x00000001);
- stackTrace_ = "";
+ clientProtocolVersion_ = 0L;
bitField0_ = (bitField0_ & ~0x00000002);
+ request_ = com.google.protobuf.ByteString.EMPTY;
+ bitField0_ = (bitField0_ & ~0x00000004);
return this;
}
@@ -1873,24 +1815,24 @@ public final class RPCProtos {
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
- return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException.getDescriptor();
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody.getDescriptor();
}
- public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException getDefaultInstanceForType() {
- return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException.getDefaultInstance();
+ public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody getDefaultInstanceForType() {
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody.getDefaultInstance();
}
- public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException build() {
- org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException result = buildPartial();
+ public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody build() {
+ org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
- private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException buildParsed()
+ private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody buildParsed()
throws com.google.protobuf.InvalidProtocolBufferException {
- org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException result = buildPartial();
+ org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(
result).asInvalidProtocolBufferException();
@@ -1898,46 +1840,53 @@ public final class RPCProtos {
return result;
}
- public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException buildPartial() {
- org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException result = new org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException(this);
+ public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody buildPartial() {
+ org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody result = new org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody(this);
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
to_bitField0_ |= 0x00000001;
}
- result.exceptionName_ = exceptionName_;
+ result.methodName_ = methodName_;
if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
to_bitField0_ |= 0x00000002;
}
- result.stackTrace_ = stackTrace_;
+ result.clientProtocolVersion_ = clientProtocolVersion_;
+ if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
+ to_bitField0_ |= 0x00000004;
+ }
+ result.request_ = request_;
result.bitField0_ = to_bitField0_;
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
- if (other instanceof org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException) {
- return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException)other);
+ if (other instanceof org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody) {
+ return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody)other);
} else {
super.mergeFrom(other);
return this;
}
}
- public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException other) {
- if (other == org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException.getDefaultInstance()) return this;
- if (other.hasExceptionName()) {
- setExceptionName(other.getExceptionName());
+ public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody other) {
+ if (other == org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcRequestBody.getDefaultInstance()) return this;
+ if (other.hasMethodName()) {
+ setMethodName(other.getMethodName());
}
- if (other.hasStackTrace()) {
- setStackTrace(other.getStackTrace());
+ if (other.hasClientProtocolVersion()) {
+ setClientProtocolVersion(other.getClientProtocolVersion());
+ }
+ if (other.hasRequest()) {
+ setRequest(other.getRequest());
}
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public final boolean isInitialized() {
- if (!hasExceptionName()) {
+ if (!hasMethodName()) {
return false;
}
@@ -1969,12 +1918,17 @@ public final class RPCProtos {
}
case 10: {
bitField0_ |= 0x00000001;
- exceptionName_ = input.readBytes();
+ methodName_ = input.readBytes();
break;
}
- case 18: {
+ case 16: {
bitField0_ |= 0x00000002;
- stackTrace_ = input.readBytes();
+ clientProtocolVersion_ = input.readUInt64();
+ break;
+ }
+ case 26: {
+ bitField0_ |= 0x00000004;
+ request_ = input.readBytes();
break;
}
}
@@ -1983,135 +1937,135 @@ public final class RPCProtos {
private int bitField0_;
- // required string exceptionName = 1;
- private java.lang.Object exceptionName_ = "";
- public boolean hasExceptionName() {
+ // required string methodName = 1;
+ private java.lang.Object methodName_ = "";
+ public boolean hasMethodName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
- public String getExceptionName() {
- java.lang.Object ref = exceptionName_;
+ public String getMethodName() {
+ java.lang.Object ref = methodName_;
if (!(ref instanceof String)) {
String s = ((com.google.protobuf.ByteString) ref).toStringUtf8();
- exceptionName_ = s;
+ methodName_ = s;
return s;
} else {
return (String) ref;
}
}
- public Builder setExceptionName(String value) {
+ public Builder setMethodName(String value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000001;
- exceptionName_ = value;
+ methodName_ = value;
onChanged();
return this;
}
- public Builder clearExceptionName() {
+ public Builder clearMethodName() {
bitField0_ = (bitField0_ & ~0x00000001);
- exceptionName_ = getDefaultInstance().getExceptionName();
+ methodName_ = getDefaultInstance().getMethodName();
onChanged();
return this;
}
- void setExceptionName(com.google.protobuf.ByteString value) {
+ void setMethodName(com.google.protobuf.ByteString value) {
bitField0_ |= 0x00000001;
- exceptionName_ = value;
+ methodName_ = value;
onChanged();
}
- // optional string stackTrace = 2;
- private java.lang.Object stackTrace_ = "";
- public boolean hasStackTrace() {
+ // optional uint64 clientProtocolVersion = 2;
+ private long clientProtocolVersion_ ;
+ public boolean hasClientProtocolVersion() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
- public String getStackTrace() {
- java.lang.Object ref = stackTrace_;
- if (!(ref instanceof String)) {
- String s = ((com.google.protobuf.ByteString) ref).toStringUtf8();
- stackTrace_ = s;
- return s;
- } else {
- return (String) ref;
- }
+ public long getClientProtocolVersion() {
+ return clientProtocolVersion_;
}
- public Builder setStackTrace(String value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000002;
- stackTrace_ = value;
+ public Builder setClientProtocolVersion(long value) {
+ bitField0_ |= 0x00000002;
+ clientProtocolVersion_ = value;
onChanged();
return this;
}
- public Builder clearStackTrace() {
+ public Builder clearClientProtocolVersion() {
bitField0_ = (bitField0_ & ~0x00000002);
- stackTrace_ = getDefaultInstance().getStackTrace();
+ clientProtocolVersion_ = 0L;
onChanged();
return this;
}
- void setStackTrace(com.google.protobuf.ByteString value) {
- bitField0_ |= 0x00000002;
- stackTrace_ = value;
+
+ // optional bytes request = 3;
+ private com.google.protobuf.ByteString request_ = com.google.protobuf.ByteString.EMPTY;
+ public boolean hasRequest() {
+ return ((bitField0_ & 0x00000004) == 0x00000004);
+ }
+ public com.google.protobuf.ByteString getRequest() {
+ return request_;
+ }
+ public Builder setRequest(com.google.protobuf.ByteString value) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ bitField0_ |= 0x00000004;
+ request_ = value;
+ onChanged();
+ return this;
+ }
+ public Builder clearRequest() {
+ bitField0_ = (bitField0_ & ~0x00000004);
+ request_ = getDefaultInstance().getRequest();
onChanged();
+ return this;
}
- // @@protoc_insertion_point(builder_scope:RpcException)
+ // @@protoc_insertion_point(builder_scope:RpcRequestBody)
}
static {
- defaultInstance = new RpcException(true);
+ defaultInstance = new RpcRequestBody(true);
defaultInstance.initFields();
}
- // @@protoc_insertion_point(class_scope:RpcException)
+ // @@protoc_insertion_point(class_scope:RpcRequestBody)
}
- public interface RpcResponseOrBuilder
+ public interface RpcResponseHeaderOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required int32 callId = 1;
boolean hasCallId();
int getCallId();
- // required .RpcResponse.Status status = 2;
+ // required .RpcResponseHeader.Status status = 2;
boolean hasStatus();
- org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse.Status getStatus();
-
- // optional bytes response = 3;
- boolean hasResponse();
- com.google.protobuf.ByteString getResponse();
-
- // optional .RpcException exception = 4;
- boolean hasException();
- org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException getException();
- org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcExceptionOrBuilder getExceptionOrBuilder();
+ org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader.Status getStatus();
}
- public static final class RpcResponse extends
+ public static final class RpcResponseHeader extends
com.google.protobuf.GeneratedMessage
- implements RpcResponseOrBuilder {
- // Use RpcResponse.newBuilder() to construct.
- private RpcResponse(Builder builder) {
+ implements RpcResponseHeaderOrBuilder {
+ // Use RpcResponseHeader.newBuilder() to construct.
+ private RpcResponseHeader(Builder builder) {
super(builder);
}
- private RpcResponse(boolean noInit) {}
+ private RpcResponseHeader(boolean noInit) {}
- private static final RpcResponse defaultInstance;
- public static RpcResponse getDefaultInstance() {
+ private static final RpcResponseHeader defaultInstance;
+ public static RpcResponseHeader getDefaultInstance() {
return defaultInstance;
}
- public RpcResponse getDefaultInstanceForType() {
+ public RpcResponseHeader getDefaultInstanceForType() {
return defaultInstance;
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
- return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcResponse_descriptor;
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcResponseHeader_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcResponse_fieldAccessorTable;
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcResponseHeader_fieldAccessorTable;
}
public enum Status
@@ -2159,7 +2113,7 @@ public final class RPCProtos {
}
public static final com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
- return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse.getDescriptor().getEnumTypes().get(0);
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader.getDescriptor().getEnumTypes().get(0);
}
private static final Status[] VALUES = {
@@ -2183,7 +2137,7 @@ public final class RPCProtos {
this.value = value;
}
- // @@protoc_insertion_point(enum_scope:RpcResponse.Status)
+ // @@protoc_insertion_point(enum_scope:RpcResponseHeader.Status)
}
private int bitField0_;
@@ -2197,44 +2151,19 @@ public final class RPCProtos {
return callId_;
}
- // required .RpcResponse.Status status = 2;
+ // required .RpcResponseHeader.Status status = 2;
public static final int STATUS_FIELD_NUMBER = 2;
- private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse.Status status_;
+ private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader.Status status_;
public boolean hasStatus() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
- public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse.Status getStatus() {
+ public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader.Status getStatus() {
return status_;
}
- // optional bytes response = 3;
- public static final int RESPONSE_FIELD_NUMBER = 3;
- private com.google.protobuf.ByteString response_;
- public boolean hasResponse() {
- return ((bitField0_ & 0x00000004) == 0x00000004);
- }
- public com.google.protobuf.ByteString getResponse() {
- return response_;
- }
-
- // optional .RpcException exception = 4;
- public static final int EXCEPTION_FIELD_NUMBER = 4;
- private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException exception_;
- public boolean hasException() {
- return ((bitField0_ & 0x00000008) == 0x00000008);
- }
- public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException getException() {
- return exception_;
- }
- public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcExceptionOrBuilder getExceptionOrBuilder() {
- return exception_;
- }
-
private void initFields() {
callId_ = 0;
- status_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse.Status.SUCCESS;
- response_ = com.google.protobuf.ByteString.EMPTY;
- exception_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException.getDefaultInstance();
+ status_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader.Status.SUCCESS;
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
@@ -2249,12 +2178,6 @@ public final class RPCProtos {
memoizedIsInitialized = 0;
return false;
}
- if (hasException()) {
- if (!getException().isInitialized()) {
- memoizedIsInitialized = 0;
- return false;
- }
- }
memoizedIsInitialized = 1;
return true;
}
@@ -2268,12 +2191,6 @@ public final class RPCProtos {
if (((bitField0_ & 0x00000002) == 0x00000002)) {
output.writeEnum(2, status_.getNumber());
}
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
- output.writeBytes(3, response_);
- }
- if (((bitField0_ & 0x00000008) == 0x00000008)) {
- output.writeMessage(4, exception_);
- }
getUnknownFields().writeTo(output);
}
@@ -2291,14 +2208,6 @@ public final class RPCProtos {
size += com.google.protobuf.CodedOutputStream
.computeEnumSize(2, status_.getNumber());
}
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
- size += com.google.protobuf.CodedOutputStream
- .computeBytesSize(3, response_);
- }
- if (((bitField0_ & 0x00000008) == 0x00000008)) {
- size += com.google.protobuf.CodedOutputStream
- .computeMessageSize(4, exception_);
- }
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
@@ -2316,10 +2225,10 @@ public final class RPCProtos {
if (obj == this) {
return true;
}
- if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse)) {
+ if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader)) {
return super.equals(obj);
}
- org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse other = (org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse) obj;
+ org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader other = (org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader) obj;
boolean result = true;
result = result && (hasCallId() == other.hasCallId());
@@ -2332,16 +2241,6 @@ public final class RPCProtos {
result = result &&
(getStatus() == other.getStatus());
}
- result = result && (hasResponse() == other.hasResponse());
- if (hasResponse()) {
- result = result && getResponse()
- .equals(other.getResponse());
- }
- result = result && (hasException() == other.hasException());
- if (hasException()) {
- result = result && getException()
- .equals(other.getException());
- }
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
@@ -2359,53 +2258,45 @@ public final class RPCProtos {
hash = (37 * hash) + STATUS_FIELD_NUMBER;
hash = (53 * hash) + hashEnum(getStatus());
}
- if (hasResponse()) {
- hash = (37 * hash) + RESPONSE_FIELD_NUMBER;
- hash = (53 * hash) + getResponse().hashCode();
- }
- if (hasException()) {
- hash = (37 * hash) + EXCEPTION_FIELD_NUMBER;
- hash = (53 * hash) + getException().hashCode();
- }
hash = (29 * hash) + getUnknownFields().hashCode();
return hash;
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse parseFrom(
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data).buildParsed();
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse parseFrom(
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data, extensionRegistry)
.buildParsed();
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse parseFrom(byte[] data)
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data).buildParsed();
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse parseFrom(
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data, extensionRegistry)
.buildParsed();
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse parseFrom(java.io.InputStream input)
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader parseFrom(java.io.InputStream input)
throws java.io.IOException {
return newBuilder().mergeFrom(input).buildParsed();
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse parseFrom(
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return newBuilder().mergeFrom(input, extensionRegistry)
.buildParsed();
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse parseDelimitedFrom(java.io.InputStream input)
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
Builder builder = newBuilder();
if (builder.mergeDelimitedFrom(input)) {
@@ -2414,7 +2305,7 @@ public final class RPCProtos {
return null;
}
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse parseDelimitedFrom(
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
@@ -2425,12 +2316,12 @@ public final class RPCProtos {
return null;
}
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse parseFrom(
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return newBuilder().mergeFrom(input).buildParsed();
}
- public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse parseFrom(
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
@@ -2440,7 +2331,7 @@ public final class RPCProtos {
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
- public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse prototype) {
+ public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@@ -2453,18 +2344,18 @@ public final class RPCProtos {
}
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseOrBuilder {
+ implements org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeaderOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
- return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcResponse_descriptor;
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcResponseHeader_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcResponse_fieldAccessorTable;
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcResponseHeader_fieldAccessorTable;
}
- // Construct using org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse.newBuilder()
+ // Construct using org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
@@ -2475,7 +2366,6 @@ public final class RPCProtos {
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
- getExceptionFieldBuilder();
}
}
private static Builder create() {
@@ -2486,16 +2376,8 @@ public final class RPCProtos {
super.clear();
callId_ = 0;
bitField0_ = (bitField0_ & ~0x00000001);
- status_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse.Status.SUCCESS;
+ status_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader.Status.SUCCESS;
bitField0_ = (bitField0_ & ~0x00000002);
- response_ = com.google.protobuf.ByteString.EMPTY;
- bitField0_ = (bitField0_ & ~0x00000004);
- if (exceptionBuilder_ == null) {
- exception_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException.getDefaultInstance();
- } else {
- exceptionBuilder_.clear();
- }
- bitField0_ = (bitField0_ & ~0x00000008);
return this;
}
@@ -2505,24 +2387,24 @@ public final class RPCProtos {
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
- return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse.getDescriptor();
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader.getDescriptor();
}
- public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse getDefaultInstanceForType() {
- return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse.getDefaultInstance();
+ public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader getDefaultInstanceForType() {
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader.getDefaultInstance();
}
- public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse build() {
- org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse result = buildPartial();
+ public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader build() {
+ org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
- private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse buildParsed()
+ private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader buildParsed()
throws com.google.protobuf.InvalidProtocolBufferException {
- org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse result = buildPartial();
+ org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(
result).asInvalidProtocolBufferException();
@@ -2530,8 +2412,8 @@ public final class RPCProtos {
return result;
}
- public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse buildPartial() {
- org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse result = new org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse(this);
+ public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader buildPartial() {
+ org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader result = new org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader(this);
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
@@ -2542,46 +2424,28 @@ public final class RPCProtos {
to_bitField0_ |= 0x00000002;
}
result.status_ = status_;
- if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
- to_bitField0_ |= 0x00000004;
- }
- result.response_ = response_;
- if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
- to_bitField0_ |= 0x00000008;
- }
- if (exceptionBuilder_ == null) {
- result.exception_ = exception_;
- } else {
- result.exception_ = exceptionBuilder_.build();
- }
result.bitField0_ = to_bitField0_;
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
- if (other instanceof org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse) {
- return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse)other);
+ if (other instanceof org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader) {
+ return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader)other);
} else {
super.mergeFrom(other);
return this;
}
}
- public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse other) {
- if (other == org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse.getDefaultInstance()) return this;
+ public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader other) {
+ if (other == org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader.getDefaultInstance()) return this;
if (other.hasCallId()) {
setCallId(other.getCallId());
}
if (other.hasStatus()) {
setStatus(other.getStatus());
}
- if (other.hasResponse()) {
- setResponse(other.getResponse());
- }
- if (other.hasException()) {
- mergeException(other.getException());
- }
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
@@ -2595,12 +2459,6 @@ public final class RPCProtos {
return false;
}
- if (hasException()) {
- if (!getException().isInitialized()) {
-
- return false;
- }
- }
return true;
}
@@ -2634,7 +2492,7 @@ public final class RPCProtos {
}
case 16: {
int rawValue = input.readEnum();
- org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse.Status value = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse.Status.valueOf(rawValue);
+ org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader.Status value = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader.Status.valueOf(rawValue);
if (value == null) {
unknownFields.mergeVarintField(2, rawValue);
} else {
@@ -2643,20 +2501,6 @@ public final class RPCProtos {
}
break;
}
- case 26: {
- bitField0_ |= 0x00000004;
- response_ = input.readBytes();
- break;
- }
- case 34: {
- org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException.newBuilder();
- if (hasException()) {
- subBuilder.mergeFrom(getException());
- }
- input.readMessage(subBuilder, extensionRegistry);
- setException(subBuilder.buildPartial());
- break;
- }
}
}
}
@@ -2684,15 +2528,15 @@ public final class RPCProtos {
return this;
}
- // required .RpcResponse.Status status = 2;
- private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse.Status status_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse.Status.SUCCESS;
+ // required .RpcResponseHeader.Status status = 2;
+ private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader.Status status_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader.Status.SUCCESS;
public boolean hasStatus() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
- public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse.Status getStatus() {
+ public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader.Status getStatus() {
return status_;
}
- public Builder setStatus(org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse.Status value) {
+ public Builder setStatus(org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader.Status value) {
if (value == null) {
throw new NullPointerException();
}
@@ -2703,161 +2547,950 @@ public final class RPCProtos {
}
public Builder clearStatus() {
bitField0_ = (bitField0_ & ~0x00000002);
- status_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponse.Status.SUCCESS;
+ status_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseHeader.Status.SUCCESS;
onChanged();
return this;
}
- // optional bytes response = 3;
- private com.google.protobuf.ByteString response_ = com.google.protobuf.ByteString.EMPTY;
- public boolean hasResponse() {
- return ((bitField0_ & 0x00000004) == 0x00000004);
- }
- public com.google.protobuf.ByteString getResponse() {
- return response_;
- }
- public Builder setResponse(com.google.protobuf.ByteString value) {
- if (value == null) {
- throw new NullPointerException();
+ // @@protoc_insertion_point(builder_scope:RpcResponseHeader)
+ }
+
+ static {
+ defaultInstance = new RpcResponseHeader(true);
+ defaultInstance.initFields();
+ }
+
+ // @@protoc_insertion_point(class_scope:RpcResponseHeader)
}
- bitField0_ |= 0x00000004;
- response_ = value;
- onChanged();
- return this;
- }
- public Builder clearResponse() {
- bitField0_ = (bitField0_ & ~0x00000004);
- response_ = getDefaultInstance().getResponse();
- onChanged();
- return this;
- }
+
+ public interface RpcResponseBodyOrBuilder
+ extends com.google.protobuf.MessageOrBuilder {
+
+ // optional bytes response = 1;
+ boolean hasResponse();
+ com.google.protobuf.ByteString getResponse();
+ }
+ public static final class RpcResponseBody extends
+ com.google.protobuf.GeneratedMessage
+ implements RpcResponseBodyOrBuilder {
+ // Use RpcResponseBody.newBuilder() to construct.
+ private RpcResponseBody(Builder builder) {
+ super(builder);
+ }
+ private RpcResponseBody(boolean noInit) {}
+
+ private static final RpcResponseBody defaultInstance;
+ public static RpcResponseBody getDefaultInstance() {
+ return defaultInstance;
+ }
+
+ public RpcResponseBody getDefaultInstanceForType() {
+ return defaultInstance;
+ }
+
+ public static final com.google.protobuf.Descriptors.Descriptor
+ getDescriptor() {
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcResponseBody_descriptor;
+ }
+
+ protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ internalGetFieldAccessorTable() {
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcResponseBody_fieldAccessorTable;
+ }
+
+ private int bitField0_;
+ // optional bytes response = 1;
+ public static final int RESPONSE_FIELD_NUMBER = 1;
+ private com.google.protobuf.ByteString response_;
+ public boolean hasResponse() {
+ return ((bitField0_ & 0x00000001) == 0x00000001);
+ }
+ public com.google.protobuf.ByteString getResponse() {
+ return response_;
+ }
+
+ private void initFields() {
+ response_ = com.google.protobuf.ByteString.EMPTY;
+ }
+ private byte memoizedIsInitialized = -1;
+ public final boolean isInitialized() {
+ byte isInitialized = memoizedIsInitialized;
+ if (isInitialized != -1) return isInitialized == 1;
- // optional .RpcException exception = 4;
- private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException exception_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException.getDefaultInstance();
- private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException.Builder, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcExceptionOrBuilder> exceptionBuilder_;
- public boolean hasException() {
- return ((bitField0_ & 0x00000008) == 0x00000008);
- }
- public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException getException() {
- if (exceptionBuilder_ == null) {
- return exception_;
- } else {
- return exceptionBuilder_.getMessage();
- }
- }
- public Builder setException(org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException value) {
- if (exceptionBuilder_ == null) {
- if (value == null) {
- throw new NullPointerException();
- }
- exception_ = value;
- onChanged();
- } else {
- exceptionBuilder_.setMessage(value);
- }
- bitField0_ |= 0x00000008;
- return this;
- }
- public Builder setException(
- org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException.Builder builderForValue) {
- if (exceptionBuilder_ == null) {
- exception_ = builderForValue.build();
- onChanged();
- } else {
- exceptionBuilder_.setMessage(builderForValue.build());
- }
- bitField0_ |= 0x00000008;
- return this;
- }
- public Builder mergeException(org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException value) {
- if (exceptionBuilder_ == null) {
- if (((bitField0_ & 0x00000008) == 0x00000008) &&
- exception_ != org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException.getDefaultInstance()) {
- exception_ =
- org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException.newBuilder(exception_).mergeFrom(value).buildPartial();
- } else {
- exception_ = value;
- }
- onChanged();
- } else {
- exceptionBuilder_.mergeFrom(value);
- }
- bitField0_ |= 0x00000008;
- return this;
- }
- public Builder clearException() {
- if (exceptionBuilder_ == null) {
- exception_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException.getDefaultInstance();
- onChanged();
- } else {
- exceptionBuilder_.clear();
- }
- bitField0_ = (bitField0_ & ~0x00000008);
- return this;
+ memoizedIsInitialized = 1;
+ return true;
+ }
+
+ public void writeTo(com.google.protobuf.CodedOutputStream output)
+ throws java.io.IOException {
+ getSerializedSize();
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ output.writeBytes(1, response_);
}
- public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException.Builder getExceptionBuilder() {
- bitField0_ |= 0x00000008;
- onChanged();
- return getExceptionFieldBuilder().getBuilder();
+ getUnknownFields().writeTo(output);
+ }
+
+ private int memoizedSerializedSize = -1;
+ public int getSerializedSize() {
+ int size = memoizedSerializedSize;
+ if (size != -1) return size;
+
+ size = 0;
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ size += com.google.protobuf.CodedOutputStream
+ .computeBytesSize(1, response_);
}
- public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcExceptionOrBuilder getExceptionOrBuilder() {
- if (exceptionBuilder_ != null) {
- return exceptionBuilder_.getMessageOrBuilder();
- } else {
- return exception_;
- }
+ size += getUnknownFields().getSerializedSize();
+ memoizedSerializedSize = size;
+ return size;
+ }
+
+ private static final long serialVersionUID = 0L;
+ @java.lang.Override
+ protected java.lang.Object writeReplace()
+ throws java.io.ObjectStreamException {
+ return super.writeReplace();
+ }
+
+ @java.lang.Override
+ public boolean equals(final java.lang.Object obj) {
+ if (obj == this) {
+ return true;
}
- private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException.Builder, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcExceptionOrBuilder>
- getExceptionFieldBuilder() {
- if (exceptionBuilder_ == null) {
- exceptionBuilder_ = new com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcException.Builder, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcExceptionOrBuilder>(
- exception_,
- getParentForChildren(),
- isClean());
- exception_ = null;
- }
- return exceptionBuilder_;
+ if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody)) {
+ return super.equals(obj);
}
+ org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody other = (org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody) obj;
- // @@protoc_insertion_point(builder_scope:RpcResponse)
+ boolean result = true;
+ result = result && (hasResponse() == other.hasResponse());
+ if (hasResponse()) {
+ result = result && getResponse()
+ .equals(other.getResponse());
+ }
+ result = result &&
+ getUnknownFields().equals(other.getUnknownFields());
+ return result;
}
- static {
- defaultInstance = new RpcResponse(true);
- defaultInstance.initFields();
+ @java.lang.Override
+ public int hashCode() {
+ int hash = 41;
+ hash = (19 * hash) + getDescriptorForType().hashCode();
+ if (hasResponse()) {
+ hash = (37 * hash) + RESPONSE_FIELD_NUMBER;
+ hash = (53 * hash) + getResponse().hashCode();
+ }
+ hash = (29 * hash) + getUnknownFields().hashCode();
+ return hash;
}
- // @@protoc_insertion_point(class_scope:RpcResponse)
- }
-
- private static com.google.protobuf.Descriptors.Descriptor
- internal_static_UserInformation_descriptor;
- private static
- com.google.protobuf.GeneratedMessage.FieldAccessorTable
- internal_static_UserInformation_fieldAccessorTable;
- private static com.google.protobuf.Descriptors.Descriptor
- internal_static_ConnectionHeader_descriptor;
- private static
- com.google.protobuf.GeneratedMessage.FieldAccessorTable
- internal_static_ConnectionHeader_fieldAccessorTable;
- private static com.google.protobuf.Descriptors.Descriptor
- internal_static_RpcRequest_descriptor;
- private static
- com.google.protobuf.GeneratedMessage.FieldAccessorTable
- internal_static_RpcRequest_fieldAccessorTable;
- private static com.google.protobuf.Descriptors.Descriptor
- internal_static_RpcException_descriptor;
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody parseFrom(
+ com.google.protobuf.ByteString data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return newBuilder().mergeFrom(data).buildParsed();
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody parseFrom(
+ com.google.protobuf.ByteString data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return newBuilder().mergeFrom(data, extensionRegistry)
+ .buildParsed();
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody parseFrom(byte[] data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return newBuilder().mergeFrom(data).buildParsed();
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody parseFrom(
+ byte[] data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return newBuilder().mergeFrom(data, extensionRegistry)
+ .buildParsed();
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody parseFrom(java.io.InputStream input)
+ throws java.io.IOException {
+ return newBuilder().mergeFrom(input).buildParsed();
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody parseFrom(
+ java.io.InputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return newBuilder().mergeFrom(input, extensionRegistry)
+ .buildParsed();
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody parseDelimitedFrom(java.io.InputStream input)
+ throws java.io.IOException {
+ Builder builder = newBuilder();
+ if (builder.mergeDelimitedFrom(input)) {
+ return builder.buildParsed();
+ } else {
+ return null;
+ }
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody parseDelimitedFrom(
+ java.io.InputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ Builder builder = newBuilder();
+ if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
+ return builder.buildParsed();
+ } else {
+ return null;
+ }
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody parseFrom(
+ com.google.protobuf.CodedInputStream input)
+ throws java.io.IOException {
+ return newBuilder().mergeFrom(input).buildParsed();
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody parseFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return newBuilder().mergeFrom(input, extensionRegistry)
+ .buildParsed();
+ }
+
+ public static Builder newBuilder() { return Builder.create(); }
+ public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody prototype) {
+ return newBuilder().mergeFrom(prototype);
+ }
+ public Builder toBuilder() { return newBuilder(this); }
+
+ @java.lang.Override
+ protected Builder newBuilderForType(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ Builder builder = new Builder(parent);
+ return builder;
+ }
+ public static final class Builder extends
+ com.google.protobuf.GeneratedMessage.Builder<Builder>
+ implements org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBodyOrBuilder {
+ public static final com.google.protobuf.Descriptors.Descriptor
+ getDescriptor() {
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcResponseBody_descriptor;
+ }
+
+ protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ internalGetFieldAccessorTable() {
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.internal_static_RpcResponseBody_fieldAccessorTable;
+ }
+
+ // Construct using org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody.newBuilder()
+ private Builder() {
+ maybeForceBuilderInitialization();
+ }
+
+ private Builder(BuilderParent parent) {
+ super(parent);
+ maybeForceBuilderInitialization();
+ }
+ private void maybeForceBuilderInitialization() {
+ if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ }
+ }
+ private static Builder create() {
+ return new Builder();
+ }
+
+ public Builder clear() {
+ super.clear();
+ response_ = com.google.protobuf.ByteString.EMPTY;
+ bitField0_ = (bitField0_ & ~0x00000001);
+ return this;
+ }
+
+ public Builder clone() {
+ return create().mergeFrom(buildPartial());
+ }
+
+ public com.google.protobuf.Descriptors.Descriptor
+ getDescriptorForType() {
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody.getDescriptor();
+ }
+
+ public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody getDefaultInstanceForType() {
+ return org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody.getDefaultInstance();
+ }
+
+ public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody build() {
+ org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody result = buildPartial();
+ if (!result.isInitialized()) {
+ throw newUninitializedMessageException(result);
+ }
+ return result;
+ }
+
+ private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody buildParsed()
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody result = buildPartial();
+ if (!result.isInitialized()) {
+ throw newUninitializedMessageException(
+ result).asInvalidProtocolBufferException();
+ }
+ return result;
+ }
+
+ public org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody buildPartial() {
+ org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody result = new org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody(this);
+ int from_bitField0_ = bitField0_;
+ int to_bitField0_ = 0;
+ if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
+ to_bitField0_ |= 0x00000001;
+ }
+ result.response_ = response_;
+ result.bitField0_ = to_bitField0_;
+ onBuilt();
+ return result;
+ }
+
+ public Builder mergeFrom(com.google.protobuf.Message other) {
+ if (other instanceof org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody) {
+ return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody)other);
+ } else {
+ super.mergeFrom(other);
+ return this;
+ }
+ }
+
+ public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody other) {
+ if (other == org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RpcResponseBody.getDefaultInstance()) return this;
+ if (other.hasResponse()) {
+ setResponse(other.getResponse());
+ }
+ this.mergeUnknownFields(other.getUnknownFields());
+ return this;
+ }
+
+ public final boolean isInitialized() {
+ return true;
+ }
+
+ public Builder mergeFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ com.google.protobuf.UnknownFieldSet.newBuilder(
+ this.getUnknownFields());
+ while (true) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
[... 709 lines stripped ...]