You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by ss...@apache.org on 2016/01/25 00:36:56 UTC
[4/8] hive git commit: HIVE-12448. Change to tracking of dag status
via dagIdentifier instead of dag name. (Siddharth Seth,
reviewed by Sergey Shelukhin) (cherry picked from commit
16d495809382cf1db54ab26ff3a7ba5d57caa9b2)
http://git-wip-us.apache.org/repos/asf/hive/blob/94e8761a/llap-server/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java
----------------------------------------------------------------------
diff --git a/llap-server/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java b/llap-server/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java
index d2180e5..4ab7b32 100644
--- a/llap-server/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java
+++ b/llap-server/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java
@@ -3245,6 +3245,16 @@ public final class LlapDaemonProtocolProtos {
com.google.protobuf.ByteString
getDagNameBytes();
+ // optional int32 dag_id = 11;
+ /**
+ * <code>optional int32 dag_id = 11;</code>
+ */
+ boolean hasDagId();
+ /**
+ * <code>optional int32 dag_id = 11;</code>
+ */
+ int getDagId();
+
// optional string vertex_name = 3;
/**
* <code>optional string vertex_name = 3;</code>
@@ -3441,13 +3451,13 @@ public final class LlapDaemonProtocolProtos {
break;
}
case 26: {
- bitField0_ |= 0x00000004;
+ bitField0_ |= 0x00000008;
vertexName_ = input.readBytes();
break;
}
case 34: {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder subBuilder = null;
- if (((bitField0_ & 0x00000008) == 0x00000008)) {
+ if (((bitField0_ & 0x00000010) == 0x00000010)) {
subBuilder = processorDescriptor_.toBuilder();
}
processorDescriptor_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.PARSER, extensionRegistry);
@@ -3455,48 +3465,53 @@ public final class LlapDaemonProtocolProtos {
subBuilder.mergeFrom(processorDescriptor_);
processorDescriptor_ = subBuilder.buildPartial();
}
- bitField0_ |= 0x00000008;
+ bitField0_ |= 0x00000010;
break;
}
case 42: {
- if (!((mutable_bitField0_ & 0x00000010) == 0x00000010)) {
+ if (!((mutable_bitField0_ & 0x00000020) == 0x00000020)) {
inputSpecs_ = new java.util.ArrayList<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto>();
- mutable_bitField0_ |= 0x00000010;
+ mutable_bitField0_ |= 0x00000020;
}
inputSpecs_.add(input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.PARSER, extensionRegistry));
break;
}
case 50: {
- if (!((mutable_bitField0_ & 0x00000020) == 0x00000020)) {
+ if (!((mutable_bitField0_ & 0x00000040) == 0x00000040)) {
outputSpecs_ = new java.util.ArrayList<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto>();
- mutable_bitField0_ |= 0x00000020;
+ mutable_bitField0_ |= 0x00000040;
}
outputSpecs_.add(input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.PARSER, extensionRegistry));
break;
}
case 58: {
- if (!((mutable_bitField0_ & 0x00000040) == 0x00000040)) {
+ if (!((mutable_bitField0_ & 0x00000080) == 0x00000080)) {
groupedInputSpecs_ = new java.util.ArrayList<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto>();
- mutable_bitField0_ |= 0x00000040;
+ mutable_bitField0_ |= 0x00000080;
}
groupedInputSpecs_.add(input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.PARSER, extensionRegistry));
break;
}
case 64: {
- bitField0_ |= 0x00000010;
+ bitField0_ |= 0x00000020;
vertexParallelism_ = input.readInt32();
break;
}
case 72: {
- bitField0_ |= 0x00000020;
+ bitField0_ |= 0x00000040;
fragmentNumber_ = input.readInt32();
break;
}
case 80: {
- bitField0_ |= 0x00000040;
+ bitField0_ |= 0x00000080;
attemptNumber_ = input.readInt32();
break;
}
+ case 88: {
+ bitField0_ |= 0x00000004;
+ dagId_ = input.readInt32();
+ break;
+ }
}
}
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
@@ -3505,13 +3520,13 @@ public final class LlapDaemonProtocolProtos {
throw new com.google.protobuf.InvalidProtocolBufferException(
e.getMessage()).setUnfinishedMessage(this);
} finally {
- if (((mutable_bitField0_ & 0x00000010) == 0x00000010)) {
+ if (((mutable_bitField0_ & 0x00000020) == 0x00000020)) {
inputSpecs_ = java.util.Collections.unmodifiableList(inputSpecs_);
}
- if (((mutable_bitField0_ & 0x00000020) == 0x00000020)) {
+ if (((mutable_bitField0_ & 0x00000040) == 0x00000040)) {
outputSpecs_ = java.util.Collections.unmodifiableList(outputSpecs_);
}
- if (((mutable_bitField0_ & 0x00000040) == 0x00000040)) {
+ if (((mutable_bitField0_ & 0x00000080) == 0x00000080)) {
groupedInputSpecs_ = java.util.Collections.unmodifiableList(groupedInputSpecs_);
}
this.unknownFields = unknownFields.build();
@@ -3632,6 +3647,22 @@ public final class LlapDaemonProtocolProtos {
}
}
+ // optional int32 dag_id = 11;
+ public static final int DAG_ID_FIELD_NUMBER = 11;
+ private int dagId_;
+ /**
+ * <code>optional int32 dag_id = 11;</code>
+ */
+ public boolean hasDagId() {
+ return ((bitField0_ & 0x00000004) == 0x00000004);
+ }
+ /**
+ * <code>optional int32 dag_id = 11;</code>
+ */
+ public int getDagId() {
+ return dagId_;
+ }
+
// optional string vertex_name = 3;
public static final int VERTEX_NAME_FIELD_NUMBER = 3;
private java.lang.Object vertexName_;
@@ -3639,7 +3670,7 @@ public final class LlapDaemonProtocolProtos {
* <code>optional string vertex_name = 3;</code>
*/
public boolean hasVertexName() {
- return ((bitField0_ & 0x00000004) == 0x00000004);
+ return ((bitField0_ & 0x00000008) == 0x00000008);
}
/**
* <code>optional string vertex_name = 3;</code>
@@ -3682,7 +3713,7 @@ public final class LlapDaemonProtocolProtos {
* <code>optional .EntityDescriptorProto processor_descriptor = 4;</code>
*/
public boolean hasProcessorDescriptor() {
- return ((bitField0_ & 0x00000008) == 0x00000008);
+ return ((bitField0_ & 0x00000010) == 0x00000010);
}
/**
* <code>optional .EntityDescriptorProto processor_descriptor = 4;</code>
@@ -3812,7 +3843,7 @@ public final class LlapDaemonProtocolProtos {
* <code>optional int32 vertex_parallelism = 8;</code>
*/
public boolean hasVertexParallelism() {
- return ((bitField0_ & 0x00000010) == 0x00000010);
+ return ((bitField0_ & 0x00000020) == 0x00000020);
}
/**
* <code>optional int32 vertex_parallelism = 8;</code>
@@ -3828,7 +3859,7 @@ public final class LlapDaemonProtocolProtos {
* <code>optional int32 fragment_number = 9;</code>
*/
public boolean hasFragmentNumber() {
- return ((bitField0_ & 0x00000020) == 0x00000020);
+ return ((bitField0_ & 0x00000040) == 0x00000040);
}
/**
* <code>optional int32 fragment_number = 9;</code>
@@ -3844,7 +3875,7 @@ public final class LlapDaemonProtocolProtos {
* <code>optional int32 attempt_number = 10;</code>
*/
public boolean hasAttemptNumber() {
- return ((bitField0_ & 0x00000040) == 0x00000040);
+ return ((bitField0_ & 0x00000080) == 0x00000080);
}
/**
* <code>optional int32 attempt_number = 10;</code>
@@ -3856,6 +3887,7 @@ public final class LlapDaemonProtocolProtos {
private void initFields() {
fragmentIdentifierString_ = "";
dagName_ = "";
+ dagId_ = 0;
vertexName_ = "";
processorDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
inputSpecs_ = java.util.Collections.emptyList();
@@ -3883,10 +3915,10 @@ public final class LlapDaemonProtocolProtos {
if (((bitField0_ & 0x00000002) == 0x00000002)) {
output.writeBytes(2, getDagNameBytes());
}
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
+ if (((bitField0_ & 0x00000008) == 0x00000008)) {
output.writeBytes(3, getVertexNameBytes());
}
- if (((bitField0_ & 0x00000008) == 0x00000008)) {
+ if (((bitField0_ & 0x00000010) == 0x00000010)) {
output.writeMessage(4, processorDescriptor_);
}
for (int i = 0; i < inputSpecs_.size(); i++) {
@@ -3898,15 +3930,18 @@ public final class LlapDaemonProtocolProtos {
for (int i = 0; i < groupedInputSpecs_.size(); i++) {
output.writeMessage(7, groupedInputSpecs_.get(i));
}
- if (((bitField0_ & 0x00000010) == 0x00000010)) {
+ if (((bitField0_ & 0x00000020) == 0x00000020)) {
output.writeInt32(8, vertexParallelism_);
}
- if (((bitField0_ & 0x00000020) == 0x00000020)) {
+ if (((bitField0_ & 0x00000040) == 0x00000040)) {
output.writeInt32(9, fragmentNumber_);
}
- if (((bitField0_ & 0x00000040) == 0x00000040)) {
+ if (((bitField0_ & 0x00000080) == 0x00000080)) {
output.writeInt32(10, attemptNumber_);
}
+ if (((bitField0_ & 0x00000004) == 0x00000004)) {
+ output.writeInt32(11, dagId_);
+ }
getUnknownFields().writeTo(output);
}
@@ -3924,11 +3959,11 @@ public final class LlapDaemonProtocolProtos {
size += com.google.protobuf.CodedOutputStream
.computeBytesSize(2, getDagNameBytes());
}
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
+ if (((bitField0_ & 0x00000008) == 0x00000008)) {
size += com.google.protobuf.CodedOutputStream
.computeBytesSize(3, getVertexNameBytes());
}
- if (((bitField0_ & 0x00000008) == 0x00000008)) {
+ if (((bitField0_ & 0x00000010) == 0x00000010)) {
size += com.google.protobuf.CodedOutputStream
.computeMessageSize(4, processorDescriptor_);
}
@@ -3944,18 +3979,22 @@ public final class LlapDaemonProtocolProtos {
size += com.google.protobuf.CodedOutputStream
.computeMessageSize(7, groupedInputSpecs_.get(i));
}
- if (((bitField0_ & 0x00000010) == 0x00000010)) {
+ if (((bitField0_ & 0x00000020) == 0x00000020)) {
size += com.google.protobuf.CodedOutputStream
.computeInt32Size(8, vertexParallelism_);
}
- if (((bitField0_ & 0x00000020) == 0x00000020)) {
+ if (((bitField0_ & 0x00000040) == 0x00000040)) {
size += com.google.protobuf.CodedOutputStream
.computeInt32Size(9, fragmentNumber_);
}
- if (((bitField0_ & 0x00000040) == 0x00000040)) {
+ if (((bitField0_ & 0x00000080) == 0x00000080)) {
size += com.google.protobuf.CodedOutputStream
.computeInt32Size(10, attemptNumber_);
}
+ if (((bitField0_ & 0x00000004) == 0x00000004)) {
+ size += com.google.protobuf.CodedOutputStream
+ .computeInt32Size(11, dagId_);
+ }
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
@@ -3989,6 +4028,11 @@ public final class LlapDaemonProtocolProtos {
result = result && getDagName()
.equals(other.getDagName());
}
+ result = result && (hasDagId() == other.hasDagId());
+ if (hasDagId()) {
+ result = result && (getDagId()
+ == other.getDagId());
+ }
result = result && (hasVertexName() == other.hasVertexName());
if (hasVertexName()) {
result = result && getVertexName()
@@ -4041,6 +4085,10 @@ public final class LlapDaemonProtocolProtos {
hash = (37 * hash) + DAG_NAME_FIELD_NUMBER;
hash = (53 * hash) + getDagName().hashCode();
}
+ if (hasDagId()) {
+ hash = (37 * hash) + DAG_ID_FIELD_NUMBER;
+ hash = (53 * hash) + getDagId();
+ }
if (hasVertexName()) {
hash = (37 * hash) + VERTEX_NAME_FIELD_NUMBER;
hash = (53 * hash) + getVertexName().hashCode();
@@ -4190,38 +4238,40 @@ public final class LlapDaemonProtocolProtos {
bitField0_ = (bitField0_ & ~0x00000001);
dagName_ = "";
bitField0_ = (bitField0_ & ~0x00000002);
- vertexName_ = "";
+ dagId_ = 0;
bitField0_ = (bitField0_ & ~0x00000004);
+ vertexName_ = "";
+ bitField0_ = (bitField0_ & ~0x00000008);
if (processorDescriptorBuilder_ == null) {
processorDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
} else {
processorDescriptorBuilder_.clear();
}
- bitField0_ = (bitField0_ & ~0x00000008);
+ bitField0_ = (bitField0_ & ~0x00000010);
if (inputSpecsBuilder_ == null) {
inputSpecs_ = java.util.Collections.emptyList();
- bitField0_ = (bitField0_ & ~0x00000010);
+ bitField0_ = (bitField0_ & ~0x00000020);
} else {
inputSpecsBuilder_.clear();
}
if (outputSpecsBuilder_ == null) {
outputSpecs_ = java.util.Collections.emptyList();
- bitField0_ = (bitField0_ & ~0x00000020);
+ bitField0_ = (bitField0_ & ~0x00000040);
} else {
outputSpecsBuilder_.clear();
}
if (groupedInputSpecsBuilder_ == null) {
groupedInputSpecs_ = java.util.Collections.emptyList();
- bitField0_ = (bitField0_ & ~0x00000040);
+ bitField0_ = (bitField0_ & ~0x00000080);
} else {
groupedInputSpecsBuilder_.clear();
}
vertexParallelism_ = 0;
- bitField0_ = (bitField0_ & ~0x00000080);
- fragmentNumber_ = 0;
bitField0_ = (bitField0_ & ~0x00000100);
- attemptNumber_ = 0;
+ fragmentNumber_ = 0;
bitField0_ = (bitField0_ & ~0x00000200);
+ attemptNumber_ = 0;
+ bitField0_ = (bitField0_ & ~0x00000400);
return this;
}
@@ -4261,53 +4311,57 @@ public final class LlapDaemonProtocolProtos {
if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
to_bitField0_ |= 0x00000004;
}
- result.vertexName_ = vertexName_;
+ result.dagId_ = dagId_;
if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
to_bitField0_ |= 0x00000008;
}
+ result.vertexName_ = vertexName_;
+ if (((from_bitField0_ & 0x00000010) == 0x00000010)) {
+ to_bitField0_ |= 0x00000010;
+ }
if (processorDescriptorBuilder_ == null) {
result.processorDescriptor_ = processorDescriptor_;
} else {
result.processorDescriptor_ = processorDescriptorBuilder_.build();
}
if (inputSpecsBuilder_ == null) {
- if (((bitField0_ & 0x00000010) == 0x00000010)) {
+ if (((bitField0_ & 0x00000020) == 0x00000020)) {
inputSpecs_ = java.util.Collections.unmodifiableList(inputSpecs_);
- bitField0_ = (bitField0_ & ~0x00000010);
+ bitField0_ = (bitField0_ & ~0x00000020);
}
result.inputSpecs_ = inputSpecs_;
} else {
result.inputSpecs_ = inputSpecsBuilder_.build();
}
if (outputSpecsBuilder_ == null) {
- if (((bitField0_ & 0x00000020) == 0x00000020)) {
+ if (((bitField0_ & 0x00000040) == 0x00000040)) {
outputSpecs_ = java.util.Collections.unmodifiableList(outputSpecs_);
- bitField0_ = (bitField0_ & ~0x00000020);
+ bitField0_ = (bitField0_ & ~0x00000040);
}
result.outputSpecs_ = outputSpecs_;
} else {
result.outputSpecs_ = outputSpecsBuilder_.build();
}
if (groupedInputSpecsBuilder_ == null) {
- if (((bitField0_ & 0x00000040) == 0x00000040)) {
+ if (((bitField0_ & 0x00000080) == 0x00000080)) {
groupedInputSpecs_ = java.util.Collections.unmodifiableList(groupedInputSpecs_);
- bitField0_ = (bitField0_ & ~0x00000040);
+ bitField0_ = (bitField0_ & ~0x00000080);
}
result.groupedInputSpecs_ = groupedInputSpecs_;
} else {
result.groupedInputSpecs_ = groupedInputSpecsBuilder_.build();
}
- if (((from_bitField0_ & 0x00000080) == 0x00000080)) {
- to_bitField0_ |= 0x00000010;
- }
- result.vertexParallelism_ = vertexParallelism_;
if (((from_bitField0_ & 0x00000100) == 0x00000100)) {
to_bitField0_ |= 0x00000020;
}
- result.fragmentNumber_ = fragmentNumber_;
+ result.vertexParallelism_ = vertexParallelism_;
if (((from_bitField0_ & 0x00000200) == 0x00000200)) {
to_bitField0_ |= 0x00000040;
}
+ result.fragmentNumber_ = fragmentNumber_;
+ if (((from_bitField0_ & 0x00000400) == 0x00000400)) {
+ to_bitField0_ |= 0x00000080;
+ }
result.attemptNumber_ = attemptNumber_;
result.bitField0_ = to_bitField0_;
onBuilt();
@@ -4335,8 +4389,11 @@ public final class LlapDaemonProtocolProtos {
dagName_ = other.dagName_;
onChanged();
}
+ if (other.hasDagId()) {
+ setDagId(other.getDagId());
+ }
if (other.hasVertexName()) {
- bitField0_ |= 0x00000004;
+ bitField0_ |= 0x00000008;
vertexName_ = other.vertexName_;
onChanged();
}
@@ -4347,7 +4404,7 @@ public final class LlapDaemonProtocolProtos {
if (!other.inputSpecs_.isEmpty()) {
if (inputSpecs_.isEmpty()) {
inputSpecs_ = other.inputSpecs_;
- bitField0_ = (bitField0_ & ~0x00000010);
+ bitField0_ = (bitField0_ & ~0x00000020);
} else {
ensureInputSpecsIsMutable();
inputSpecs_.addAll(other.inputSpecs_);
@@ -4360,7 +4417,7 @@ public final class LlapDaemonProtocolProtos {
inputSpecsBuilder_.dispose();
inputSpecsBuilder_ = null;
inputSpecs_ = other.inputSpecs_;
- bitField0_ = (bitField0_ & ~0x00000010);
+ bitField0_ = (bitField0_ & ~0x00000020);
inputSpecsBuilder_ =
com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ?
getInputSpecsFieldBuilder() : null;
@@ -4373,7 +4430,7 @@ public final class LlapDaemonProtocolProtos {
if (!other.outputSpecs_.isEmpty()) {
if (outputSpecs_.isEmpty()) {
outputSpecs_ = other.outputSpecs_;
- bitField0_ = (bitField0_ & ~0x00000020);
+ bitField0_ = (bitField0_ & ~0x00000040);
} else {
ensureOutputSpecsIsMutable();
outputSpecs_.addAll(other.outputSpecs_);
@@ -4386,7 +4443,7 @@ public final class LlapDaemonProtocolProtos {
outputSpecsBuilder_.dispose();
outputSpecsBuilder_ = null;
outputSpecs_ = other.outputSpecs_;
- bitField0_ = (bitField0_ & ~0x00000020);
+ bitField0_ = (bitField0_ & ~0x00000040);
outputSpecsBuilder_ =
com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ?
getOutputSpecsFieldBuilder() : null;
@@ -4399,7 +4456,7 @@ public final class LlapDaemonProtocolProtos {
if (!other.groupedInputSpecs_.isEmpty()) {
if (groupedInputSpecs_.isEmpty()) {
groupedInputSpecs_ = other.groupedInputSpecs_;
- bitField0_ = (bitField0_ & ~0x00000040);
+ bitField0_ = (bitField0_ & ~0x00000080);
} else {
ensureGroupedInputSpecsIsMutable();
groupedInputSpecs_.addAll(other.groupedInputSpecs_);
@@ -4412,7 +4469,7 @@ public final class LlapDaemonProtocolProtos {
groupedInputSpecsBuilder_.dispose();
groupedInputSpecsBuilder_ = null;
groupedInputSpecs_ = other.groupedInputSpecs_;
- bitField0_ = (bitField0_ & ~0x00000040);
+ bitField0_ = (bitField0_ & ~0x00000080);
groupedInputSpecsBuilder_ =
com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ?
getGroupedInputSpecsFieldBuilder() : null;
@@ -4605,13 +4662,46 @@ public final class LlapDaemonProtocolProtos {
return this;
}
+ // optional int32 dag_id = 11;
+ private int dagId_ ;
+ /**
+ * <code>optional int32 dag_id = 11;</code>
+ */
+ public boolean hasDagId() {
+ return ((bitField0_ & 0x00000004) == 0x00000004);
+ }
+ /**
+ * <code>optional int32 dag_id = 11;</code>
+ */
+ public int getDagId() {
+ return dagId_;
+ }
+ /**
+ * <code>optional int32 dag_id = 11;</code>
+ */
+ public Builder setDagId(int value) {
+ bitField0_ |= 0x00000004;
+ dagId_ = value;
+ onChanged();
+ return this;
+ }
+ /**
+ * <code>optional int32 dag_id = 11;</code>
+ */
+ public Builder clearDagId() {
+ bitField0_ = (bitField0_ & ~0x00000004);
+ dagId_ = 0;
+ onChanged();
+ return this;
+ }
+
// optional string vertex_name = 3;
private java.lang.Object vertexName_ = "";
/**
* <code>optional string vertex_name = 3;</code>
*/
public boolean hasVertexName() {
- return ((bitField0_ & 0x00000004) == 0x00000004);
+ return ((bitField0_ & 0x00000008) == 0x00000008);
}
/**
* <code>optional string vertex_name = 3;</code>
@@ -4651,7 +4741,7 @@ public final class LlapDaemonProtocolProtos {
if (value == null) {
throw new NullPointerException();
}
- bitField0_ |= 0x00000004;
+ bitField0_ |= 0x00000008;
vertexName_ = value;
onChanged();
return this;
@@ -4660,7 +4750,7 @@ public final class LlapDaemonProtocolProtos {
* <code>optional string vertex_name = 3;</code>
*/
public Builder clearVertexName() {
- bitField0_ = (bitField0_ & ~0x00000004);
+ bitField0_ = (bitField0_ & ~0x00000008);
vertexName_ = getDefaultInstance().getVertexName();
onChanged();
return this;
@@ -4673,7 +4763,7 @@ public final class LlapDaemonProtocolProtos {
if (value == null) {
throw new NullPointerException();
}
- bitField0_ |= 0x00000004;
+ bitField0_ |= 0x00000008;
vertexName_ = value;
onChanged();
return this;
@@ -4687,7 +4777,7 @@ public final class LlapDaemonProtocolProtos {
* <code>optional .EntityDescriptorProto processor_descriptor = 4;</code>
*/
public boolean hasProcessorDescriptor() {
- return ((bitField0_ & 0x00000008) == 0x00000008);
+ return ((bitField0_ & 0x00000010) == 0x00000010);
}
/**
* <code>optional .EntityDescriptorProto processor_descriptor = 4;</code>
@@ -4712,7 +4802,7 @@ public final class LlapDaemonProtocolProtos {
} else {
processorDescriptorBuilder_.setMessage(value);
}
- bitField0_ |= 0x00000008;
+ bitField0_ |= 0x00000010;
return this;
}
/**
@@ -4726,7 +4816,7 @@ public final class LlapDaemonProtocolProtos {
} else {
processorDescriptorBuilder_.setMessage(builderForValue.build());
}
- bitField0_ |= 0x00000008;
+ bitField0_ |= 0x00000010;
return this;
}
/**
@@ -4734,7 +4824,7 @@ public final class LlapDaemonProtocolProtos {
*/
public Builder mergeProcessorDescriptor(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto value) {
if (processorDescriptorBuilder_ == null) {
- if (((bitField0_ & 0x00000008) == 0x00000008) &&
+ if (((bitField0_ & 0x00000010) == 0x00000010) &&
processorDescriptor_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) {
processorDescriptor_ =
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.newBuilder(processorDescriptor_).mergeFrom(value).buildPartial();
@@ -4745,7 +4835,7 @@ public final class LlapDaemonProtocolProtos {
} else {
processorDescriptorBuilder_.mergeFrom(value);
}
- bitField0_ |= 0x00000008;
+ bitField0_ |= 0x00000010;
return this;
}
/**
@@ -4758,14 +4848,14 @@ public final class LlapDaemonProtocolProtos {
} else {
processorDescriptorBuilder_.clear();
}
- bitField0_ = (bitField0_ & ~0x00000008);
+ bitField0_ = (bitField0_ & ~0x00000010);
return this;
}
/**
* <code>optional .EntityDescriptorProto processor_descriptor = 4;</code>
*/
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder getProcessorDescriptorBuilder() {
- bitField0_ |= 0x00000008;
+ bitField0_ |= 0x00000010;
onChanged();
return getProcessorDescriptorFieldBuilder().getBuilder();
}
@@ -4800,9 +4890,9 @@ public final class LlapDaemonProtocolProtos {
private java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> inputSpecs_ =
java.util.Collections.emptyList();
private void ensureInputSpecsIsMutable() {
- if (!((bitField0_ & 0x00000010) == 0x00000010)) {
+ if (!((bitField0_ & 0x00000020) == 0x00000020)) {
inputSpecs_ = new java.util.ArrayList<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto>(inputSpecs_);
- bitField0_ |= 0x00000010;
+ bitField0_ |= 0x00000020;
}
}
@@ -4951,7 +5041,7 @@ public final class LlapDaemonProtocolProtos {
public Builder clearInputSpecs() {
if (inputSpecsBuilder_ == null) {
inputSpecs_ = java.util.Collections.emptyList();
- bitField0_ = (bitField0_ & ~0x00000010);
+ bitField0_ = (bitField0_ & ~0x00000020);
onChanged();
} else {
inputSpecsBuilder_.clear();
@@ -5028,7 +5118,7 @@ public final class LlapDaemonProtocolProtos {
inputSpecsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder>(
inputSpecs_,
- ((bitField0_ & 0x00000010) == 0x00000010),
+ ((bitField0_ & 0x00000020) == 0x00000020),
getParentForChildren(),
isClean());
inputSpecs_ = null;
@@ -5040,9 +5130,9 @@ public final class LlapDaemonProtocolProtos {
private java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> outputSpecs_ =
java.util.Collections.emptyList();
private void ensureOutputSpecsIsMutable() {
- if (!((bitField0_ & 0x00000020) == 0x00000020)) {
+ if (!((bitField0_ & 0x00000040) == 0x00000040)) {
outputSpecs_ = new java.util.ArrayList<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto>(outputSpecs_);
- bitField0_ |= 0x00000020;
+ bitField0_ |= 0x00000040;
}
}
@@ -5191,7 +5281,7 @@ public final class LlapDaemonProtocolProtos {
public Builder clearOutputSpecs() {
if (outputSpecsBuilder_ == null) {
outputSpecs_ = java.util.Collections.emptyList();
- bitField0_ = (bitField0_ & ~0x00000020);
+ bitField0_ = (bitField0_ & ~0x00000040);
onChanged();
} else {
outputSpecsBuilder_.clear();
@@ -5268,7 +5358,7 @@ public final class LlapDaemonProtocolProtos {
outputSpecsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder>(
outputSpecs_,
- ((bitField0_ & 0x00000020) == 0x00000020),
+ ((bitField0_ & 0x00000040) == 0x00000040),
getParentForChildren(),
isClean());
outputSpecs_ = null;
@@ -5280,9 +5370,9 @@ public final class LlapDaemonProtocolProtos {
private java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto> groupedInputSpecs_ =
java.util.Collections.emptyList();
private void ensureGroupedInputSpecsIsMutable() {
- if (!((bitField0_ & 0x00000040) == 0x00000040)) {
+ if (!((bitField0_ & 0x00000080) == 0x00000080)) {
groupedInputSpecs_ = new java.util.ArrayList<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto>(groupedInputSpecs_);
- bitField0_ |= 0x00000040;
+ bitField0_ |= 0x00000080;
}
}
@@ -5431,7 +5521,7 @@ public final class LlapDaemonProtocolProtos {
public Builder clearGroupedInputSpecs() {
if (groupedInputSpecsBuilder_ == null) {
groupedInputSpecs_ = java.util.Collections.emptyList();
- bitField0_ = (bitField0_ & ~0x00000040);
+ bitField0_ = (bitField0_ & ~0x00000080);
onChanged();
} else {
groupedInputSpecsBuilder_.clear();
@@ -5508,7 +5598,7 @@ public final class LlapDaemonProtocolProtos {
groupedInputSpecsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder>(
groupedInputSpecs_,
- ((bitField0_ & 0x00000040) == 0x00000040),
+ ((bitField0_ & 0x00000080) == 0x00000080),
getParentForChildren(),
isClean());
groupedInputSpecs_ = null;
@@ -5522,7 +5612,7 @@ public final class LlapDaemonProtocolProtos {
* <code>optional int32 vertex_parallelism = 8;</code>
*/
public boolean hasVertexParallelism() {
- return ((bitField0_ & 0x00000080) == 0x00000080);
+ return ((bitField0_ & 0x00000100) == 0x00000100);
}
/**
* <code>optional int32 vertex_parallelism = 8;</code>
@@ -5534,7 +5624,7 @@ public final class LlapDaemonProtocolProtos {
* <code>optional int32 vertex_parallelism = 8;</code>
*/
public Builder setVertexParallelism(int value) {
- bitField0_ |= 0x00000080;
+ bitField0_ |= 0x00000100;
vertexParallelism_ = value;
onChanged();
return this;
@@ -5543,7 +5633,7 @@ public final class LlapDaemonProtocolProtos {
* <code>optional int32 vertex_parallelism = 8;</code>
*/
public Builder clearVertexParallelism() {
- bitField0_ = (bitField0_ & ~0x00000080);
+ bitField0_ = (bitField0_ & ~0x00000100);
vertexParallelism_ = 0;
onChanged();
return this;
@@ -5555,7 +5645,7 @@ public final class LlapDaemonProtocolProtos {
* <code>optional int32 fragment_number = 9;</code>
*/
public boolean hasFragmentNumber() {
- return ((bitField0_ & 0x00000100) == 0x00000100);
+ return ((bitField0_ & 0x00000200) == 0x00000200);
}
/**
* <code>optional int32 fragment_number = 9;</code>
@@ -5567,7 +5657,7 @@ public final class LlapDaemonProtocolProtos {
* <code>optional int32 fragment_number = 9;</code>
*/
public Builder setFragmentNumber(int value) {
- bitField0_ |= 0x00000100;
+ bitField0_ |= 0x00000200;
fragmentNumber_ = value;
onChanged();
return this;
@@ -5576,7 +5666,7 @@ public final class LlapDaemonProtocolProtos {
* <code>optional int32 fragment_number = 9;</code>
*/
public Builder clearFragmentNumber() {
- bitField0_ = (bitField0_ & ~0x00000100);
+ bitField0_ = (bitField0_ & ~0x00000200);
fragmentNumber_ = 0;
onChanged();
return this;
@@ -5588,7 +5678,7 @@ public final class LlapDaemonProtocolProtos {
* <code>optional int32 attempt_number = 10;</code>
*/
public boolean hasAttemptNumber() {
- return ((bitField0_ & 0x00000200) == 0x00000200);
+ return ((bitField0_ & 0x00000400) == 0x00000400);
}
/**
* <code>optional int32 attempt_number = 10;</code>
@@ -5600,7 +5690,7 @@ public final class LlapDaemonProtocolProtos {
* <code>optional int32 attempt_number = 10;</code>
*/
public Builder setAttemptNumber(int value) {
- bitField0_ |= 0x00000200;
+ bitField0_ |= 0x00000400;
attemptNumber_ = value;
onChanged();
return this;
@@ -5609,7 +5699,7 @@ public final class LlapDaemonProtocolProtos {
* <code>optional int32 attempt_number = 10;</code>
*/
public Builder clearAttemptNumber() {
- bitField0_ = (bitField0_ & ~0x00000200);
+ bitField0_ = (bitField0_ & ~0x00000400);
attemptNumber_ = 0;
onChanged();
return this;
@@ -6510,76 +6600,675 @@ public final class LlapDaemonProtocolProtos {
// @@protoc_insertion_point(class_scope:FragmentRuntimeInfo)
}
- public interface SubmitWorkRequestProtoOrBuilder
+ public interface QueryIdentifierProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
- // optional string container_id_string = 1;
- /**
- * <code>optional string container_id_string = 1;</code>
- */
- boolean hasContainerIdString();
- /**
- * <code>optional string container_id_string = 1;</code>
- */
- java.lang.String getContainerIdString();
- /**
- * <code>optional string container_id_string = 1;</code>
- */
- com.google.protobuf.ByteString
- getContainerIdStringBytes();
-
- // optional string am_host = 2;
+ // optional string app_identifier = 1;
/**
- * <code>optional string am_host = 2;</code>
+ * <code>optional string app_identifier = 1;</code>
*/
- boolean hasAmHost();
+ boolean hasAppIdentifier();
/**
- * <code>optional string am_host = 2;</code>
+ * <code>optional string app_identifier = 1;</code>
*/
- java.lang.String getAmHost();
+ java.lang.String getAppIdentifier();
/**
- * <code>optional string am_host = 2;</code>
+ * <code>optional string app_identifier = 1;</code>
*/
com.google.protobuf.ByteString
- getAmHostBytes();
+ getAppIdentifierBytes();
- // optional int32 am_port = 3;
+ // optional int32 dag_identifier = 2;
/**
- * <code>optional int32 am_port = 3;</code>
+ * <code>optional int32 dag_identifier = 2;</code>
*/
- boolean hasAmPort();
+ boolean hasDagIdentifier();
/**
- * <code>optional int32 am_port = 3;</code>
+ * <code>optional int32 dag_identifier = 2;</code>
*/
- int getAmPort();
+ int getDagIdentifier();
+ }
+ /**
+ * Protobuf type {@code QueryIdentifierProto}
+ */
+ public static final class QueryIdentifierProto extends
+ com.google.protobuf.GeneratedMessage
+ implements QueryIdentifierProtoOrBuilder {
+ // Use QueryIdentifierProto.newBuilder() to construct.
+ private QueryIdentifierProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ super(builder);
+ this.unknownFields = builder.getUnknownFields();
+ }
+ private QueryIdentifierProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
- // optional string token_identifier = 4;
- /**
- * <code>optional string token_identifier = 4;</code>
- */
- boolean hasTokenIdentifier();
- /**
- * <code>optional string token_identifier = 4;</code>
- */
- java.lang.String getTokenIdentifier();
- /**
- * <code>optional string token_identifier = 4;</code>
- */
- com.google.protobuf.ByteString
- getTokenIdentifierBytes();
+ private static final QueryIdentifierProto defaultInstance;
+ public static QueryIdentifierProto getDefaultInstance() {
+ return defaultInstance;
+ }
- // optional bytes credentials_binary = 5;
- /**
- * <code>optional bytes credentials_binary = 5;</code>
- */
- boolean hasCredentialsBinary();
- /**
- * <code>optional bytes credentials_binary = 5;</code>
- */
- com.google.protobuf.ByteString getCredentialsBinary();
+ public QueryIdentifierProto getDefaultInstanceForType() {
+ return defaultInstance;
+ }
- // optional string user = 6;
- /**
+ private final com.google.protobuf.UnknownFieldSet unknownFields;
+ @java.lang.Override
+ public final com.google.protobuf.UnknownFieldSet
+ getUnknownFields() {
+ return this.unknownFields;
+ }
+ private QueryIdentifierProto(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ initFields();
+ int mutable_bitField0_ = 0;
+ com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ com.google.protobuf.UnknownFieldSet.newBuilder();
+ try {
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ default: {
+ if (!parseUnknownField(input, unknownFields,
+ extensionRegistry, tag)) {
+ done = true;
+ }
+ break;
+ }
+ case 10: {
+ bitField0_ |= 0x00000001;
+ appIdentifier_ = input.readBytes();
+ break;
+ }
+ case 16: {
+ bitField0_ |= 0x00000002;
+ dagIdentifier_ = input.readInt32();
+ break;
+ }
+ }
+ }
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(this);
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(
+ e.getMessage()).setUnfinishedMessage(this);
+ } finally {
+ this.unknownFields = unknownFields.build();
+ makeExtensionsImmutable();
+ }
+ }
+ public static final com.google.protobuf.Descriptors.Descriptor
+ getDescriptor() {
+ return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_descriptor;
+ }
+
+ protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ internalGetFieldAccessorTable() {
+ return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder.class);
+ }
+
+ public static com.google.protobuf.Parser<QueryIdentifierProto> PARSER =
+ new com.google.protobuf.AbstractParser<QueryIdentifierProto>() {
+ public QueryIdentifierProto parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new QueryIdentifierProto(input, extensionRegistry);
+ }
+ };
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<QueryIdentifierProto> getParserForType() {
+ return PARSER;
+ }
+
+ private int bitField0_;
+ // optional string app_identifier = 1;
+ public static final int APP_IDENTIFIER_FIELD_NUMBER = 1;
+ private java.lang.Object appIdentifier_;
+ /**
+ * <code>optional string app_identifier = 1;</code>
+ */
+ public boolean hasAppIdentifier() {
+ return ((bitField0_ & 0x00000001) == 0x00000001);
+ }
+ /**
+ * <code>optional string app_identifier = 1;</code>
+ */
+ public java.lang.String getAppIdentifier() {
+ java.lang.Object ref = appIdentifier_;
+ if (ref instanceof java.lang.String) {
+ return (java.lang.String) ref;
+ } else {
+ com.google.protobuf.ByteString bs =
+ (com.google.protobuf.ByteString) ref;
+ java.lang.String s = bs.toStringUtf8();
+ if (bs.isValidUtf8()) {
+ appIdentifier_ = s;
+ }
+ return s;
+ }
+ }
+ /**
+ * <code>optional string app_identifier = 1;</code>
+ */
+ public com.google.protobuf.ByteString
+ getAppIdentifierBytes() {
+ java.lang.Object ref = appIdentifier_;
+ if (ref instanceof java.lang.String) {
+ com.google.protobuf.ByteString b =
+ com.google.protobuf.ByteString.copyFromUtf8(
+ (java.lang.String) ref);
+ appIdentifier_ = b;
+ return b;
+ } else {
+ return (com.google.protobuf.ByteString) ref;
+ }
+ }
+
+ // optional int32 dag_identifier = 2;
+ public static final int DAG_IDENTIFIER_FIELD_NUMBER = 2;
+ private int dagIdentifier_;
+ /**
+ * <code>optional int32 dag_identifier = 2;</code>
+ */
+ public boolean hasDagIdentifier() {
+ return ((bitField0_ & 0x00000002) == 0x00000002);
+ }
+ /**
+ * <code>optional int32 dag_identifier = 2;</code>
+ */
+ public int getDagIdentifier() {
+ return dagIdentifier_;
+ }
+
+ private void initFields() {
+ appIdentifier_ = "";
+ dagIdentifier_ = 0;
+ }
+ private byte memoizedIsInitialized = -1;
+ public final boolean isInitialized() {
+ byte isInitialized = memoizedIsInitialized;
+ if (isInitialized != -1) return isInitialized == 1;
+
+ memoizedIsInitialized = 1;
+ return true;
+ }
+
+ public void writeTo(com.google.protobuf.CodedOutputStream output)
+ throws java.io.IOException {
+ getSerializedSize();
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ output.writeBytes(1, getAppIdentifierBytes());
+ }
+ if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ output.writeInt32(2, dagIdentifier_);
+ }
+ getUnknownFields().writeTo(output);
+ }
+
+ private int memoizedSerializedSize = -1;
+ public int getSerializedSize() {
+ int size = memoizedSerializedSize;
+ if (size != -1) return size;
+
+ size = 0;
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ size += com.google.protobuf.CodedOutputStream
+ .computeBytesSize(1, getAppIdentifierBytes());
+ }
+ if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ size += com.google.protobuf.CodedOutputStream
+ .computeInt32Size(2, dagIdentifier_);
+ }
+ size += getUnknownFields().getSerializedSize();
+ memoizedSerializedSize = size;
+ return size;
+ }
+
+ private static final long serialVersionUID = 0L;
+ @java.lang.Override
+ protected java.lang.Object writeReplace()
+ throws java.io.ObjectStreamException {
+ return super.writeReplace();
+ }
+
+ @java.lang.Override
+ public boolean equals(final java.lang.Object obj) {
+ if (obj == this) {
+ return true;
+ }
+ if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto)) {
+ return super.equals(obj);
+ }
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto) obj;
+
+ boolean result = true;
+ result = result && (hasAppIdentifier() == other.hasAppIdentifier());
+ if (hasAppIdentifier()) {
+ result = result && getAppIdentifier()
+ .equals(other.getAppIdentifier());
+ }
+ result = result && (hasDagIdentifier() == other.hasDagIdentifier());
+ if (hasDagIdentifier()) {
+ result = result && (getDagIdentifier()
+ == other.getDagIdentifier());
+ }
+ result = result &&
+ getUnknownFields().equals(other.getUnknownFields());
+ return result;
+ }
+
+ private int memoizedHashCode = 0;
+ @java.lang.Override
+ public int hashCode() {
+ if (memoizedHashCode != 0) {
+ return memoizedHashCode;
+ }
+ int hash = 41;
+ hash = (19 * hash) + getDescriptorForType().hashCode();
+ if (hasAppIdentifier()) {
+ hash = (37 * hash) + APP_IDENTIFIER_FIELD_NUMBER;
+ hash = (53 * hash) + getAppIdentifier().hashCode();
+ }
+ if (hasDagIdentifier()) {
+ hash = (37 * hash) + DAG_IDENTIFIER_FIELD_NUMBER;
+ hash = (53 * hash) + getDagIdentifier();
+ }
+ hash = (29 * hash) + getUnknownFields().hashCode();
+ memoizedHashCode = hash;
+ return hash;
+ }
+
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(
+ com.google.protobuf.ByteString data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(
+ com.google.protobuf.ByteString data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(byte[] data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(
+ byte[] data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(java.io.InputStream input)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input);
+ }
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(
+ java.io.InputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input, extensionRegistry);
+ }
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseDelimitedFrom(java.io.InputStream input)
+ throws java.io.IOException {
+ return PARSER.parseDelimitedFrom(input);
+ }
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseDelimitedFrom(
+ java.io.InputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ }
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(
+ com.google.protobuf.CodedInputStream input)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input);
+ }
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input, extensionRegistry);
+ }
+
+ public static Builder newBuilder() { return Builder.create(); }
+ public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto prototype) {
+ return newBuilder().mergeFrom(prototype);
+ }
+ public Builder toBuilder() { return newBuilder(this); }
+
+ @java.lang.Override
+ protected Builder newBuilderForType(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ Builder builder = new Builder(parent);
+ return builder;
+ }
+ /**
+ * Protobuf type {@code QueryIdentifierProto}
+ */
+ public static final class Builder extends
+ com.google.protobuf.GeneratedMessage.Builder<Builder>
+ implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder {
+ public static final com.google.protobuf.Descriptors.Descriptor
+ getDescriptor() {
+ return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_descriptor;
+ }
+
+ protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ internalGetFieldAccessorTable() {
+ return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder.class);
+ }
+
+ // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.newBuilder()
+ private Builder() {
+ maybeForceBuilderInitialization();
+ }
+
+ private Builder(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ super(parent);
+ maybeForceBuilderInitialization();
+ }
+ private void maybeForceBuilderInitialization() {
+ if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ }
+ }
+ private static Builder create() {
+ return new Builder();
+ }
+
+ public Builder clear() {
+ super.clear();
+ appIdentifier_ = "";
+ bitField0_ = (bitField0_ & ~0x00000001);
+ dagIdentifier_ = 0;
+ bitField0_ = (bitField0_ & ~0x00000002);
+ return this;
+ }
+
+ public Builder clone() {
+ return create().mergeFrom(buildPartial());
+ }
+
+ public com.google.protobuf.Descriptors.Descriptor
+ getDescriptorForType() {
+ return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_descriptor;
+ }
+
+ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getDefaultInstanceForType() {
+ return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance();
+ }
+
+ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto build() {
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto result = buildPartial();
+ if (!result.isInitialized()) {
+ throw newUninitializedMessageException(result);
+ }
+ return result;
+ }
+
+ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto buildPartial() {
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto(this);
+ int from_bitField0_ = bitField0_;
+ int to_bitField0_ = 0;
+ if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
+ to_bitField0_ |= 0x00000001;
+ }
+ result.appIdentifier_ = appIdentifier_;
+ if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
+ to_bitField0_ |= 0x00000002;
+ }
+ result.dagIdentifier_ = dagIdentifier_;
+ result.bitField0_ = to_bitField0_;
+ onBuilt();
+ return result;
+ }
+
+ public Builder mergeFrom(com.google.protobuf.Message other) {
+ if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto) {
+ return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto)other);
+ } else {
+ super.mergeFrom(other);
+ return this;
+ }
+ }
+
+ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto other) {
+ if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance()) return this;
+ if (other.hasAppIdentifier()) {
+ bitField0_ |= 0x00000001;
+ appIdentifier_ = other.appIdentifier_;
+ onChanged();
+ }
+ if (other.hasDagIdentifier()) {
+ setDagIdentifier(other.getDagIdentifier());
+ }
+ this.mergeUnknownFields(other.getUnknownFields());
+ return this;
+ }
+
+ public final boolean isInitialized() {
+ return true;
+ }
+
+ public Builder mergeFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parsedMessage = null;
+ try {
+ parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto) e.getUnfinishedMessage();
+ throw e;
+ } finally {
+ if (parsedMessage != null) {
+ mergeFrom(parsedMessage);
+ }
+ }
+ return this;
+ }
+ private int bitField0_;
+
+ // optional string app_identifier = 1;
+ private java.lang.Object appIdentifier_ = "";
+ /**
+ * <code>optional string app_identifier = 1;</code>
+ */
+ public boolean hasAppIdentifier() {
+ return ((bitField0_ & 0x00000001) == 0x00000001);
+ }
+ /**
+ * <code>optional string app_identifier = 1;</code>
+ */
+ public java.lang.String getAppIdentifier() {
+ java.lang.Object ref = appIdentifier_;
+ if (!(ref instanceof java.lang.String)) {
+ java.lang.String s = ((com.google.protobuf.ByteString) ref)
+ .toStringUtf8();
+ appIdentifier_ = s;
+ return s;
+ } else {
+ return (java.lang.String) ref;
+ }
+ }
+ /**
+ * <code>optional string app_identifier = 1;</code>
+ */
+ public com.google.protobuf.ByteString
+ getAppIdentifierBytes() {
+ java.lang.Object ref = appIdentifier_;
+ if (ref instanceof String) {
+ com.google.protobuf.ByteString b =
+ com.google.protobuf.ByteString.copyFromUtf8(
+ (java.lang.String) ref);
+ appIdentifier_ = b;
+ return b;
+ } else {
+ return (com.google.protobuf.ByteString) ref;
+ }
+ }
+ /**
+ * <code>optional string app_identifier = 1;</code>
+ */
+ public Builder setAppIdentifier(
+ java.lang.String value) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ bitField0_ |= 0x00000001;
+ appIdentifier_ = value;
+ onChanged();
+ return this;
+ }
+ /**
+ * <code>optional string app_identifier = 1;</code>
+ */
+ public Builder clearAppIdentifier() {
+ bitField0_ = (bitField0_ & ~0x00000001);
+ appIdentifier_ = getDefaultInstance().getAppIdentifier();
+ onChanged();
+ return this;
+ }
+ /**
+ * <code>optional string app_identifier = 1;</code>
+ */
+ public Builder setAppIdentifierBytes(
+ com.google.protobuf.ByteString value) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ bitField0_ |= 0x00000001;
+ appIdentifier_ = value;
+ onChanged();
+ return this;
+ }
+
+ // optional int32 dag_identifier = 2;
+ private int dagIdentifier_ ;
+ /**
+ * <code>optional int32 dag_identifier = 2;</code>
+ */
+ public boolean hasDagIdentifier() {
+ return ((bitField0_ & 0x00000002) == 0x00000002);
+ }
+ /**
+ * <code>optional int32 dag_identifier = 2;</code>
+ */
+ public int getDagIdentifier() {
+ return dagIdentifier_;
+ }
+ /**
+ * <code>optional int32 dag_identifier = 2;</code>
+ */
+ public Builder setDagIdentifier(int value) {
+ bitField0_ |= 0x00000002;
+ dagIdentifier_ = value;
+ onChanged();
+ return this;
+ }
+ /**
+ * <code>optional int32 dag_identifier = 2;</code>
+ */
+ public Builder clearDagIdentifier() {
+ bitField0_ = (bitField0_ & ~0x00000002);
+ dagIdentifier_ = 0;
+ onChanged();
+ return this;
+ }
+
+ // @@protoc_insertion_point(builder_scope:QueryIdentifierProto)
+ }
+
+ static {
+ defaultInstance = new QueryIdentifierProto(true);
+ defaultInstance.initFields();
+ }
+
+ // @@protoc_insertion_point(class_scope:QueryIdentifierProto)
+ }
+
+ public interface SubmitWorkRequestProtoOrBuilder
+ extends com.google.protobuf.MessageOrBuilder {
+
+ // optional string container_id_string = 1;
+ /**
+ * <code>optional string container_id_string = 1;</code>
+ */
+ boolean hasContainerIdString();
+ /**
+ * <code>optional string container_id_string = 1;</code>
+ */
+ java.lang.String getContainerIdString();
+ /**
+ * <code>optional string container_id_string = 1;</code>
+ */
+ com.google.protobuf.ByteString
+ getContainerIdStringBytes();
+
+ // optional string am_host = 2;
+ /**
+ * <code>optional string am_host = 2;</code>
+ */
+ boolean hasAmHost();
+ /**
+ * <code>optional string am_host = 2;</code>
+ */
+ java.lang.String getAmHost();
+ /**
+ * <code>optional string am_host = 2;</code>
+ */
+ com.google.protobuf.ByteString
+ getAmHostBytes();
+
+ // optional int32 am_port = 3;
+ /**
+ * <code>optional int32 am_port = 3;</code>
+ */
+ boolean hasAmPort();
+ /**
+ * <code>optional int32 am_port = 3;</code>
+ */
+ int getAmPort();
+
+ // optional string token_identifier = 4;
+ /**
+ * <code>optional string token_identifier = 4;</code>
+ */
+ boolean hasTokenIdentifier();
+ /**
+ * <code>optional string token_identifier = 4;</code>
+ */
+ java.lang.String getTokenIdentifier();
+ /**
+ * <code>optional string token_identifier = 4;</code>
+ */
+ com.google.protobuf.ByteString
+ getTokenIdentifierBytes();
+
+ // optional bytes credentials_binary = 5;
+ /**
+ * <code>optional bytes credentials_binary = 5;</code>
+ */
+ boolean hasCredentialsBinary();
+ /**
+ * <code>optional bytes credentials_binary = 5;</code>
+ */
+ com.google.protobuf.ByteString getCredentialsBinary();
+
+ // optional string user = 6;
+ /**
* <code>optional string user = 6;</code>
*/
boolean hasUser();
@@ -8800,20 +9489,19 @@ public final class LlapDaemonProtocolProtos {
public interface SourceStateUpdatedRequestProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
- // optional string dag_name = 1;
+ // optional .QueryIdentifierProto query_identifier = 1;
/**
- * <code>optional string dag_name = 1;</code>
+ * <code>optional .QueryIdentifierProto query_identifier = 1;</code>
*/
- boolean hasDagName();
+ boolean hasQueryIdentifier();
/**
- * <code>optional string dag_name = 1;</code>
+ * <code>optional .QueryIdentifierProto query_identifier = 1;</code>
*/
- java.lang.String getDagName();
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier();
/**
- * <code>optional string dag_name = 1;</code>
+ * <code>optional .QueryIdentifierProto query_identifier = 1;</code>
*/
- com.google.protobuf.ByteString
- getDagNameBytes();
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder();
// optional string src_name = 2;
/**
@@ -8892,8 +9580,16 @@ public final class LlapDaemonProtocolProtos {
break;
}
case 10: {
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder subBuilder = null;
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ subBuilder = queryIdentifier_.toBuilder();
+ }
+ queryIdentifier_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.PARSER, extensionRegistry);
+ if (subBuilder != null) {
+ subBuilder.mergeFrom(queryIdentifier_);
+ queryIdentifier_ = subBuilder.buildPartial();
+ }
bitField0_ |= 0x00000001;
- dagName_ = input.readBytes();
break;
}
case 18: {
@@ -8952,47 +9648,26 @@ public final class LlapDaemonProtocolProtos {
}
private int bitField0_;
- // optional string dag_name = 1;
- public static final int DAG_NAME_FIELD_NUMBER = 1;
- private java.lang.Object dagName_;
+ // optional .QueryIdentifierProto query_identifier = 1;
+ public static final int QUERY_IDENTIFIER_FIELD_NUMBER = 1;
+ private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_;
/**
- * <code>optional string dag_name = 1;</code>
+ * <code>optional .QueryIdentifierProto query_identifier = 1;</code>
*/
- public boolean hasDagName() {
+ public boolean hasQueryIdentifier() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
- * <code>optional string dag_name = 1;</code>
+ * <code>optional .QueryIdentifierProto query_identifier = 1;</code>
*/
- public java.lang.String getDagName() {
- java.lang.Object ref = dagName_;
- if (ref instanceof java.lang.String) {
- return (java.lang.String) ref;
- } else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
- java.lang.String s = bs.toStringUtf8();
- if (bs.isValidUtf8()) {
- dagName_ = s;
- }
- return s;
- }
+ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() {
+ return queryIdentifier_;
}
/**
- * <code>optional string dag_name = 1;</code>
+ * <code>optional .QueryIdentifierProto query_identifier = 1;</code>
*/
- public com.google.protobuf.ByteString
- getDagNameBytes() {
- java.lang.Object ref = dagName_;
- if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
- (java.lang.String) ref);
- dagName_ = b;
- return b;
- } else {
- return (com.google.protobuf.ByteString) ref;
- }
+ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder() {
+ return queryIdentifier_;
}
// optional string src_name = 2;
@@ -9055,7 +9730,7 @@ public final class LlapDaemonProtocolProtos {
}
private void initFields() {
- dagName_ = "";
+ queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance();
srcName_ = "";
state_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto.S_SUCCEEDED;
}
@@ -9072,7 +9747,7 @@ public final class LlapDaemonProtocolProtos {
throws java.io.IOException {
getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- output.writeBytes(1, getDagNameBytes());
+ output.writeMessage(1, queryIdentifier_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
output.writeBytes(2, getSrcNameBytes());
@@ -9091,7 +9766,7 @@ public final class LlapDaemonProtocolProtos {
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
- .computeBytesSize(1, getDagNameBytes());
+ .computeMessageSize(1, queryIdentifier_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
size += com.google.protobuf.CodedOutputStream
@@ -9124,10 +9799,10 @@ public final class LlapDaemonProtocolProtos {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto) obj;
boolean result = true;
- result = result && (hasDagName() == other.hasDagName());
- if (hasDagName()) {
- result = result && getDagName()
- .equals(other.getDagName());
+ result = result && (hasQueryIdentifier() == other.hasQueryIdentifier());
+ if (hasQueryIdentifier()) {
+ result = result && getQueryIdentifier()
+ .equals(other.getQueryIdentifier());
}
result = result && (hasSrcName() == other.hasSrcName());
if (hasSrcName()) {
@@ -9152,9 +9827,9 @@ public final class LlapDaemonProtocolProtos {
}
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
- if (hasDagName()) {
- hash = (37 * hash) + DAG_NAME_FIELD_NUMBER;
- hash = (53 * hash) + getDagName().hashCode();
+ if (hasQueryIdentifier()) {
+ hash = (37 * hash) + QUERY_IDENTIFIER_FIELD_NUMBER;
+ hash = (53 * hash) + getQueryIdentifier().hashCode();
}
if (hasSrcName()) {
hash = (37 * hash) + SRC_NAME_FIELD_NUMBER;
@@ -9265,6 +9940,7 @@ public final class LlapDaemonProtocolProtos {
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ getQueryIdentifierFieldBuilder();
}
}
private static Builder create() {
@@ -9273,7 +9949,11 @@ public final class LlapDaemonProtocolProtos {
public Builder clear() {
super.clear();
- dagName_ = "";
+ if (queryIdentifierBuilder_ == null) {
+ queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance();
+ } else {
+ queryIdentifierBuilder_.clear();
+ }
bitField0_ = (bitField0_ & ~0x00000001);
srcName_ = "";
bitField0_ = (bitField0_ & ~0x00000002);
@@ -9310,7 +9990,11 @@ public final class LlapDaemonProtocolProtos {
if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
to_bitField0_ |= 0x00000001;
}
- result.dagName_ = dagName_;
+ if (queryIdentifierBuilder_ == null) {
+ result.queryIdentifier_ = queryIdentifier_;
+ } else {
+ result.queryIdentifier_ = queryIdentifierBuilder_.build();
+ }
if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
to_bitField0_ |= 0x00000002;
}
@@ -9335,10 +10019,8 @@ public final class LlapDaemonProtocolProtos {
public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto other) {
if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto.getDefaultInstance()) return this;
- if (other.hasDagName()) {
- bitField0_ |= 0x00000001;
- dagName_ = other.dagName_;
- onChanged();
+ if (other.hasQueryIdentifier()) {
+ mergeQueryIdentifier(other.getQueryIdentifier());
}
if (other.hasSrcName()) {
bitField0_ |= 0x00000002;
@@ -9375,78 +10057,121 @@ public final class LlapDaemonProtocolProtos {
}
private int bitField0_;
- // optional string dag_name = 1;
- private java.lang.Object dagName_ = "";
+ // optional .QueryIdentifierProto query_identifier = 1;
+ private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance();
+ private com.google.protobuf.SingleFieldBuilder<
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> queryIdentifierBuilder_;
/**
- * <code>optional string dag_name = 1;</code>
+ * <code>optional .QueryIdentifierProto query_identifier = 1;</code>
*/
- public boolean hasDagName() {
+ public boolean hasQueryIdentifier() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
- * <code>optional string dag_name = 1;</code>
+ * <code>optional .QueryIdentifierProto query_identifier = 1;</code>
*/
- public java.lang.String getDagName() {
- java.lang.Object ref = dagName_;
- if (!(ref instanceof java.lang.String)) {
- java.lang.String s = ((com.google.protobuf.ByteString) ref)
- .toStringUtf8();
- dagName_ = s;
- return s;
+ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() {
+ if (queryIdentifierBuilder_ == null) {
+ return queryIdentifier_;
} else {
- return (java.lang.String) ref;
+ return queryIdentifierBuilder_.getMessage();
}
}
/**
- * <code>optional string dag_name = 1;</code>
+ * <code>optional .QueryIdentifierProto query_identifier = 1;</code>
*/
- public com.google.protobuf.ByteString
- getDagNameBytes() {
- java.lang.Object ref = dagName_;
- if (ref instanceof String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
- (java.lang.String) ref);
- dagName_ = b;
- return b;
+ public Builder setQueryIdentifier(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto value) {
+ if (queryIdentifierBuilder_ == null) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ queryIdentifier_ = value;
+ onChanged();
} else {
- return (com.google.protobuf.ByteString) ref;
+ queryIdentifierBuilder_.setMessage(value);
}
+ bitField0_ |= 0x00000001;
+ return this;
}
/**
- * <code>optional string dag_name = 1;</code>
+ * <code>optional .QueryIdentifierProto query_identifier = 1;</code>
*/
- public Builder setDagName(
- java.lang.String value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000001;
- dagName_ = value;
- onChanged();
+ public Builder setQueryIdentifier(
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder builderForValue) {
+ if (queryIdentifierBuilder_ == null) {
+ queryIdentifier_ = builderForValue.build();
+ onChanged();
+ } else {
+ queryIdentifierBuilder_.setMessage(builderForValue.build());
+ }
+ bitField0_ |= 0x00000001;
+ return this;
+ }
+ /**
+ * <code>optional .QueryIdentifierProto query_identifier = 1;</code>
+ */
+ public Builder mergeQueryIdentifier(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto value) {
+ if (queryIdentifierBuilder_ == null) {
+ if (((bitField0_ & 0x00000001) == 0x00000001) &&
+ queryIdentifier_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance()) {
+ queryIdentifier_ =
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.newBuilder(queryIdentifier_).mergeFrom(value).buildPartial();
+ } else {
+ queryIdentifier_ = value;
+ }
+ onChanged();
+ } else {
+ queryIdentifierBuilder_.mergeFrom(value);
+ }
+ bitField0_ |= 0x00000001;
+ return this;
+ }
+ /**
+ * <code>optional .QueryIdentifierProto query_identifier = 1;</code>
+ */
+ public Builder clearQueryIdentifier() {
+ if (queryIdentifierBuilder_ == null) {
+ queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance();
+ onChanged();
+ } else {
+ queryIdentifierBuilder_.clear();
+ }
+ bitField0_ = (bitField0_ & ~0x00000001);
return this;
}
/**
- * <code>optional string dag_name = 1;</code>
+ * <code>optional .QueryIdentifierProto query_identifier = 1;</code>
*/
- public Builder clearDagName() {
- bitField0_ = (bitField0_ & ~0x00000001);
- dagName_ = getDefaultInstance().getDagName();
+ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder getQueryIdentifierBuilder() {
+ bitField0_ |= 0x00000001;
onChanged();
- return this;
+ return getQueryIdentifierFieldBuilder().getBuilder();
}
/**
- * <code>optional string dag_name = 1;</code>
+ * <code>optional .QueryIdentifierProto query_identifier = 1;</code>
*/
- public Builder setDagNameBytes(
- com.google.protobuf.ByteString value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000001;
- dagName_ = value;
- onChanged();
- return this;
+ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder() {
+ if (queryIdentifierBuilder_ != null) {
+ return queryIdentifierBuilder_.getMessageOrBuilder();
+ } else {
+ return queryIdentifier_;
+ }
+ }
+ /**
+ * <code>optional .QueryIdentifierProto query_identifier = 1;</code>
+ */
+ private com.google.protobuf.SingleFieldBuilder<
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder>
+ getQueryIdentifierFieldBuilder() {
+ if (queryIdentifierBuilder_ == null) {
+ queryIdentifierBuilder_ = new com.google.protobuf.SingleFieldBuilder<
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder>(
+ queryIdentifier_,
+ getParentForChildren(),
+ isClean());
+ queryIdentifier_ = null;
+ }
+ return queryIdentifierBuilder_;
}
// optional string src_name = 2;
@@ -9926,28 +10651,27 @@ public final class LlapDaemonProtocolProtos {
com.google.protobuf.ByteString
getQueryIdBytes();
- // optional string dag_name = 2;
+ // optional .QueryIdentifierProto query_identifier = 2;
/**
- * <code>optional string dag_name = 2;</code>
+ * <code>optional .QueryIdentifierProto query_identifier = 2;</code>
*/
- boolean hasDagName();
+ boolean hasQueryIdentifier();
/**
- * <code>optional string dag_name = 2;</code>
+ * <code>optional .QueryIdentifierProto query_identifier = 2;</code>
*/
- java.lang.String getDagName();
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier();
/**
- * <code>optional string dag_name = 2;</code>
+ * <code>optional .QueryIdentifierProto query_identifier = 2;</code>
*/
- com.google.protobuf.ByteString
- getDagNameBytes();
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder();
- // optional int64 delete_delay = 3 [default = 0];
+ // optional int64 delete_delay = 4 [default = 0];
/**
- * <code>optional int64 delete_delay = 3 [default = 0];</code>
+ * <code>optional int64 delete_delay = 4 [default = 0];</code>
*/
boolean hasDeleteDelay();
/**
- * <code>optional int64 delete_delay = 3 [default = 0];</code>
+ * <code>optional int64 delete_delay = 4 [default = 0];</code>
*/
long getDeleteDelay();
}
@@ -10008,11 +10732,19 @@ public final class LlapDaemonProtocolProtos {
break;
}
case 18: {
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder subBuilder = null;
+ if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ subBuilder = queryIdentifier_.toBuilder();
+ }
+ queryIdentifier_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.PARSER, extensionRegistry);
+ if (subBuilder != null) {
+ subBuilder.mergeFrom(queryIdentifier_);
+ queryIdentifier_ = subBuilder.buildPartial();
+ }
bitField0_ |= 0x00000002;
- dagName_ = input.readBytes();
break;
}
- case 24: {
+ case 32: {
bitField0_ |= 0x00000004;
deleteDelay_ = input.readInt64();
break;
@@ -10100,60 +10832,39 @@ public final class LlapDaemonProtocolProtos {
}
}
- // optional string dag_name = 2;
- public static final int DAG_NAME_FIELD_NUMBER = 2;
- private java.lang.Object dagName_;
+ // optional .QueryIdentifierProto query_identifier = 2;
+ public static final int QUERY_IDENTIFIER_FIELD_NUMBER = 2;
+ private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_;
/**
- * <code>optional string dag_name = 2;</code>
+ * <code>optional .QueryIdentifierProto query_identifier = 2;</code>
*/
- public boolean hasDagName() {
+ public boolean hasQueryIdentifier() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
/**
- * <code>optional string dag_name = 2;</code>
+ * <code>optional .QueryIdentifierProto query_identifier = 2;</code>
*/
- public java.lang.String getDagName() {
- java.lang.Object ref = dagName_;
- if (ref instanceof java.lang.String) {
- return (java.lang.String) ref;
- } else {
- com.google.protobuf.ByteString bs =
- (com.google.protobuf.ByteString) ref;
- java.lang.String s = bs.toStringUtf8();
- if (bs.isValidUtf8()) {
- dagName_ = s;
- }
- return s;
- }
+ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() {
+ return queryIdentifier_;
}
/**
- * <code>optional string dag_name = 2;</code>
+ * <code>optional .QueryIdentifierProto query_identifier = 2;</code>
*/
- public com.google.protobuf.ByteString
- getDagNameBytes() {
- java.lang.Object ref = dagName_;
- if (ref instanceof java.lang.String) {
- com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8(
- (java.lang.String) ref);
- dagName_ = b;
- return b;
- } else {
- return (com.google.protobuf.ByteString) ref;
- }
+ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder() {
+ return queryIdentifier_;
}
- // optional int64 delete_delay = 3 [default = 0];
- public static final int DELETE_DELAY_FIELD_NUMBER = 3;
+ // optional int64 delete_delay = 4 [default = 0];
+ public static final int DELETE_DELAY_FIELD_NUMBER = 4;
private long deleteDelay_;
/**
- * <code>optional int64 delete_delay = 3 [default = 0];</code>
+ * <code>optional int64 delete_delay = 4 [default = 0];</code>
*/
public boolean hasDeleteDelay() {
return ((bitField0_ & 0x00000004) == 0x00000004);
}
/**
- * <code>optional int64 delete_delay = 3 [default = 0];</code>
+ * <code>optional int64 delete_delay = 4 [default = 0];</code>
*/
public long getDeleteDelay() {
return deleteDelay_;
@@ -10161,7 +10872,7 @@ public final class LlapDaemonProtocolProtos {
private void initFields() {
queryId_ = "";
- dagName_ = "";
+ queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance();
deleteDelay_ = 0L;
}
private byte memoizedIsInitialized = -1;
@@ -10180,10 +10891,10 @@ public final class LlapDaemonProtocolProtos {
output.writeBytes(1, getQueryIdBytes());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- output.writeBytes(2, getDagNameBytes());
+ output.writeMessage(2, queryIdentifier_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
- output.writeInt64(3, deleteDelay_);
+ output.writeInt64(4, deleteDelay_);
}
getUnknownFields().writeTo(output);
}
@@ -10200,11 +10911,11 @@ public final class LlapDaemonProtocolProtos {
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
size += com.google.protobuf.CodedOutputStream
- .computeBytesSize(2, getDagNameBytes());
+ .computeMessageSize(2, queryIdentifier_);
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
size += com.google.protobuf.CodedOutputStream
- .computeInt64Size(3, deleteDelay_);
+ .computeInt64Size(4, deleteDelay_);
}
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
@@ -10234,10 +10945,10 @@ public final class LlapDaemonProtocolProtos {
result = result && getQueryId()
.equals(other.getQueryId());
}
- result = result && (hasDagName() == other.hasDagName());
- if (hasDagName()) {
- result = result && getDagName()
- .equals(other.getDagName());
+ result = result && (hasQueryIdentifier() == other.hasQueryIdentifier());
+ if (hasQueryIdentifier()) {
+ result = result && getQueryIdentifier()
+ .equals(other.getQueryIdentifier());
}
result = result && (hasDeleteDelay() == other.hasDeleteDelay());
if (hasDeleteDelay()) {
@@ -10261,9 +10972,9 @@ public final class LlapDaemonProtocolProtos {
hash = (37 * hash) + QUERY_ID_FIELD_NUMBER;
hash = (53 * hash) + getQueryId().hashCode();
}
- if (hasDagName()) {
- hash = (37 * hash) + DAG_NAME_FIELD_NUMBER;
- hash = (53 * hash) + getDagName().hashCode();
+ if (hasQueryIdentifier()) {
+ hash = (37 * hash) + QUERY_IDENTIFIER_FIELD_NUMBER;
+ hash = (53 * hash) + getQueryIdentifier().hashCode();
}
if (hasDeleteDelay()) {
hash = (37 * hash) + DELETE_DELAY_FIELD_NUMBER;
@@ -10370,6 +11081,7 @@ public final class LlapDaemonProtocolProtos {
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ getQueryIdentifierFieldBuilder();
}
}
private static Builder create() {
@@ -10380,7 +11092,11 @@ public final class LlapDaemonProtocolProtos {
super.clear();
queryId_ = "";
bitField0_ = (bitField0_ & ~0x00000001);
- dagName_ = "";
+ if (queryIdentifierBuilder_ == null) {
+ queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance();
+ } else {
+ queryIdentifierBuilder_.clear();
+ }
bitField0_ = (bitField0_ & ~0x00000002);
deleteDelay_ = 0L;
bitField0_ = (bitField0_ & ~0x00000004);
@@ -10419,7 +11135,11 @@ public final class LlapDaemonProtocolProtos {
if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
to_bitField0_ |= 0x00000002;
}
- result.dagName_ = dagName_;
+ if (queryIdentifierBuilder_ == null) {
+ result.queryIdentifier_ = queryIdentifier_;
+ } else {
+ result.queryIdentifier_ = queryIdentifierBuilder_.build();
+ }
if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
to_bitField0_ |= 0x00000004;
}
@@ -10445,10 +11165,8 @@ public final class LlapDaemonProtocolProtos {
queryId_ = other.queryId_;
onChanged();
}
- if (other.hasDagName()) {
- bitField0_ |= 0x00000002;
- dagName_ = other.dagName_;
- onChanged();
+ if (other.hasQueryIdentifier()) {
+ mergeQueryIdentifier(other.getQueryIdentifier());
}
if (other.hasDeleteDelay()) {
setDeleteDelay(other.getDeleteDelay());
@@ -10554,96 +11272,139 @@ public final class LlapDaemonProtocolProtos {
return this;
}
- // optional string dag_name = 2;
- private java.lang.Object dagName_ = "";
+ // optional .QueryIdentifierProto query_identifier = 2;
+ private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance();
+ private com.google.protobuf.SingleFieldBuilder<
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> queryIdentifierBuilder_;
/**
- * <code>optional string dag_name = 2;</code>
+ * <code>optional .QueryIdentifierProto query_
<TRUNCATED>