You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by ay...@apache.org on 2024/03/19 06:53:18 UTC
(hive) branch master updated: HIVE-28123: Add Generated Protobuf code for 3.24.4 Upgrade (#5132). (Indhumathi Muthumurugesh, reviewed by Ayush Saxena, Tanishq Chugh)
This is an automated email from the ASF dual-hosted git repository.
ayushsaxena pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hive.git
The following commit(s) were added to refs/heads/master by this push:
new 5256d3c2669 HIVE-28123: Add Generated Protobuf code for 3.24.4 Upgrade (#5132). (Indhumathi Muthumurugesh, reviewed by Ayush Saxena, Tanishq Chugh)
5256d3c2669 is described below
commit 5256d3c2669fb523913dd79c8878c3fa94abdbde
Author: Indhumathi27 <34...@users.noreply.github.com>
AuthorDate: Tue Mar 19 12:23:12 2024 +0530
HIVE-28123: Add Generated Protobuf code for 3.24.4 Upgrade (#5132). (Indhumathi Muthumurugesh, reviewed by Ayush Saxena, Tanishq Chugh)
---
.../llap/daemon/rpc/LlapDaemonProtocolProtos.java | 21594 +++++++++++--------
.../llap/plugin/rpc/LlapPluginProtocolProtos.java | 749 +-
.../hadoop/hive/ql/hooks/proto/HiveHookEvents.java | 1908 +-
.../hadoop/hive/ql/io/protobuf/SampleProtos.java | 2955 ++-
4 files changed, 15983 insertions(+), 11223 deletions(-)
diff --git a/llap-common/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java b/llap-common/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java
index b50448053bf..dafa1e7a78d 100644
--- a/llap-common/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java
+++ b/llap-common/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java
@@ -5,8 +5,14 @@ package org.apache.hadoop.hive.llap.daemon.rpc;
public final class LlapDaemonProtocolProtos {
private LlapDaemonProtocolProtos() {}
+ public static void registerAllExtensions(
+ com.google.protobuf.ExtensionRegistryLite registry) {
+ }
+
public static void registerAllExtensions(
com.google.protobuf.ExtensionRegistry registry) {
+ registerAllExtensions(
+ (com.google.protobuf.ExtensionRegistryLite) registry);
}
/**
* Protobuf enum {@code SourceStateProto}
@@ -16,11 +22,11 @@ public final class LlapDaemonProtocolProtos {
/**
* <code>S_SUCCEEDED = 1;</code>
*/
- S_SUCCEEDED(0, 1),
+ S_SUCCEEDED(1),
/**
* <code>S_RUNNING = 2;</code>
*/
- S_RUNNING(1, 2),
+ S_RUNNING(2),
;
/**
@@ -33,9 +39,25 @@ public final class LlapDaemonProtocolProtos {
public static final int S_RUNNING_VALUE = 2;
- public final int getNumber() { return value; }
+ public final int getNumber() {
+ return value;
+ }
+ /**
+ * @param value The numeric wire value of the corresponding enum entry.
+ * @return The enum associated with the given numeric wire value.
+ * @deprecated Use {@link #forNumber(int)} instead.
+ */
+ @java.lang.Deprecated
public static SourceStateProto valueOf(int value) {
+ return forNumber(value);
+ }
+
+ /**
+ * @param value The numeric wire value of the corresponding enum entry.
+ * @return The enum associated with the given numeric wire value.
+ */
+ public static SourceStateProto forNumber(int value) {
switch (value) {
case 1: return S_SUCCEEDED;
case 2: return S_RUNNING;
@@ -47,17 +69,17 @@ public final class LlapDaemonProtocolProtos {
internalGetValueMap() {
return internalValueMap;
}
- private static com.google.protobuf.Internal.EnumLiteMap<SourceStateProto>
- internalValueMap =
+ private static final com.google.protobuf.Internal.EnumLiteMap<
+ SourceStateProto> internalValueMap =
new com.google.protobuf.Internal.EnumLiteMap<SourceStateProto>() {
public SourceStateProto findValueByNumber(int number) {
- return SourceStateProto.valueOf(number);
+ return SourceStateProto.forNumber(number);
}
};
public final com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
- return getDescriptor().getValues().get(index);
+ return getDescriptor().getValues().get(ordinal());
}
public final com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
@@ -79,11 +101,9 @@ public final class LlapDaemonProtocolProtos {
return VALUES[desc.getIndex()];
}
- private final int index;
private final int value;
- private SourceStateProto(int index, int value) {
- this.index = index;
+ private SourceStateProto(int value) {
this.value = value;
}
@@ -98,15 +118,15 @@ public final class LlapDaemonProtocolProtos {
/**
* <code>ACCEPTED = 1;</code>
*/
- ACCEPTED(0, 1),
+ ACCEPTED(1),
/**
* <code>REJECTED = 2;</code>
*/
- REJECTED(1, 2),
+ REJECTED(2),
/**
* <code>EVICTED_OTHER = 3;</code>
*/
- EVICTED_OTHER(2, 3),
+ EVICTED_OTHER(3),
;
/**
@@ -123,9 +143,25 @@ public final class LlapDaemonProtocolProtos {
public static final int EVICTED_OTHER_VALUE = 3;
- public final int getNumber() { return value; }
+ public final int getNumber() {
+ return value;
+ }
+ /**
+ * @param value The numeric wire value of the corresponding enum entry.
+ * @return The enum associated with the given numeric wire value.
+ * @deprecated Use {@link #forNumber(int)} instead.
+ */
+ @java.lang.Deprecated
public static SubmissionStateProto valueOf(int value) {
+ return forNumber(value);
+ }
+
+ /**
+ * @param value The numeric wire value of the corresponding enum entry.
+ * @return The enum associated with the given numeric wire value.
+ */
+ public static SubmissionStateProto forNumber(int value) {
switch (value) {
case 1: return ACCEPTED;
case 2: return REJECTED;
@@ -138,17 +174,17 @@ public final class LlapDaemonProtocolProtos {
internalGetValueMap() {
return internalValueMap;
}
- private static com.google.protobuf.Internal.EnumLiteMap<SubmissionStateProto>
- internalValueMap =
+ private static final com.google.protobuf.Internal.EnumLiteMap<
+ SubmissionStateProto> internalValueMap =
new com.google.protobuf.Internal.EnumLiteMap<SubmissionStateProto>() {
public SubmissionStateProto findValueByNumber(int number) {
- return SubmissionStateProto.valueOf(number);
+ return SubmissionStateProto.forNumber(number);
}
};
public final com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
- return getDescriptor().getValues().get(index);
+ return getDescriptor().getValues().get(ordinal());
}
public final com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
@@ -170,37 +206,38 @@ public final class LlapDaemonProtocolProtos {
return VALUES[desc.getIndex()];
}
- private final int index;
private final int value;
- private SubmissionStateProto(int index, int value) {
- this.index = index;
+ private SubmissionStateProto(int value) {
this.value = value;
}
// @@protoc_insertion_point(enum_scope:SubmissionStateProto)
}
- public interface UserPayloadProtoOrBuilder
- extends com.google.protobuf.MessageOrBuilder {
+ public interface UserPayloadProtoOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:UserPayloadProto)
+ com.google.protobuf.MessageOrBuilder {
- // optional bytes user_payload = 1;
/**
* <code>optional bytes user_payload = 1;</code>
+ * @return Whether the userPayload field is set.
*/
boolean hasUserPayload();
/**
* <code>optional bytes user_payload = 1;</code>
+ * @return The userPayload.
*/
com.google.protobuf.ByteString getUserPayload();
- // optional int32 version = 2;
/**
* <code>optional int32 version = 2;</code>
+ * @return Whether the version field is set.
*/
boolean hasVersion();
/**
* <code>optional int32 version = 2;</code>
+ * @return The version.
*/
int getVersion();
}
@@ -208,186 +245,119 @@ public final class LlapDaemonProtocolProtos {
* Protobuf type {@code UserPayloadProto}
*/
public static final class UserPayloadProto extends
- com.google.protobuf.GeneratedMessage
- implements UserPayloadProtoOrBuilder {
+ com.google.protobuf.GeneratedMessageV3 implements
+ // @@protoc_insertion_point(message_implements:UserPayloadProto)
+ UserPayloadProtoOrBuilder {
+ private static final long serialVersionUID = 0L;
// Use UserPayloadProto.newBuilder() to construct.
- private UserPayloadProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ private UserPayloadProto(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
- this.unknownFields = builder.getUnknownFields();
- }
- private UserPayloadProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
- private static final UserPayloadProto defaultInstance;
- public static UserPayloadProto getDefaultInstance() {
- return defaultInstance;
}
-
- public UserPayloadProto getDefaultInstanceForType() {
- return defaultInstance;
+ private UserPayloadProto() {
+ userPayload_ = com.google.protobuf.ByteString.EMPTY;
}
- private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
- return this.unknownFields;
- }
- private UserPayloadProto(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- initFields();
- int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
- try {
- boolean done = false;
- while (!done) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
- done = true;
- break;
- default: {
- if (!parseUnknownField(input, unknownFields,
- extensionRegistry, tag)) {
- done = true;
- }
- break;
- }
- case 10: {
- bitField0_ |= 0x00000001;
- userPayload_ = input.readBytes();
- break;
- }
- case 16: {
- bitField0_ |= 0x00000002;
- version_ = input.readInt32();
- break;
- }
- }
- }
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
- throw e.setUnfinishedMessage(this);
- } catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
- e.getMessage()).setUnfinishedMessage(this);
- } finally {
- this.unknownFields = unknownFields.build();
- makeExtensionsImmutable();
- }
+ @SuppressWarnings({"unused"})
+ protected java.lang.Object newInstance(
+ UnusedPrivateParameter unused) {
+ return new UserPayloadProto();
}
+
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ @java.lang.Override
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder.class);
}
- public static com.google.protobuf.Parser<UserPayloadProto> PARSER =
- new com.google.protobuf.AbstractParser<UserPayloadProto>() {
- public UserPayloadProto parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return new UserPayloadProto(input, extensionRegistry);
- }
- };
-
- @java.lang.Override
- public com.google.protobuf.Parser<UserPayloadProto> getParserForType() {
- return PARSER;
- }
-
private int bitField0_;
- // optional bytes user_payload = 1;
public static final int USER_PAYLOAD_FIELD_NUMBER = 1;
- private com.google.protobuf.ByteString userPayload_;
+ private com.google.protobuf.ByteString userPayload_ = com.google.protobuf.ByteString.EMPTY;
/**
* <code>optional bytes user_payload = 1;</code>
+ * @return Whether the userPayload field is set.
*/
+ @java.lang.Override
public boolean hasUserPayload() {
- return ((bitField0_ & 0x00000001) == 0x00000001);
+ return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bytes user_payload = 1;</code>
+ * @return The userPayload.
*/
+ @java.lang.Override
public com.google.protobuf.ByteString getUserPayload() {
return userPayload_;
}
- // optional int32 version = 2;
public static final int VERSION_FIELD_NUMBER = 2;
- private int version_;
+ private int version_ = 0;
/**
* <code>optional int32 version = 2;</code>
+ * @return Whether the version field is set.
*/
+ @java.lang.Override
public boolean hasVersion() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
+ return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional int32 version = 2;</code>
+ * @return The version.
*/
+ @java.lang.Override
public int getVersion() {
return version_;
}
- private void initFields() {
- userPayload_ = com.google.protobuf.ByteString.EMPTY;
- version_ = 0;
- }
private byte memoizedIsInitialized = -1;
+ @java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
- if (isInitialized != -1) return isInitialized == 1;
+ if (isInitialized == 1) return true;
+ if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
+ @java.lang.Override
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
- getSerializedSize();
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ if (((bitField0_ & 0x00000001) != 0)) {
output.writeBytes(1, userPayload_);
}
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ if (((bitField0_ & 0x00000002) != 0)) {
output.writeInt32(2, version_);
}
getUnknownFields().writeTo(output);
}
- private int memoizedSerializedSize = -1;
+ @java.lang.Override
public int getSerializedSize() {
- int size = memoizedSerializedSize;
+ int size = memoizedSize;
if (size != -1) return size;
size = 0;
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ if (((bitField0_ & 0x00000001) != 0)) {
size += com.google.protobuf.CodedOutputStream
.computeBytesSize(1, userPayload_);
}
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ if (((bitField0_ & 0x00000002) != 0)) {
size += com.google.protobuf.CodedOutputStream
.computeInt32Size(2, version_);
}
size += getUnknownFields().getSerializedSize();
- memoizedSerializedSize = size;
+ memoizedSize = size;
return size;
}
- private static final long serialVersionUID = 0L;
- @java.lang.Override
- protected java.lang.Object writeReplace()
- throws java.io.ObjectStreamException {
- return super.writeReplace();
- }
-
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
@@ -398,30 +368,27 @@ public final class LlapDaemonProtocolProtos {
}
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto) obj;
- boolean result = true;
- result = result && (hasUserPayload() == other.hasUserPayload());
+ if (hasUserPayload() != other.hasUserPayload()) return false;
if (hasUserPayload()) {
- result = result && getUserPayload()
- .equals(other.getUserPayload());
+ if (!getUserPayload()
+ .equals(other.getUserPayload())) return false;
}
- result = result && (hasVersion() == other.hasVersion());
+ if (hasVersion() != other.hasVersion()) return false;
if (hasVersion()) {
- result = result && (getVersion()
- == other.getVersion());
+ if (getVersion()
+ != other.getVersion()) return false;
}
- result = result &&
- getUnknownFields().equals(other.getUnknownFields());
- return result;
+ if (!getUnknownFields().equals(other.getUnknownFields())) return false;
+ return true;
}
- private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
- hash = (19 * hash) + getDescriptorForType().hashCode();
+ hash = (19 * hash) + getDescriptor().hashCode();
if (hasUserPayload()) {
hash = (37 * hash) + USER_PAYLOAD_FIELD_NUMBER;
hash = (53 * hash) + getUserPayload().hashCode();
@@ -435,6 +402,17 @@ public final class LlapDaemonProtocolProtos {
return hash;
}
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom(
+ java.nio.ByteBuffer data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom(
+ java.nio.ByteBuffer data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
@@ -458,46 +436,61 @@ public final class LlapDaemonProtocolProtos {
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
+
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input);
}
+
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
- public static Builder newBuilder() { return Builder.create(); }
+ @java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder() {
+ return DEFAULT_INSTANCE.toBuilder();
+ }
public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto prototype) {
- return newBuilder().mergeFrom(prototype);
+ return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+ }
+ @java.lang.Override
+ public Builder toBuilder() {
+ return this == DEFAULT_INSTANCE
+ ? new Builder() : new Builder().mergeFrom(this);
}
- public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -505,14 +498,16 @@ public final class LlapDaemonProtocolProtos {
* Protobuf type {@code UserPayloadProto}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder {
+ com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ // @@protoc_insertion_point(builder_implements:UserPayloadProto)
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ @java.lang.Override
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -521,44 +516,35 @@ public final class LlapDaemonProtocolProtos {
// Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.newBuilder()
private Builder() {
- maybeForceBuilderInitialization();
+
}
private Builder(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
- maybeForceBuilderInitialization();
- }
- private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
- }
- }
- private static Builder create() {
- return new Builder();
- }
+ }
+ @java.lang.Override
public Builder clear() {
super.clear();
+ bitField0_ = 0;
userPayload_ = com.google.protobuf.ByteString.EMPTY;
- bitField0_ = (bitField0_ & ~0x00000001);
version_ = 0;
- bitField0_ = (bitField0_ & ~0x00000002);
return this;
}
- public Builder clone() {
- return create().mergeFrom(buildPartial());
- }
-
+ @java.lang.Override
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_descriptor;
}
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getDefaultInstanceForType() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance();
}
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto build() {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto result = buildPartial();
if (!result.isInitialized()) {
@@ -567,23 +553,61 @@ public final class LlapDaemonProtocolProtos {
return result;
}
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto buildPartial() {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto(this);
+ if (bitField0_ != 0) { buildPartial0(result); }
+ onBuilt();
+ return result;
+ }
+
+ private void buildPartial0(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
- if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
+ if (((from_bitField0_ & 0x00000001) != 0)) {
+ result.userPayload_ = userPayload_;
to_bitField0_ |= 0x00000001;
}
- result.userPayload_ = userPayload_;
- if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
+ if (((from_bitField0_ & 0x00000002) != 0)) {
+ result.version_ = version_;
to_bitField0_ |= 0x00000002;
}
- result.version_ = version_;
- result.bitField0_ = to_bitField0_;
- onBuilt();
- return result;
+ result.bitField0_ |= to_bitField0_;
}
+ @java.lang.Override
+ public Builder clone() {
+ return super.clone();
+ }
+ @java.lang.Override
+ public Builder setField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ java.lang.Object value) {
+ return super.setField(field, value);
+ }
+ @java.lang.Override
+ public Builder clearField(
+ com.google.protobuf.Descriptors.FieldDescriptor field) {
+ return super.clearField(field);
+ }
+ @java.lang.Override
+ public Builder clearOneof(
+ com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ return super.clearOneof(oneof);
+ }
+ @java.lang.Override
+ public Builder setRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ int index, java.lang.Object value) {
+ return super.setRepeatedField(field, index, value);
+ }
+ @java.lang.Override
+ public Builder addRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ java.lang.Object value) {
+ return super.addRepeatedField(field, value);
+ }
+ @java.lang.Override
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto) {
return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto)other);
@@ -602,60 +626,90 @@ public final class LlapDaemonProtocolProtos {
setVersion(other.getVersion());
}
this.mergeUnknownFields(other.getUnknownFields());
+ onChanged();
return this;
}
+ @java.lang.Override
public final boolean isInitialized() {
return true;
}
+ @java.lang.Override
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parsedMessage = null;
+ if (extensionRegistry == null) {
+ throw new java.lang.NullPointerException();
+ }
try {
- parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ case 10: {
+ userPayload_ = input.readBytes();
+ bitField0_ |= 0x00000001;
+ break;
+ } // case 10
+ case 16: {
+ version_ = input.readInt32();
+ bitField0_ |= 0x00000002;
+ break;
+ } // case 16
+ default: {
+ if (!super.parseUnknownField(input, extensionRegistry, tag)) {
+ done = true; // was an endgroup tag
+ }
+ break;
+ } // default:
+ } // switch (tag)
+ } // while (!done)
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
- parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto) e.getUnfinishedMessage();
- throw e;
+ throw e.unwrapIOException();
} finally {
- if (parsedMessage != null) {
- mergeFrom(parsedMessage);
- }
- }
+ onChanged();
+ } // finally
return this;
}
private int bitField0_;
- // optional bytes user_payload = 1;
private com.google.protobuf.ByteString userPayload_ = com.google.protobuf.ByteString.EMPTY;
/**
* <code>optional bytes user_payload = 1;</code>
+ * @return Whether the userPayload field is set.
*/
+ @java.lang.Override
public boolean hasUserPayload() {
- return ((bitField0_ & 0x00000001) == 0x00000001);
+ return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bytes user_payload = 1;</code>
+ * @return The userPayload.
*/
+ @java.lang.Override
public com.google.protobuf.ByteString getUserPayload() {
return userPayload_;
}
/**
* <code>optional bytes user_payload = 1;</code>
+ * @param value The userPayload to set.
+ * @return This builder for chaining.
*/
public Builder setUserPayload(com.google.protobuf.ByteString value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000001;
+ if (value == null) { throw new NullPointerException(); }
userPayload_ = value;
+ bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional bytes user_payload = 1;</code>
+ * @return This builder for chaining.
*/
public Builder clearUserPayload() {
bitField0_ = (bitField0_ & ~0x00000001);
@@ -664,31 +718,38 @@ public final class LlapDaemonProtocolProtos {
return this;
}
- // optional int32 version = 2;
private int version_ ;
/**
* <code>optional int32 version = 2;</code>
+ * @return Whether the version field is set.
*/
+ @java.lang.Override
public boolean hasVersion() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
+ return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional int32 version = 2;</code>
+ * @return The version.
*/
+ @java.lang.Override
public int getVersion() {
return version_;
}
/**
* <code>optional int32 version = 2;</code>
+ * @param value The version to set.
+ * @return This builder for chaining.
*/
public Builder setVersion(int value) {
- bitField0_ |= 0x00000002;
+
version_ = value;
+ bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* <code>optional int32 version = 2;</code>
+ * @return This builder for chaining.
*/
public Builder clearVersion() {
bitField0_ = (bitField0_ & ~0x00000002);
@@ -696,43 +757,99 @@ public final class LlapDaemonProtocolProtos {
onChanged();
return this;
}
+ @java.lang.Override
+ public final Builder setUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.setUnknownFields(unknownFields);
+ }
+
+ @java.lang.Override
+ public final Builder mergeUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.mergeUnknownFields(unknownFields);
+ }
+
// @@protoc_insertion_point(builder_scope:UserPayloadProto)
}
+ // @@protoc_insertion_point(class_scope:UserPayloadProto)
+ private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto DEFAULT_INSTANCE;
static {
- defaultInstance = new UserPayloadProto(true);
- defaultInstance.initFields();
+ DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto();
+ }
+
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getDefaultInstance() {
+ return DEFAULT_INSTANCE;
+ }
+
+ @java.lang.Deprecated public static final com.google.protobuf.Parser<UserPayloadProto>
+ PARSER = new com.google.protobuf.AbstractParser<UserPayloadProto>() {
+ @java.lang.Override
+ public UserPayloadProto parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ Builder builder = newBuilder();
+ try {
+ builder.mergeFrom(input, extensionRegistry);
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(builder.buildPartial());
+ } catch (com.google.protobuf.UninitializedMessageException e) {
+ throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(e)
+ .setUnfinishedMessage(builder.buildPartial());
+ }
+ return builder.buildPartial();
+ }
+ };
+
+ public static com.google.protobuf.Parser<UserPayloadProto> parser() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<UserPayloadProto> getParserForType() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getDefaultInstanceForType() {
+ return DEFAULT_INSTANCE;
}
- // @@protoc_insertion_point(class_scope:UserPayloadProto)
}
- public interface EntityDescriptorProtoOrBuilder
- extends com.google.protobuf.MessageOrBuilder {
+ public interface EntityDescriptorProtoOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:EntityDescriptorProto)
+ com.google.protobuf.MessageOrBuilder {
- // optional string class_name = 1;
/**
* <code>optional string class_name = 1;</code>
+ * @return Whether the className field is set.
*/
boolean hasClassName();
/**
* <code>optional string class_name = 1;</code>
+ * @return The className.
*/
java.lang.String getClassName();
/**
* <code>optional string class_name = 1;</code>
+ * @return The bytes for className.
*/
com.google.protobuf.ByteString
getClassNameBytes();
- // optional .UserPayloadProto user_payload = 2;
/**
* <code>optional .UserPayloadProto user_payload = 2;</code>
+ * @return Whether the userPayload field is set.
*/
boolean hasUserPayload();
/**
* <code>optional .UserPayloadProto user_payload = 2;</code>
+ * @return The userPayload.
*/
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getUserPayload();
/**
@@ -740,13 +857,14 @@ public final class LlapDaemonProtocolProtos {
*/
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder getUserPayloadOrBuilder();
- // optional bytes history_text = 3;
/**
* <code>optional bytes history_text = 3;</code>
+ * @return Whether the historyText field is set.
*/
boolean hasHistoryText();
/**
* <code>optional bytes history_text = 3;</code>
+ * @return The historyText.
*/
com.google.protobuf.ByteString getHistoryText();
}
@@ -754,128 +872,56 @@ public final class LlapDaemonProtocolProtos {
* Protobuf type {@code EntityDescriptorProto}
*/
public static final class EntityDescriptorProto extends
- com.google.protobuf.GeneratedMessage
- implements EntityDescriptorProtoOrBuilder {
+ com.google.protobuf.GeneratedMessageV3 implements
+ // @@protoc_insertion_point(message_implements:EntityDescriptorProto)
+ EntityDescriptorProtoOrBuilder {
+ private static final long serialVersionUID = 0L;
// Use EntityDescriptorProto.newBuilder() to construct.
- private EntityDescriptorProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ private EntityDescriptorProto(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
- this.unknownFields = builder.getUnknownFields();
- }
- private EntityDescriptorProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
- private static final EntityDescriptorProto defaultInstance;
- public static EntityDescriptorProto getDefaultInstance() {
- return defaultInstance;
}
-
- public EntityDescriptorProto getDefaultInstanceForType() {
- return defaultInstance;
+ private EntityDescriptorProto() {
+ className_ = "";
+ historyText_ = com.google.protobuf.ByteString.EMPTY;
}
- private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
- return this.unknownFields;
- }
- private EntityDescriptorProto(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- initFields();
- int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
- try {
- boolean done = false;
- while (!done) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
- done = true;
- break;
- default: {
- if (!parseUnknownField(input, unknownFields,
- extensionRegistry, tag)) {
- done = true;
- }
- break;
- }
- case 10: {
- bitField0_ |= 0x00000001;
- className_ = input.readBytes();
- break;
- }
- case 18: {
- org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder subBuilder = null;
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
- subBuilder = userPayload_.toBuilder();
- }
- userPayload_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.PARSER, extensionRegistry);
- if (subBuilder != null) {
- subBuilder.mergeFrom(userPayload_);
- userPayload_ = subBuilder.buildPartial();
- }
- bitField0_ |= 0x00000002;
- break;
- }
- case 26: {
- bitField0_ |= 0x00000004;
- historyText_ = input.readBytes();
- break;
- }
- }
- }
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
- throw e.setUnfinishedMessage(this);
- } catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
- e.getMessage()).setUnfinishedMessage(this);
- } finally {
- this.unknownFields = unknownFields.build();
- makeExtensionsImmutable();
- }
+ @SuppressWarnings({"unused"})
+ protected java.lang.Object newInstance(
+ UnusedPrivateParameter unused) {
+ return new EntityDescriptorProto();
}
+
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ @java.lang.Override
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder.class);
}
- public static com.google.protobuf.Parser<EntityDescriptorProto> PARSER =
- new com.google.protobuf.AbstractParser<EntityDescriptorProto>() {
- public EntityDescriptorProto parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return new EntityDescriptorProto(input, extensionRegistry);
- }
- };
-
- @java.lang.Override
- public com.google.protobuf.Parser<EntityDescriptorProto> getParserForType() {
- return PARSER;
- }
-
private int bitField0_;
- // optional string class_name = 1;
public static final int CLASS_NAME_FIELD_NUMBER = 1;
- private java.lang.Object className_;
+ @SuppressWarnings("serial")
+ private volatile java.lang.Object className_ = "";
/**
* <code>optional string class_name = 1;</code>
+ * @return Whether the className field is set.
*/
+ @java.lang.Override
public boolean hasClassName() {
- return ((bitField0_ & 0x00000001) == 0x00000001);
+ return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string class_name = 1;</code>
+ * @return The className.
*/
+ @java.lang.Override
public java.lang.String getClassName() {
java.lang.Object ref = className_;
if (ref instanceof java.lang.String) {
@@ -892,7 +938,9 @@ public final class LlapDaemonProtocolProtos {
}
/**
* <code>optional string class_name = 1;</code>
+ * @return The bytes for className.
*/
+ @java.lang.Override
public com.google.protobuf.ByteString
getClassNameBytes() {
java.lang.Object ref = className_;
@@ -907,103 +955,99 @@ public final class LlapDaemonProtocolProtos {
}
}
- // optional .UserPayloadProto user_payload = 2;
public static final int USER_PAYLOAD_FIELD_NUMBER = 2;
private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto userPayload_;
/**
* <code>optional .UserPayloadProto user_payload = 2;</code>
+ * @return Whether the userPayload field is set.
*/
+ @java.lang.Override
public boolean hasUserPayload() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
+ return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional .UserPayloadProto user_payload = 2;</code>
+ * @return The userPayload.
*/
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getUserPayload() {
- return userPayload_;
+ return userPayload_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance() : userPayload_;
}
/**
* <code>optional .UserPayloadProto user_payload = 2;</code>
*/
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder getUserPayloadOrBuilder() {
- return userPayload_;
+ return userPayload_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance() : userPayload_;
}
- // optional bytes history_text = 3;
public static final int HISTORY_TEXT_FIELD_NUMBER = 3;
- private com.google.protobuf.ByteString historyText_;
+ private com.google.protobuf.ByteString historyText_ = com.google.protobuf.ByteString.EMPTY;
/**
* <code>optional bytes history_text = 3;</code>
+ * @return Whether the historyText field is set.
*/
+ @java.lang.Override
public boolean hasHistoryText() {
- return ((bitField0_ & 0x00000004) == 0x00000004);
+ return ((bitField0_ & 0x00000004) != 0);
}
/**
* <code>optional bytes history_text = 3;</code>
+ * @return The historyText.
*/
+ @java.lang.Override
public com.google.protobuf.ByteString getHistoryText() {
return historyText_;
}
- private void initFields() {
- className_ = "";
- userPayload_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance();
- historyText_ = com.google.protobuf.ByteString.EMPTY;
- }
private byte memoizedIsInitialized = -1;
+ @java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
- if (isInitialized != -1) return isInitialized == 1;
+ if (isInitialized == 1) return true;
+ if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
+ @java.lang.Override
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
- getSerializedSize();
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
- output.writeBytes(1, getClassNameBytes());
+ if (((bitField0_ & 0x00000001) != 0)) {
+ com.google.protobuf.GeneratedMessageV3.writeString(output, 1, className_);
}
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
- output.writeMessage(2, userPayload_);
+ if (((bitField0_ & 0x00000002) != 0)) {
+ output.writeMessage(2, getUserPayload());
}
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
+ if (((bitField0_ & 0x00000004) != 0)) {
output.writeBytes(3, historyText_);
}
getUnknownFields().writeTo(output);
}
- private int memoizedSerializedSize = -1;
+ @java.lang.Override
public int getSerializedSize() {
- int size = memoizedSerializedSize;
+ int size = memoizedSize;
if (size != -1) return size;
size = 0;
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
- .computeBytesSize(1, getClassNameBytes());
+ if (((bitField0_ & 0x00000001) != 0)) {
+ size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, className_);
}
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ if (((bitField0_ & 0x00000002) != 0)) {
size += com.google.protobuf.CodedOutputStream
- .computeMessageSize(2, userPayload_);
+ .computeMessageSize(2, getUserPayload());
}
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
+ if (((bitField0_ & 0x00000004) != 0)) {
size += com.google.protobuf.CodedOutputStream
.computeBytesSize(3, historyText_);
}
size += getUnknownFields().getSerializedSize();
- memoizedSerializedSize = size;
+ memoizedSize = size;
return size;
}
- private static final long serialVersionUID = 0L;
- @java.lang.Override
- protected java.lang.Object writeReplace()
- throws java.io.ObjectStreamException {
- return super.writeReplace();
- }
-
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
@@ -1014,35 +1058,32 @@ public final class LlapDaemonProtocolProtos {
}
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto) obj;
- boolean result = true;
- result = result && (hasClassName() == other.hasClassName());
+ if (hasClassName() != other.hasClassName()) return false;
if (hasClassName()) {
- result = result && getClassName()
- .equals(other.getClassName());
+ if (!getClassName()
+ .equals(other.getClassName())) return false;
}
- result = result && (hasUserPayload() == other.hasUserPayload());
+ if (hasUserPayload() != other.hasUserPayload()) return false;
if (hasUserPayload()) {
- result = result && getUserPayload()
- .equals(other.getUserPayload());
+ if (!getUserPayload()
+ .equals(other.getUserPayload())) return false;
}
- result = result && (hasHistoryText() == other.hasHistoryText());
+ if (hasHistoryText() != other.hasHistoryText()) return false;
if (hasHistoryText()) {
- result = result && getHistoryText()
- .equals(other.getHistoryText());
+ if (!getHistoryText()
+ .equals(other.getHistoryText())) return false;
}
- result = result &&
- getUnknownFields().equals(other.getUnknownFields());
- return result;
+ if (!getUnknownFields().equals(other.getUnknownFields())) return false;
+ return true;
}
- private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
- hash = (19 * hash) + getDescriptorForType().hashCode();
+ hash = (19 * hash) + getDescriptor().hashCode();
if (hasClassName()) {
hash = (37 * hash) + CLASS_NAME_FIELD_NUMBER;
hash = (53 * hash) + getClassName().hashCode();
@@ -1060,6 +1101,17 @@ public final class LlapDaemonProtocolProtos {
return hash;
}
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom(
+ java.nio.ByteBuffer data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom(
+ java.nio.ByteBuffer data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
@@ -1083,46 +1135,61 @@ public final class LlapDaemonProtocolProtos {
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
+
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input);
}
+
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
- public static Builder newBuilder() { return Builder.create(); }
+ @java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder() {
+ return DEFAULT_INSTANCE.toBuilder();
+ }
public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto prototype) {
- return newBuilder().mergeFrom(prototype);
+ return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+ }
+ @java.lang.Override
+ public Builder toBuilder() {
+ return this == DEFAULT_INSTANCE
+ ? new Builder() : new Builder().mergeFrom(this);
}
- public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1130,14 +1197,16 @@ public final class LlapDaemonProtocolProtos {
* Protobuf type {@code EntityDescriptorProto}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder {
+ com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ // @@protoc_insertion_point(builder_implements:EntityDescriptorProto)
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ @java.lang.Override
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1150,47 +1219,42 @@ public final class LlapDaemonProtocolProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ if (com.google.protobuf.GeneratedMessageV3
+ .alwaysUseFieldBuilders) {
getUserPayloadFieldBuilder();
}
}
- private static Builder create() {
- return new Builder();
- }
-
+ @java.lang.Override
public Builder clear() {
super.clear();
+ bitField0_ = 0;
className_ = "";
- bitField0_ = (bitField0_ & ~0x00000001);
- if (userPayloadBuilder_ == null) {
- userPayload_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance();
- } else {
- userPayloadBuilder_.clear();
+ userPayload_ = null;
+ if (userPayloadBuilder_ != null) {
+ userPayloadBuilder_.dispose();
+ userPayloadBuilder_ = null;
}
- bitField0_ = (bitField0_ & ~0x00000002);
historyText_ = com.google.protobuf.ByteString.EMPTY;
- bitField0_ = (bitField0_ & ~0x00000004);
return this;
}
- public Builder clone() {
- return create().mergeFrom(buildPartial());
- }
-
+ @java.lang.Override
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_descriptor;
}
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getDefaultInstanceForType() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
}
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto build() {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto result = buildPartial();
if (!result.isInitialized()) {
@@ -1199,31 +1263,67 @@ public final class LlapDaemonProtocolProtos {
return result;
}
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto buildPartial() {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto(this);
+ if (bitField0_ != 0) { buildPartial0(result); }
+ onBuilt();
+ return result;
+ }
+
+ private void buildPartial0(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
- if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
+ if (((from_bitField0_ & 0x00000001) != 0)) {
+ result.className_ = className_;
to_bitField0_ |= 0x00000001;
}
- result.className_ = className_;
- if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
+ if (((from_bitField0_ & 0x00000002) != 0)) {
+ result.userPayload_ = userPayloadBuilder_ == null
+ ? userPayload_
+ : userPayloadBuilder_.build();
to_bitField0_ |= 0x00000002;
}
- if (userPayloadBuilder_ == null) {
- result.userPayload_ = userPayload_;
- } else {
- result.userPayload_ = userPayloadBuilder_.build();
- }
- if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
+ if (((from_bitField0_ & 0x00000004) != 0)) {
+ result.historyText_ = historyText_;
to_bitField0_ |= 0x00000004;
}
- result.historyText_ = historyText_;
- result.bitField0_ = to_bitField0_;
- onBuilt();
- return result;
+ result.bitField0_ |= to_bitField0_;
}
+ @java.lang.Override
+ public Builder clone() {
+ return super.clone();
+ }
+ @java.lang.Override
+ public Builder setField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ java.lang.Object value) {
+ return super.setField(field, value);
+ }
+ @java.lang.Override
+ public Builder clearField(
+ com.google.protobuf.Descriptors.FieldDescriptor field) {
+ return super.clearField(field);
+ }
+ @java.lang.Override
+ public Builder clearOneof(
+ com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ return super.clearOneof(oneof);
+ }
+ @java.lang.Override
+ public Builder setRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ int index, java.lang.Object value) {
+ return super.setRepeatedField(field, index, value);
+ }
+ @java.lang.Override
+ public Builder addRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ java.lang.Object value) {
+ return super.addRepeatedField(field, value);
+ }
+ @java.lang.Override
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto) {
return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto)other);
@@ -1236,8 +1336,8 @@ public final class LlapDaemonProtocolProtos {
public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto other) {
if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) return this;
if (other.hasClassName()) {
- bitField0_ |= 0x00000001;
className_ = other.className_;
+ bitField0_ |= 0x00000001;
onChanged();
}
if (other.hasUserPayload()) {
@@ -1247,49 +1347,86 @@ public final class LlapDaemonProtocolProtos {
setHistoryText(other.getHistoryText());
}
this.mergeUnknownFields(other.getUnknownFields());
+ onChanged();
return this;
}
+ @java.lang.Override
public final boolean isInitialized() {
return true;
}
+ @java.lang.Override
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parsedMessage = null;
+ if (extensionRegistry == null) {
+ throw new java.lang.NullPointerException();
+ }
try {
- parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ case 10: {
+ className_ = input.readBytes();
+ bitField0_ |= 0x00000001;
+ break;
+ } // case 10
+ case 18: {
+ input.readMessage(
+ getUserPayloadFieldBuilder().getBuilder(),
+ extensionRegistry);
+ bitField0_ |= 0x00000002;
+ break;
+ } // case 18
+ case 26: {
+ historyText_ = input.readBytes();
+ bitField0_ |= 0x00000004;
+ break;
+ } // case 26
+ default: {
+ if (!super.parseUnknownField(input, extensionRegistry, tag)) {
+ done = true; // was an endgroup tag
+ }
+ break;
+ } // default:
+ } // switch (tag)
+ } // while (!done)
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
- parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto) e.getUnfinishedMessage();
- throw e;
+ throw e.unwrapIOException();
} finally {
- if (parsedMessage != null) {
- mergeFrom(parsedMessage);
- }
- }
+ onChanged();
+ } // finally
return this;
}
private int bitField0_;
- // optional string class_name = 1;
private java.lang.Object className_ = "";
/**
* <code>optional string class_name = 1;</code>
+ * @return Whether the className field is set.
*/
public boolean hasClassName() {
- return ((bitField0_ & 0x00000001) == 0x00000001);
+ return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string class_name = 1;</code>
+ * @return The className.
*/
public java.lang.String getClassName() {
java.lang.Object ref = className_;
if (!(ref instanceof java.lang.String)) {
- java.lang.String s = ((com.google.protobuf.ByteString) ref)
- .toStringUtf8();
- className_ = s;
+ com.google.protobuf.ByteString bs =
+ (com.google.protobuf.ByteString) ref;
+ java.lang.String s = bs.toStringUtf8();
+ if (bs.isValidUtf8()) {
+ className_ = s;
+ }
return s;
} else {
return (java.lang.String) ref;
@@ -1297,6 +1434,7 @@ public final class LlapDaemonProtocolProtos {
}
/**
* <code>optional string class_name = 1;</code>
+ * @return The bytes for className.
*/
public com.google.protobuf.ByteString
getClassNameBytes() {
@@ -1313,56 +1451,58 @@ public final class LlapDaemonProtocolProtos {
}
/**
* <code>optional string class_name = 1;</code>
+ * @param value The className to set.
+ * @return This builder for chaining.
*/
public Builder setClassName(
java.lang.String value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000001;
+ if (value == null) { throw new NullPointerException(); }
className_ = value;
+ bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional string class_name = 1;</code>
+ * @return This builder for chaining.
*/
public Builder clearClassName() {
- bitField0_ = (bitField0_ & ~0x00000001);
className_ = getDefaultInstance().getClassName();
+ bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
return this;
}
/**
* <code>optional string class_name = 1;</code>
+ * @param value The bytes for className to set.
+ * @return This builder for chaining.
*/
public Builder setClassNameBytes(
com.google.protobuf.ByteString value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000001;
+ if (value == null) { throw new NullPointerException(); }
className_ = value;
+ bitField0_ |= 0x00000001;
onChanged();
return this;
}
- // optional .UserPayloadProto user_payload = 2;
- private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto userPayload_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance();
- private com.google.protobuf.SingleFieldBuilder<
+ private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto userPayload_;
+ private com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder> userPayloadBuilder_;
/**
* <code>optional .UserPayloadProto user_payload = 2;</code>
+ * @return Whether the userPayload field is set.
*/
public boolean hasUserPayload() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
+ return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional .UserPayloadProto user_payload = 2;</code>
+ * @return The userPayload.
*/
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getUserPayload() {
if (userPayloadBuilder_ == null) {
- return userPayload_;
+ return userPayload_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance() : userPayload_;
} else {
return userPayloadBuilder_.getMessage();
}
@@ -1376,11 +1516,11 @@ public final class LlapDaemonProtocolProtos {
throw new NullPointerException();
}
userPayload_ = value;
- onChanged();
} else {
userPayloadBuilder_.setMessage(value);
}
bitField0_ |= 0x00000002;
+ onChanged();
return this;
}
/**
@@ -1390,11 +1530,11 @@ public final class LlapDaemonProtocolProtos {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder builderForValue) {
if (userPayloadBuilder_ == null) {
userPayload_ = builderForValue.build();
- onChanged();
} else {
userPayloadBuilder_.setMessage(builderForValue.build());
}
bitField0_ |= 0x00000002;
+ onChanged();
return this;
}
/**
@@ -1402,31 +1542,33 @@ public final class LlapDaemonProtocolProtos {
*/
public Builder mergeUserPayload(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto value) {
if (userPayloadBuilder_ == null) {
- if (((bitField0_ & 0x00000002) == 0x00000002) &&
- userPayload_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance()) {
- userPayload_ =
- org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.newBuilder(userPayload_).mergeFrom(value).buildPartial();
+ if (((bitField0_ & 0x00000002) != 0) &&
+ userPayload_ != null &&
+ userPayload_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance()) {
+ getUserPayloadBuilder().mergeFrom(value);
} else {
userPayload_ = value;
}
- onChanged();
} else {
userPayloadBuilder_.mergeFrom(value);
}
- bitField0_ |= 0x00000002;
- return this;
+ if (userPayload_ != null) {
+ bitField0_ |= 0x00000002;
+ onChanged();
+ }
+ return this;
}
/**
* <code>optional .UserPayloadProto user_payload = 2;</code>
*/
public Builder clearUserPayload() {
- if (userPayloadBuilder_ == null) {
- userPayload_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance();
- onChanged();
- } else {
- userPayloadBuilder_.clear();
- }
bitField0_ = (bitField0_ & ~0x00000002);
+ userPayload_ = null;
+ if (userPayloadBuilder_ != null) {
+ userPayloadBuilder_.dispose();
+ userPayloadBuilder_ = null;
+ }
+ onChanged();
return this;
}
/**
@@ -1444,19 +1586,20 @@ public final class LlapDaemonProtocolProtos {
if (userPayloadBuilder_ != null) {
return userPayloadBuilder_.getMessageOrBuilder();
} else {
- return userPayload_;
+ return userPayload_ == null ?
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance() : userPayload_;
}
}
/**
* <code>optional .UserPayloadProto user_payload = 2;</code>
*/
- private com.google.protobuf.SingleFieldBuilder<
+ private com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder>
getUserPayloadFieldBuilder() {
if (userPayloadBuilder_ == null) {
- userPayloadBuilder_ = new com.google.protobuf.SingleFieldBuilder<
+ userPayloadBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder>(
- userPayload_,
+ getUserPayload(),
getParentForChildren(),
isClean());
userPayload_ = null;
@@ -1464,34 +1607,38 @@ public final class LlapDaemonProtocolProtos {
return userPayloadBuilder_;
}
- // optional bytes history_text = 3;
private com.google.protobuf.ByteString historyText_ = com.google.protobuf.ByteString.EMPTY;
/**
* <code>optional bytes history_text = 3;</code>
+ * @return Whether the historyText field is set.
*/
+ @java.lang.Override
public boolean hasHistoryText() {
- return ((bitField0_ & 0x00000004) == 0x00000004);
+ return ((bitField0_ & 0x00000004) != 0);
}
/**
* <code>optional bytes history_text = 3;</code>
+ * @return The historyText.
*/
+ @java.lang.Override
public com.google.protobuf.ByteString getHistoryText() {
return historyText_;
}
/**
* <code>optional bytes history_text = 3;</code>
+ * @param value The historyText to set.
+ * @return This builder for chaining.
*/
public Builder setHistoryText(com.google.protobuf.ByteString value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000004;
+ if (value == null) { throw new NullPointerException(); }
historyText_ = value;
+ bitField0_ |= 0x00000004;
onChanged();
return this;
}
/**
* <code>optional bytes history_text = 3;</code>
+ * @return This builder for chaining.
*/
public Builder clearHistoryText() {
bitField0_ = (bitField0_ & ~0x00000004);
@@ -1499,43 +1646,99 @@ public final class LlapDaemonProtocolProtos {
onChanged();
return this;
}
+ @java.lang.Override
+ public final Builder setUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.setUnknownFields(unknownFields);
+ }
+
+ @java.lang.Override
+ public final Builder mergeUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.mergeUnknownFields(unknownFields);
+ }
+
// @@protoc_insertion_point(builder_scope:EntityDescriptorProto)
}
+ // @@protoc_insertion_point(class_scope:EntityDescriptorProto)
+ private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto DEFAULT_INSTANCE;
static {
- defaultInstance = new EntityDescriptorProto(true);
- defaultInstance.initFields();
+ DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto();
+ }
+
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getDefaultInstance() {
+ return DEFAULT_INSTANCE;
+ }
+
+ @java.lang.Deprecated public static final com.google.protobuf.Parser<EntityDescriptorProto>
+ PARSER = new com.google.protobuf.AbstractParser<EntityDescriptorProto>() {
+ @java.lang.Override
+ public EntityDescriptorProto parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ Builder builder = newBuilder();
+ try {
+ builder.mergeFrom(input, extensionRegistry);
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(builder.buildPartial());
+ } catch (com.google.protobuf.UninitializedMessageException e) {
+ throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(e)
+ .setUnfinishedMessage(builder.buildPartial());
+ }
+ return builder.buildPartial();
+ }
+ };
+
+ public static com.google.protobuf.Parser<EntityDescriptorProto> parser() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<EntityDescriptorProto> getParserForType() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getDefaultInstanceForType() {
+ return DEFAULT_INSTANCE;
}
- // @@protoc_insertion_point(class_scope:EntityDescriptorProto)
}
- public interface IOSpecProtoOrBuilder
- extends com.google.protobuf.MessageOrBuilder {
+ public interface IOSpecProtoOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:IOSpecProto)
+ com.google.protobuf.MessageOrBuilder {
- // optional string connected_vertex_name = 1;
/**
* <code>optional string connected_vertex_name = 1;</code>
+ * @return Whether the connectedVertexName field is set.
*/
boolean hasConnectedVertexName();
/**
* <code>optional string connected_vertex_name = 1;</code>
+ * @return The connectedVertexName.
*/
java.lang.String getConnectedVertexName();
/**
* <code>optional string connected_vertex_name = 1;</code>
+ * @return The bytes for connectedVertexName.
*/
com.google.protobuf.ByteString
getConnectedVertexNameBytes();
- // optional .EntityDescriptorProto io_descriptor = 2;
/**
* <code>optional .EntityDescriptorProto io_descriptor = 2;</code>
+ * @return Whether the ioDescriptor field is set.
*/
boolean hasIoDescriptor();
/**
* <code>optional .EntityDescriptorProto io_descriptor = 2;</code>
+ * @return The ioDescriptor.
*/
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getIoDescriptor();
/**
@@ -1543,13 +1746,14 @@ public final class LlapDaemonProtocolProtos {
*/
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getIoDescriptorOrBuilder();
- // optional int32 physical_edge_count = 3;
/**
* <code>optional int32 physical_edge_count = 3;</code>
+ * @return Whether the physicalEdgeCount field is set.
*/
boolean hasPhysicalEdgeCount();
/**
* <code>optional int32 physical_edge_count = 3;</code>
+ * @return The physicalEdgeCount.
*/
int getPhysicalEdgeCount();
}
@@ -1557,128 +1761,55 @@ public final class LlapDaemonProtocolProtos {
* Protobuf type {@code IOSpecProto}
*/
public static final class IOSpecProto extends
- com.google.protobuf.GeneratedMessage
- implements IOSpecProtoOrBuilder {
+ com.google.protobuf.GeneratedMessageV3 implements
+ // @@protoc_insertion_point(message_implements:IOSpecProto)
+ IOSpecProtoOrBuilder {
+ private static final long serialVersionUID = 0L;
// Use IOSpecProto.newBuilder() to construct.
- private IOSpecProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ private IOSpecProto(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
- this.unknownFields = builder.getUnknownFields();
- }
- private IOSpecProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
- private static final IOSpecProto defaultInstance;
- public static IOSpecProto getDefaultInstance() {
- return defaultInstance;
}
-
- public IOSpecProto getDefaultInstanceForType() {
- return defaultInstance;
+ private IOSpecProto() {
+ connectedVertexName_ = "";
}
- private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
- return this.unknownFields;
- }
- private IOSpecProto(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- initFields();
- int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
- try {
- boolean done = false;
- while (!done) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
- done = true;
- break;
- default: {
- if (!parseUnknownField(input, unknownFields,
- extensionRegistry, tag)) {
- done = true;
- }
- break;
- }
- case 10: {
- bitField0_ |= 0x00000001;
- connectedVertexName_ = input.readBytes();
- break;
- }
- case 18: {
- org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder subBuilder = null;
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
- subBuilder = ioDescriptor_.toBuilder();
- }
- ioDescriptor_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.PARSER, extensionRegistry);
- if (subBuilder != null) {
- subBuilder.mergeFrom(ioDescriptor_);
- ioDescriptor_ = subBuilder.buildPartial();
- }
- bitField0_ |= 0x00000002;
- break;
- }
- case 24: {
- bitField0_ |= 0x00000004;
- physicalEdgeCount_ = input.readInt32();
- break;
- }
- }
- }
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
- throw e.setUnfinishedMessage(this);
- } catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
- e.getMessage()).setUnfinishedMessage(this);
- } finally {
- this.unknownFields = unknownFields.build();
- makeExtensionsImmutable();
- }
+ @SuppressWarnings({"unused"})
+ protected java.lang.Object newInstance(
+ UnusedPrivateParameter unused) {
+ return new IOSpecProto();
}
+
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ @java.lang.Override
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder.class);
}
- public static com.google.protobuf.Parser<IOSpecProto> PARSER =
- new com.google.protobuf.AbstractParser<IOSpecProto>() {
- public IOSpecProto parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return new IOSpecProto(input, extensionRegistry);
- }
- };
-
- @java.lang.Override
- public com.google.protobuf.Parser<IOSpecProto> getParserForType() {
- return PARSER;
- }
-
private int bitField0_;
- // optional string connected_vertex_name = 1;
public static final int CONNECTED_VERTEX_NAME_FIELD_NUMBER = 1;
- private java.lang.Object connectedVertexName_;
+ @SuppressWarnings("serial")
+ private volatile java.lang.Object connectedVertexName_ = "";
/**
* <code>optional string connected_vertex_name = 1;</code>
+ * @return Whether the connectedVertexName field is set.
*/
+ @java.lang.Override
public boolean hasConnectedVertexName() {
- return ((bitField0_ & 0x00000001) == 0x00000001);
+ return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string connected_vertex_name = 1;</code>
+ * @return The connectedVertexName.
*/
+ @java.lang.Override
public java.lang.String getConnectedVertexName() {
java.lang.Object ref = connectedVertexName_;
if (ref instanceof java.lang.String) {
@@ -1695,7 +1826,9 @@ public final class LlapDaemonProtocolProtos {
}
/**
* <code>optional string connected_vertex_name = 1;</code>
+ * @return The bytes for connectedVertexName.
*/
+ @java.lang.Override
public com.google.protobuf.ByteString
getConnectedVertexNameBytes() {
java.lang.Object ref = connectedVertexName_;
@@ -1710,103 +1843,99 @@ public final class LlapDaemonProtocolProtos {
}
}
- // optional .EntityDescriptorProto io_descriptor = 2;
public static final int IO_DESCRIPTOR_FIELD_NUMBER = 2;
private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto ioDescriptor_;
/**
* <code>optional .EntityDescriptorProto io_descriptor = 2;</code>
+ * @return Whether the ioDescriptor field is set.
*/
+ @java.lang.Override
public boolean hasIoDescriptor() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
+ return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional .EntityDescriptorProto io_descriptor = 2;</code>
+ * @return The ioDescriptor.
*/
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getIoDescriptor() {
- return ioDescriptor_;
+ return ioDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : ioDescriptor_;
}
/**
* <code>optional .EntityDescriptorProto io_descriptor = 2;</code>
*/
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getIoDescriptorOrBuilder() {
- return ioDescriptor_;
+ return ioDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : ioDescriptor_;
}
- // optional int32 physical_edge_count = 3;
public static final int PHYSICAL_EDGE_COUNT_FIELD_NUMBER = 3;
- private int physicalEdgeCount_;
+ private int physicalEdgeCount_ = 0;
/**
* <code>optional int32 physical_edge_count = 3;</code>
+ * @return Whether the physicalEdgeCount field is set.
*/
+ @java.lang.Override
public boolean hasPhysicalEdgeCount() {
- return ((bitField0_ & 0x00000004) == 0x00000004);
+ return ((bitField0_ & 0x00000004) != 0);
}
/**
* <code>optional int32 physical_edge_count = 3;</code>
+ * @return The physicalEdgeCount.
*/
+ @java.lang.Override
public int getPhysicalEdgeCount() {
return physicalEdgeCount_;
}
- private void initFields() {
- connectedVertexName_ = "";
- ioDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
- physicalEdgeCount_ = 0;
- }
private byte memoizedIsInitialized = -1;
+ @java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
- if (isInitialized != -1) return isInitialized == 1;
+ if (isInitialized == 1) return true;
+ if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
+ @java.lang.Override
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
- getSerializedSize();
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
- output.writeBytes(1, getConnectedVertexNameBytes());
+ if (((bitField0_ & 0x00000001) != 0)) {
+ com.google.protobuf.GeneratedMessageV3.writeString(output, 1, connectedVertexName_);
}
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
- output.writeMessage(2, ioDescriptor_);
+ if (((bitField0_ & 0x00000002) != 0)) {
+ output.writeMessage(2, getIoDescriptor());
}
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
+ if (((bitField0_ & 0x00000004) != 0)) {
output.writeInt32(3, physicalEdgeCount_);
}
getUnknownFields().writeTo(output);
}
- private int memoizedSerializedSize = -1;
+ @java.lang.Override
public int getSerializedSize() {
- int size = memoizedSerializedSize;
+ int size = memoizedSize;
if (size != -1) return size;
size = 0;
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
- .computeBytesSize(1, getConnectedVertexNameBytes());
+ if (((bitField0_ & 0x00000001) != 0)) {
+ size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, connectedVertexName_);
}
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ if (((bitField0_ & 0x00000002) != 0)) {
size += com.google.protobuf.CodedOutputStream
- .computeMessageSize(2, ioDescriptor_);
+ .computeMessageSize(2, getIoDescriptor());
}
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
+ if (((bitField0_ & 0x00000004) != 0)) {
size += com.google.protobuf.CodedOutputStream
.computeInt32Size(3, physicalEdgeCount_);
}
size += getUnknownFields().getSerializedSize();
- memoizedSerializedSize = size;
+ memoizedSize = size;
return size;
}
- private static final long serialVersionUID = 0L;
- @java.lang.Override
- protected java.lang.Object writeReplace()
- throws java.io.ObjectStreamException {
- return super.writeReplace();
- }
-
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
@@ -1817,35 +1946,32 @@ public final class LlapDaemonProtocolProtos {
}
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto) obj;
- boolean result = true;
- result = result && (hasConnectedVertexName() == other.hasConnectedVertexName());
+ if (hasConnectedVertexName() != other.hasConnectedVertexName()) return false;
if (hasConnectedVertexName()) {
- result = result && getConnectedVertexName()
- .equals(other.getConnectedVertexName());
+ if (!getConnectedVertexName()
+ .equals(other.getConnectedVertexName())) return false;
}
- result = result && (hasIoDescriptor() == other.hasIoDescriptor());
+ if (hasIoDescriptor() != other.hasIoDescriptor()) return false;
if (hasIoDescriptor()) {
- result = result && getIoDescriptor()
- .equals(other.getIoDescriptor());
+ if (!getIoDescriptor()
+ .equals(other.getIoDescriptor())) return false;
}
- result = result && (hasPhysicalEdgeCount() == other.hasPhysicalEdgeCount());
+ if (hasPhysicalEdgeCount() != other.hasPhysicalEdgeCount()) return false;
if (hasPhysicalEdgeCount()) {
- result = result && (getPhysicalEdgeCount()
- == other.getPhysicalEdgeCount());
+ if (getPhysicalEdgeCount()
+ != other.getPhysicalEdgeCount()) return false;
}
- result = result &&
- getUnknownFields().equals(other.getUnknownFields());
- return result;
+ if (!getUnknownFields().equals(other.getUnknownFields())) return false;
+ return true;
}
- private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
- hash = (19 * hash) + getDescriptorForType().hashCode();
+ hash = (19 * hash) + getDescriptor().hashCode();
if (hasConnectedVertexName()) {
hash = (37 * hash) + CONNECTED_VERTEX_NAME_FIELD_NUMBER;
hash = (53 * hash) + getConnectedVertexName().hashCode();
@@ -1863,6 +1989,17 @@ public final class LlapDaemonProtocolProtos {
return hash;
}
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom(
+ java.nio.ByteBuffer data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom(
+ java.nio.ByteBuffer data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
@@ -1886,46 +2023,61 @@ public final class LlapDaemonProtocolProtos {
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
+
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input);
}
+
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
- public static Builder newBuilder() { return Builder.create(); }
+ @java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder() {
+ return DEFAULT_INSTANCE.toBuilder();
+ }
public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto prototype) {
- return newBuilder().mergeFrom(prototype);
+ return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+ }
+ @java.lang.Override
+ public Builder toBuilder() {
+ return this == DEFAULT_INSTANCE
+ ? new Builder() : new Builder().mergeFrom(this);
}
- public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -1933,14 +2085,16 @@ public final class LlapDaemonProtocolProtos {
* Protobuf type {@code IOSpecProto}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder {
+ com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ // @@protoc_insertion_point(builder_implements:IOSpecProto)
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ @java.lang.Override
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -1953,47 +2107,42 @@ public final class LlapDaemonProtocolProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ if (com.google.protobuf.GeneratedMessageV3
+ .alwaysUseFieldBuilders) {
getIoDescriptorFieldBuilder();
}
}
- private static Builder create() {
- return new Builder();
- }
-
+ @java.lang.Override
public Builder clear() {
super.clear();
+ bitField0_ = 0;
connectedVertexName_ = "";
- bitField0_ = (bitField0_ & ~0x00000001);
- if (ioDescriptorBuilder_ == null) {
- ioDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
- } else {
- ioDescriptorBuilder_.clear();
+ ioDescriptor_ = null;
+ if (ioDescriptorBuilder_ != null) {
+ ioDescriptorBuilder_.dispose();
+ ioDescriptorBuilder_ = null;
}
- bitField0_ = (bitField0_ & ~0x00000002);
physicalEdgeCount_ = 0;
- bitField0_ = (bitField0_ & ~0x00000004);
return this;
}
- public Builder clone() {
- return create().mergeFrom(buildPartial());
- }
-
+ @java.lang.Override
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_descriptor;
}
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getDefaultInstanceForType() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.getDefaultInstance();
}
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto build() {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto result = buildPartial();
if (!result.isInitialized()) {
@@ -2002,31 +2151,67 @@ public final class LlapDaemonProtocolProtos {
return result;
}
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto buildPartial() {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto(this);
+ if (bitField0_ != 0) { buildPartial0(result); }
+ onBuilt();
+ return result;
+ }
+
+ private void buildPartial0(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
- if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
+ if (((from_bitField0_ & 0x00000001) != 0)) {
+ result.connectedVertexName_ = connectedVertexName_;
to_bitField0_ |= 0x00000001;
}
- result.connectedVertexName_ = connectedVertexName_;
- if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
+ if (((from_bitField0_ & 0x00000002) != 0)) {
+ result.ioDescriptor_ = ioDescriptorBuilder_ == null
+ ? ioDescriptor_
+ : ioDescriptorBuilder_.build();
to_bitField0_ |= 0x00000002;
}
- if (ioDescriptorBuilder_ == null) {
- result.ioDescriptor_ = ioDescriptor_;
- } else {
- result.ioDescriptor_ = ioDescriptorBuilder_.build();
- }
- if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
+ if (((from_bitField0_ & 0x00000004) != 0)) {
+ result.physicalEdgeCount_ = physicalEdgeCount_;
to_bitField0_ |= 0x00000004;
}
- result.physicalEdgeCount_ = physicalEdgeCount_;
- result.bitField0_ = to_bitField0_;
- onBuilt();
- return result;
+ result.bitField0_ |= to_bitField0_;
}
+ @java.lang.Override
+ public Builder clone() {
+ return super.clone();
+ }
+ @java.lang.Override
+ public Builder setField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ java.lang.Object value) {
+ return super.setField(field, value);
+ }
+ @java.lang.Override
+ public Builder clearField(
+ com.google.protobuf.Descriptors.FieldDescriptor field) {
+ return super.clearField(field);
+ }
+ @java.lang.Override
+ public Builder clearOneof(
+ com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ return super.clearOneof(oneof);
+ }
+ @java.lang.Override
+ public Builder setRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ int index, java.lang.Object value) {
+ return super.setRepeatedField(field, index, value);
+ }
+ @java.lang.Override
+ public Builder addRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ java.lang.Object value) {
+ return super.addRepeatedField(field, value);
+ }
+ @java.lang.Override
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto) {
return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto)other);
@@ -2039,8 +2224,8 @@ public final class LlapDaemonProtocolProtos {
public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto other) {
if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.getDefaultInstance()) return this;
if (other.hasConnectedVertexName()) {
- bitField0_ |= 0x00000001;
connectedVertexName_ = other.connectedVertexName_;
+ bitField0_ |= 0x00000001;
onChanged();
}
if (other.hasIoDescriptor()) {
@@ -2050,49 +2235,86 @@ public final class LlapDaemonProtocolProtos {
setPhysicalEdgeCount(other.getPhysicalEdgeCount());
}
this.mergeUnknownFields(other.getUnknownFields());
+ onChanged();
return this;
}
+ @java.lang.Override
public final boolean isInitialized() {
return true;
}
+ @java.lang.Override
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parsedMessage = null;
+ if (extensionRegistry == null) {
+ throw new java.lang.NullPointerException();
+ }
try {
- parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ case 10: {
+ connectedVertexName_ = input.readBytes();
+ bitField0_ |= 0x00000001;
+ break;
+ } // case 10
+ case 18: {
+ input.readMessage(
+ getIoDescriptorFieldBuilder().getBuilder(),
+ extensionRegistry);
+ bitField0_ |= 0x00000002;
+ break;
+ } // case 18
+ case 24: {
+ physicalEdgeCount_ = input.readInt32();
+ bitField0_ |= 0x00000004;
+ break;
+ } // case 24
+ default: {
+ if (!super.parseUnknownField(input, extensionRegistry, tag)) {
+ done = true; // was an endgroup tag
+ }
+ break;
+ } // default:
+ } // switch (tag)
+ } // while (!done)
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
- parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto) e.getUnfinishedMessage();
- throw e;
+ throw e.unwrapIOException();
} finally {
- if (parsedMessage != null) {
- mergeFrom(parsedMessage);
- }
- }
+ onChanged();
+ } // finally
return this;
}
private int bitField0_;
- // optional string connected_vertex_name = 1;
private java.lang.Object connectedVertexName_ = "";
/**
* <code>optional string connected_vertex_name = 1;</code>
+ * @return Whether the connectedVertexName field is set.
*/
public boolean hasConnectedVertexName() {
- return ((bitField0_ & 0x00000001) == 0x00000001);
+ return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string connected_vertex_name = 1;</code>
+ * @return The connectedVertexName.
*/
public java.lang.String getConnectedVertexName() {
java.lang.Object ref = connectedVertexName_;
if (!(ref instanceof java.lang.String)) {
- java.lang.String s = ((com.google.protobuf.ByteString) ref)
- .toStringUtf8();
- connectedVertexName_ = s;
+ com.google.protobuf.ByteString bs =
+ (com.google.protobuf.ByteString) ref;
+ java.lang.String s = bs.toStringUtf8();
+ if (bs.isValidUtf8()) {
+ connectedVertexName_ = s;
+ }
return s;
} else {
return (java.lang.String) ref;
@@ -2100,6 +2322,7 @@ public final class LlapDaemonProtocolProtos {
}
/**
* <code>optional string connected_vertex_name = 1;</code>
+ * @return The bytes for connectedVertexName.
*/
public com.google.protobuf.ByteString
getConnectedVertexNameBytes() {
@@ -2116,56 +2339,58 @@ public final class LlapDaemonProtocolProtos {
}
/**
* <code>optional string connected_vertex_name = 1;</code>
+ * @param value The connectedVertexName to set.
+ * @return This builder for chaining.
*/
public Builder setConnectedVertexName(
java.lang.String value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000001;
+ if (value == null) { throw new NullPointerException(); }
connectedVertexName_ = value;
+ bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional string connected_vertex_name = 1;</code>
+ * @return This builder for chaining.
*/
public Builder clearConnectedVertexName() {
- bitField0_ = (bitField0_ & ~0x00000001);
connectedVertexName_ = getDefaultInstance().getConnectedVertexName();
+ bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
return this;
}
/**
* <code>optional string connected_vertex_name = 1;</code>
+ * @param value The bytes for connectedVertexName to set.
+ * @return This builder for chaining.
*/
public Builder setConnectedVertexNameBytes(
com.google.protobuf.ByteString value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000001;
+ if (value == null) { throw new NullPointerException(); }
connectedVertexName_ = value;
+ bitField0_ |= 0x00000001;
onChanged();
return this;
}
- // optional .EntityDescriptorProto io_descriptor = 2;
- private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto ioDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
- private com.google.protobuf.SingleFieldBuilder<
+ private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto ioDescriptor_;
+ private com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> ioDescriptorBuilder_;
/**
* <code>optional .EntityDescriptorProto io_descriptor = 2;</code>
+ * @return Whether the ioDescriptor field is set.
*/
public boolean hasIoDescriptor() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
+ return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional .EntityDescriptorProto io_descriptor = 2;</code>
+ * @return The ioDescriptor.
*/
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getIoDescriptor() {
if (ioDescriptorBuilder_ == null) {
- return ioDescriptor_;
+ return ioDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : ioDescriptor_;
} else {
return ioDescriptorBuilder_.getMessage();
}
@@ -2179,11 +2404,11 @@ public final class LlapDaemonProtocolProtos {
throw new NullPointerException();
}
ioDescriptor_ = value;
- onChanged();
} else {
ioDescriptorBuilder_.setMessage(value);
}
bitField0_ |= 0x00000002;
+ onChanged();
return this;
}
/**
@@ -2193,11 +2418,11 @@ public final class LlapDaemonProtocolProtos {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder builderForValue) {
if (ioDescriptorBuilder_ == null) {
ioDescriptor_ = builderForValue.build();
- onChanged();
} else {
ioDescriptorBuilder_.setMessage(builderForValue.build());
}
bitField0_ |= 0x00000002;
+ onChanged();
return this;
}
/**
@@ -2205,31 +2430,33 @@ public final class LlapDaemonProtocolProtos {
*/
public Builder mergeIoDescriptor(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto value) {
if (ioDescriptorBuilder_ == null) {
- if (((bitField0_ & 0x00000002) == 0x00000002) &&
- ioDescriptor_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) {
- ioDescriptor_ =
- org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.newBuilder(ioDescriptor_).mergeFrom(value).buildPartial();
+ if (((bitField0_ & 0x00000002) != 0) &&
+ ioDescriptor_ != null &&
+ ioDescriptor_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) {
+ getIoDescriptorBuilder().mergeFrom(value);
} else {
ioDescriptor_ = value;
}
- onChanged();
} else {
ioDescriptorBuilder_.mergeFrom(value);
}
- bitField0_ |= 0x00000002;
+ if (ioDescriptor_ != null) {
+ bitField0_ |= 0x00000002;
+ onChanged();
+ }
return this;
}
/**
* <code>optional .EntityDescriptorProto io_descriptor = 2;</code>
*/
public Builder clearIoDescriptor() {
- if (ioDescriptorBuilder_ == null) {
- ioDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
- onChanged();
- } else {
- ioDescriptorBuilder_.clear();
- }
bitField0_ = (bitField0_ & ~0x00000002);
+ ioDescriptor_ = null;
+ if (ioDescriptorBuilder_ != null) {
+ ioDescriptorBuilder_.dispose();
+ ioDescriptorBuilder_ = null;
+ }
+ onChanged();
return this;
}
/**
@@ -2247,19 +2474,20 @@ public final class LlapDaemonProtocolProtos {
if (ioDescriptorBuilder_ != null) {
return ioDescriptorBuilder_.getMessageOrBuilder();
} else {
- return ioDescriptor_;
+ return ioDescriptor_ == null ?
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : ioDescriptor_;
}
}
/**
* <code>optional .EntityDescriptorProto io_descriptor = 2;</code>
*/
- private com.google.protobuf.SingleFieldBuilder<
+ private com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder>
getIoDescriptorFieldBuilder() {
if (ioDescriptorBuilder_ == null) {
- ioDescriptorBuilder_ = new com.google.protobuf.SingleFieldBuilder<
+ ioDescriptorBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder>(
- ioDescriptor_,
+ getIoDescriptor(),
getParentForChildren(),
isClean());
ioDescriptor_ = null;
@@ -2267,31 +2495,38 @@ public final class LlapDaemonProtocolProtos {
return ioDescriptorBuilder_;
}
- // optional int32 physical_edge_count = 3;
private int physicalEdgeCount_ ;
/**
* <code>optional int32 physical_edge_count = 3;</code>
+ * @return Whether the physicalEdgeCount field is set.
*/
+ @java.lang.Override
public boolean hasPhysicalEdgeCount() {
- return ((bitField0_ & 0x00000004) == 0x00000004);
+ return ((bitField0_ & 0x00000004) != 0);
}
/**
* <code>optional int32 physical_edge_count = 3;</code>
+ * @return The physicalEdgeCount.
*/
+ @java.lang.Override
public int getPhysicalEdgeCount() {
return physicalEdgeCount_;
}
/**
* <code>optional int32 physical_edge_count = 3;</code>
+ * @param value The physicalEdgeCount to set.
+ * @return This builder for chaining.
*/
public Builder setPhysicalEdgeCount(int value) {
- bitField0_ |= 0x00000004;
+
physicalEdgeCount_ = value;
+ bitField0_ |= 0x00000004;
onChanged();
return this;
}
/**
* <code>optional int32 physical_edge_count = 3;</code>
+ * @return This builder for chaining.
*/
public Builder clearPhysicalEdgeCount() {
bitField0_ = (bitField0_ & ~0x00000004);
@@ -2299,63 +2534,124 @@ public final class LlapDaemonProtocolProtos {
onChanged();
return this;
}
+ @java.lang.Override
+ public final Builder setUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.setUnknownFields(unknownFields);
+ }
+
+ @java.lang.Override
+ public final Builder mergeUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.mergeUnknownFields(unknownFields);
+ }
+
// @@protoc_insertion_point(builder_scope:IOSpecProto)
}
+ // @@protoc_insertion_point(class_scope:IOSpecProto)
+ private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto DEFAULT_INSTANCE;
static {
- defaultInstance = new IOSpecProto(true);
- defaultInstance.initFields();
+ DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto();
+ }
+
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getDefaultInstance() {
+ return DEFAULT_INSTANCE;
+ }
+
+ @java.lang.Deprecated public static final com.google.protobuf.Parser<IOSpecProto>
+ PARSER = new com.google.protobuf.AbstractParser<IOSpecProto>() {
+ @java.lang.Override
+ public IOSpecProto parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ Builder builder = newBuilder();
+ try {
+ builder.mergeFrom(input, extensionRegistry);
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(builder.buildPartial());
+ } catch (com.google.protobuf.UninitializedMessageException e) {
+ throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(e)
+ .setUnfinishedMessage(builder.buildPartial());
+ }
+ return builder.buildPartial();
+ }
+ };
+
+ public static com.google.protobuf.Parser<IOSpecProto> parser() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<IOSpecProto> getParserForType() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getDefaultInstanceForType() {
+ return DEFAULT_INSTANCE;
}
- // @@protoc_insertion_point(class_scope:IOSpecProto)
}
- public interface GroupInputSpecProtoOrBuilder
- extends com.google.protobuf.MessageOrBuilder {
+ public interface GroupInputSpecProtoOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:GroupInputSpecProto)
+ com.google.protobuf.MessageOrBuilder {
- // optional string group_name = 1;
/**
* <code>optional string group_name = 1;</code>
+ * @return Whether the groupName field is set.
*/
boolean hasGroupName();
/**
* <code>optional string group_name = 1;</code>
+ * @return The groupName.
*/
java.lang.String getGroupName();
/**
* <code>optional string group_name = 1;</code>
+ * @return The bytes for groupName.
*/
com.google.protobuf.ByteString
getGroupNameBytes();
- // repeated string group_vertices = 2;
/**
* <code>repeated string group_vertices = 2;</code>
+ * @return A list containing the groupVertices.
*/
java.util.List<java.lang.String>
- getGroupVerticesList();
+ getGroupVerticesList();
/**
* <code>repeated string group_vertices = 2;</code>
+ * @return The count of groupVertices.
*/
int getGroupVerticesCount();
/**
* <code>repeated string group_vertices = 2;</code>
+ * @param index The index of the element to return.
+ * @return The groupVertices at the given index.
*/
java.lang.String getGroupVertices(int index);
/**
* <code>repeated string group_vertices = 2;</code>
+ * @param index The index of the value to return.
+ * @return The bytes of the groupVertices at the given index.
*/
com.google.protobuf.ByteString
getGroupVerticesBytes(int index);
- // optional .EntityDescriptorProto merged_input_descriptor = 3;
/**
* <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code>
+ * @return Whether the mergedInputDescriptor field is set.
*/
boolean hasMergedInputDescriptor();
/**
* <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code>
+ * @return The mergedInputDescriptor.
*/
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getMergedInputDescriptor();
/**
@@ -2367,134 +2663,57 @@ public final class LlapDaemonProtocolProtos {
* Protobuf type {@code GroupInputSpecProto}
*/
public static final class GroupInputSpecProto extends
- com.google.protobuf.GeneratedMessage
- implements GroupInputSpecProtoOrBuilder {
+ com.google.protobuf.GeneratedMessageV3 implements
+ // @@protoc_insertion_point(message_implements:GroupInputSpecProto)
+ GroupInputSpecProtoOrBuilder {
+ private static final long serialVersionUID = 0L;
// Use GroupInputSpecProto.newBuilder() to construct.
- private GroupInputSpecProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ private GroupInputSpecProto(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
- this.unknownFields = builder.getUnknownFields();
}
- private GroupInputSpecProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
- private static final GroupInputSpecProto defaultInstance;
- public static GroupInputSpecProto getDefaultInstance() {
- return defaultInstance;
- }
-
- public GroupInputSpecProto getDefaultInstanceForType() {
- return defaultInstance;
+ private GroupInputSpecProto() {
+ groupName_ = "";
+ groupVertices_ =
+ com.google.protobuf.LazyStringArrayList.emptyList();
}
- private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
- return this.unknownFields;
- }
- private GroupInputSpecProto(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- initFields();
- int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
- try {
- boolean done = false;
- while (!done) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
- done = true;
- break;
- default: {
- if (!parseUnknownField(input, unknownFields,
- extensionRegistry, tag)) {
- done = true;
- }
- break;
- }
- case 10: {
- bitField0_ |= 0x00000001;
- groupName_ = input.readBytes();
- break;
- }
- case 18: {
- if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
- groupVertices_ = new com.google.protobuf.LazyStringArrayList();
- mutable_bitField0_ |= 0x00000002;
- }
- groupVertices_.add(input.readBytes());
- break;
- }
- case 26: {
- org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder subBuilder = null;
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
- subBuilder = mergedInputDescriptor_.toBuilder();
- }
- mergedInputDescriptor_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.PARSER, extensionRegistry);
- if (subBuilder != null) {
- subBuilder.mergeFrom(mergedInputDescriptor_);
- mergedInputDescriptor_ = subBuilder.buildPartial();
- }
- bitField0_ |= 0x00000002;
- break;
- }
- }
- }
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
- throw e.setUnfinishedMessage(this);
- } catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
- e.getMessage()).setUnfinishedMessage(this);
- } finally {
- if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
- groupVertices_ = new com.google.protobuf.UnmodifiableLazyStringList(groupVertices_);
- }
- this.unknownFields = unknownFields.build();
- makeExtensionsImmutable();
- }
+ @SuppressWarnings({"unused"})
+ protected java.lang.Object newInstance(
+ UnusedPrivateParameter unused) {
+ return new GroupInputSpecProto();
}
+
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GroupInputSpecProto_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ @java.lang.Override
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GroupInputSpecProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder.class);
}
- public static com.google.protobuf.Parser<GroupInputSpecProto> PARSER =
- new com.google.protobuf.AbstractParser<GroupInputSpecProto>() {
- public GroupInputSpecProto parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return new GroupInputSpecProto(input, extensionRegistry);
- }
- };
-
- @java.lang.Override
- public com.google.protobuf.Parser<GroupInputSpecProto> getParserForType() {
- return PARSER;
- }
-
private int bitField0_;
- // optional string group_name = 1;
public static final int GROUP_NAME_FIELD_NUMBER = 1;
- private java.lang.Object groupName_;
+ @SuppressWarnings("serial")
+ private volatile java.lang.Object groupName_ = "";
/**
* <code>optional string group_name = 1;</code>
+ * @return Whether the groupName field is set.
*/
+ @java.lang.Override
public boolean hasGroupName() {
- return ((bitField0_ & 0x00000001) == 0x00000001);
+ return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string group_name = 1;</code>
+ * @return The groupName.
*/
+ @java.lang.Override
public java.lang.String getGroupName() {
java.lang.Object ref = groupName_;
if (ref instanceof java.lang.String) {
@@ -2511,7 +2730,9 @@ public final class LlapDaemonProtocolProtos {
}
/**
* <code>optional string group_name = 1;</code>
+ * @return The bytes for groupName.
*/
+ @java.lang.Override
public com.google.protobuf.ByteString
getGroupNameBytes() {
java.lang.Object ref = groupName_;
@@ -2526,122 +2747,121 @@ public final class LlapDaemonProtocolProtos {
}
}
- // repeated string group_vertices = 2;
public static final int GROUP_VERTICES_FIELD_NUMBER = 2;
- private com.google.protobuf.LazyStringList groupVertices_;
+ @SuppressWarnings("serial")
+ private com.google.protobuf.LazyStringArrayList groupVertices_ =
+ com.google.protobuf.LazyStringArrayList.emptyList();
/**
* <code>repeated string group_vertices = 2;</code>
+ * @return A list containing the groupVertices.
*/
- public java.util.List<java.lang.String>
+ public com.google.protobuf.ProtocolStringList
getGroupVerticesList() {
return groupVertices_;
}
/**
* <code>repeated string group_vertices = 2;</code>
+ * @return The count of groupVertices.
*/
public int getGroupVerticesCount() {
return groupVertices_.size();
}
/**
* <code>repeated string group_vertices = 2;</code>
+ * @param index The index of the element to return.
+ * @return The groupVertices at the given index.
*/
public java.lang.String getGroupVertices(int index) {
return groupVertices_.get(index);
}
/**
* <code>repeated string group_vertices = 2;</code>
+ * @param index The index of the value to return.
+ * @return The bytes of the groupVertices at the given index.
*/
public com.google.protobuf.ByteString
getGroupVerticesBytes(int index) {
return groupVertices_.getByteString(index);
}
- // optional .EntityDescriptorProto merged_input_descriptor = 3;
public static final int MERGED_INPUT_DESCRIPTOR_FIELD_NUMBER = 3;
private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto mergedInputDescriptor_;
/**
* <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code>
+ * @return Whether the mergedInputDescriptor field is set.
*/
+ @java.lang.Override
public boolean hasMergedInputDescriptor() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
+ return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code>
+ * @return The mergedInputDescriptor.
*/
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getMergedInputDescriptor() {
- return mergedInputDescriptor_;
+ return mergedInputDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : mergedInputDescriptor_;
}
/**
* <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code>
*/
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getMergedInputDescriptorOrBuilder() {
- return mergedInputDescriptor_;
+ return mergedInputDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : mergedInputDescriptor_;
}
- private void initFields() {
- groupName_ = "";
- groupVertices_ = com.google.protobuf.LazyStringArrayList.EMPTY;
- mergedInputDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
- }
private byte memoizedIsInitialized = -1;
+ @java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
- if (isInitialized != -1) return isInitialized == 1;
+ if (isInitialized == 1) return true;
+ if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
+ @java.lang.Override
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
- getSerializedSize();
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
- output.writeBytes(1, getGroupNameBytes());
+ if (((bitField0_ & 0x00000001) != 0)) {
+ com.google.protobuf.GeneratedMessageV3.writeString(output, 1, groupName_);
}
for (int i = 0; i < groupVertices_.size(); i++) {
- output.writeBytes(2, groupVertices_.getByteString(i));
+ com.google.protobuf.GeneratedMessageV3.writeString(output, 2, groupVertices_.getRaw(i));
}
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
- output.writeMessage(3, mergedInputDescriptor_);
+ if (((bitField0_ & 0x00000002) != 0)) {
+ output.writeMessage(3, getMergedInputDescriptor());
}
getUnknownFields().writeTo(output);
}
- private int memoizedSerializedSize = -1;
+ @java.lang.Override
public int getSerializedSize() {
- int size = memoizedSerializedSize;
+ int size = memoizedSize;
if (size != -1) return size;
size = 0;
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
- .computeBytesSize(1, getGroupNameBytes());
+ if (((bitField0_ & 0x00000001) != 0)) {
+ size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, groupName_);
}
{
int dataSize = 0;
for (int i = 0; i < groupVertices_.size(); i++) {
- dataSize += com.google.protobuf.CodedOutputStream
- .computeBytesSizeNoTag(groupVertices_.getByteString(i));
+ dataSize += computeStringSizeNoTag(groupVertices_.getRaw(i));
}
size += dataSize;
size += 1 * getGroupVerticesList().size();
}
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ if (((bitField0_ & 0x00000002) != 0)) {
size += com.google.protobuf.CodedOutputStream
- .computeMessageSize(3, mergedInputDescriptor_);
+ .computeMessageSize(3, getMergedInputDescriptor());
}
size += getUnknownFields().getSerializedSize();
- memoizedSerializedSize = size;
+ memoizedSize = size;
return size;
}
- private static final long serialVersionUID = 0L;
- @java.lang.Override
- protected java.lang.Object writeReplace()
- throws java.io.ObjectStreamException {
- return super.writeReplace();
- }
-
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
@@ -2652,32 +2872,29 @@ public final class LlapDaemonProtocolProtos {
}
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto) obj;
- boolean result = true;
- result = result && (hasGroupName() == other.hasGroupName());
+ if (hasGroupName() != other.hasGroupName()) return false;
if (hasGroupName()) {
- result = result && getGroupName()
- .equals(other.getGroupName());
+ if (!getGroupName()
+ .equals(other.getGroupName())) return false;
}
- result = result && getGroupVerticesList()
- .equals(other.getGroupVerticesList());
- result = result && (hasMergedInputDescriptor() == other.hasMergedInputDescriptor());
+ if (!getGroupVerticesList()
+ .equals(other.getGroupVerticesList())) return false;
+ if (hasMergedInputDescriptor() != other.hasMergedInputDescriptor()) return false;
if (hasMergedInputDescriptor()) {
- result = result && getMergedInputDescriptor()
- .equals(other.getMergedInputDescriptor());
+ if (!getMergedInputDescriptor()
+ .equals(other.getMergedInputDescriptor())) return false;
}
- result = result &&
- getUnknownFields().equals(other.getUnknownFields());
- return result;
+ if (!getUnknownFields().equals(other.getUnknownFields())) return false;
+ return true;
}
- private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
- hash = (19 * hash) + getDescriptorForType().hashCode();
+ hash = (19 * hash) + getDescriptor().hashCode();
if (hasGroupName()) {
hash = (37 * hash) + GROUP_NAME_FIELD_NUMBER;
hash = (53 * hash) + getGroupName().hashCode();
@@ -2695,6 +2912,17 @@ public final class LlapDaemonProtocolProtos {
return hash;
}
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom(
+ java.nio.ByteBuffer data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom(
+ java.nio.ByteBuffer data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
@@ -2718,46 +2946,61 @@ public final class LlapDaemonProtocolProtos {
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
+
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input);
}
+
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
- public static Builder newBuilder() { return Builder.create(); }
+ @java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder() {
+ return DEFAULT_INSTANCE.toBuilder();
+ }
public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto prototype) {
- return newBuilder().mergeFrom(prototype);
+ return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+ }
+ @java.lang.Override
+ public Builder toBuilder() {
+ return this == DEFAULT_INSTANCE
+ ? new Builder() : new Builder().mergeFrom(this);
}
- public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -2765,14 +3008,16 @@ public final class LlapDaemonProtocolProtos {
* Protobuf type {@code GroupInputSpecProto}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder {
+ com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ // @@protoc_insertion_point(builder_implements:GroupInputSpecProto)
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GroupInputSpecProto_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ @java.lang.Override
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GroupInputSpecProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -2785,47 +3030,43 @@ public final class LlapDaemonProtocolProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ if (com.google.protobuf.GeneratedMessageV3
+ .alwaysUseFieldBuilders) {
getMergedInputDescriptorFieldBuilder();
}
}
- private static Builder create() {
- return new Builder();
- }
-
+ @java.lang.Override
public Builder clear() {
super.clear();
+ bitField0_ = 0;
groupName_ = "";
- bitField0_ = (bitField0_ & ~0x00000001);
- groupVertices_ = com.google.protobuf.LazyStringArrayList.EMPTY;
- bitField0_ = (bitField0_ & ~0x00000002);
- if (mergedInputDescriptorBuilder_ == null) {
- mergedInputDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
- } else {
- mergedInputDescriptorBuilder_.clear();
+ groupVertices_ =
+ com.google.protobuf.LazyStringArrayList.emptyList();
+ mergedInputDescriptor_ = null;
+ if (mergedInputDescriptorBuilder_ != null) {
+ mergedInputDescriptorBuilder_.dispose();
+ mergedInputDescriptorBuilder_ = null;
}
- bitField0_ = (bitField0_ & ~0x00000004);
return this;
}
- public Builder clone() {
- return create().mergeFrom(buildPartial());
- }
-
+ @java.lang.Override
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GroupInputSpecProto_descriptor;
}
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto getDefaultInstanceForType() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.getDefaultInstance();
}
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto build() {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto result = buildPartial();
if (!result.isInitialized()) {
@@ -2834,33 +3075,67 @@ public final class LlapDaemonProtocolProtos {
return result;
}
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto buildPartial() {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto(this);
+ if (bitField0_ != 0) { buildPartial0(result); }
+ onBuilt();
+ return result;
+ }
+
+ private void buildPartial0(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
- if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
+ if (((from_bitField0_ & 0x00000001) != 0)) {
+ result.groupName_ = groupName_;
to_bitField0_ |= 0x00000001;
}
- result.groupName_ = groupName_;
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
- groupVertices_ = new com.google.protobuf.UnmodifiableLazyStringList(
- groupVertices_);
- bitField0_ = (bitField0_ & ~0x00000002);
+ if (((from_bitField0_ & 0x00000002) != 0)) {
+ groupVertices_.makeImmutable();
+ result.groupVertices_ = groupVertices_;
}
- result.groupVertices_ = groupVertices_;
- if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
+ if (((from_bitField0_ & 0x00000004) != 0)) {
+ result.mergedInputDescriptor_ = mergedInputDescriptorBuilder_ == null
+ ? mergedInputDescriptor_
+ : mergedInputDescriptorBuilder_.build();
to_bitField0_ |= 0x00000002;
}
- if (mergedInputDescriptorBuilder_ == null) {
- result.mergedInputDescriptor_ = mergedInputDescriptor_;
- } else {
- result.mergedInputDescriptor_ = mergedInputDescriptorBuilder_.build();
- }
- result.bitField0_ = to_bitField0_;
- onBuilt();
- return result;
+ result.bitField0_ |= to_bitField0_;
}
+ @java.lang.Override
+ public Builder clone() {
+ return super.clone();
+ }
+ @java.lang.Override
+ public Builder setField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ java.lang.Object value) {
+ return super.setField(field, value);
+ }
+ @java.lang.Override
+ public Builder clearField(
+ com.google.protobuf.Descriptors.FieldDescriptor field) {
+ return super.clearField(field);
+ }
+ @java.lang.Override
+ public Builder clearOneof(
+ com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ return super.clearOneof(oneof);
+ }
+ @java.lang.Override
+ public Builder setRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ int index, java.lang.Object value) {
+ return super.setRepeatedField(field, index, value);
+ }
+ @java.lang.Override
+ public Builder addRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ java.lang.Object value) {
+ return super.addRepeatedField(field, value);
+ }
+ @java.lang.Override
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto) {
return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto)other);
@@ -2873,14 +3148,14 @@ public final class LlapDaemonProtocolProtos {
public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto other) {
if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.getDefaultInstance()) return this;
if (other.hasGroupName()) {
- bitField0_ |= 0x00000001;
groupName_ = other.groupName_;
+ bitField0_ |= 0x00000001;
onChanged();
}
if (!other.groupVertices_.isEmpty()) {
if (groupVertices_.isEmpty()) {
groupVertices_ = other.groupVertices_;
- bitField0_ = (bitField0_ & ~0x00000002);
+ bitField0_ |= 0x00000002;
} else {
ensureGroupVerticesIsMutable();
groupVertices_.addAll(other.groupVertices_);
@@ -2891,49 +3166,87 @@ public final class LlapDaemonProtocolProtos {
mergeMergedInputDescriptor(other.getMergedInputDescriptor());
}
this.mergeUnknownFields(other.getUnknownFields());
+ onChanged();
return this;
}
+ @java.lang.Override
public final boolean isInitialized() {
return true;
}
+ @java.lang.Override
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parsedMessage = null;
+ if (extensionRegistry == null) {
+ throw new java.lang.NullPointerException();
+ }
try {
- parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ case 10: {
+ groupName_ = input.readBytes();
+ bitField0_ |= 0x00000001;
+ break;
+ } // case 10
+ case 18: {
+ com.google.protobuf.ByteString bs = input.readBytes();
+ ensureGroupVerticesIsMutable();
+ groupVertices_.add(bs);
+ break;
+ } // case 18
+ case 26: {
+ input.readMessage(
+ getMergedInputDescriptorFieldBuilder().getBuilder(),
+ extensionRegistry);
+ bitField0_ |= 0x00000004;
+ break;
+ } // case 26
+ default: {
+ if (!super.parseUnknownField(input, extensionRegistry, tag)) {
+ done = true; // was an endgroup tag
+ }
+ break;
+ } // default:
+ } // switch (tag)
+ } // while (!done)
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
- parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto) e.getUnfinishedMessage();
- throw e;
+ throw e.unwrapIOException();
} finally {
- if (parsedMessage != null) {
- mergeFrom(parsedMessage);
- }
- }
+ onChanged();
+ } // finally
return this;
}
private int bitField0_;
- // optional string group_name = 1;
private java.lang.Object groupName_ = "";
/**
* <code>optional string group_name = 1;</code>
+ * @return Whether the groupName field is set.
*/
public boolean hasGroupName() {
- return ((bitField0_ & 0x00000001) == 0x00000001);
+ return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string group_name = 1;</code>
+ * @return The groupName.
*/
public java.lang.String getGroupName() {
java.lang.Object ref = groupName_;
if (!(ref instanceof java.lang.String)) {
- java.lang.String s = ((com.google.protobuf.ByteString) ref)
- .toStringUtf8();
- groupName_ = s;
+ com.google.protobuf.ByteString bs =
+ (com.google.protobuf.ByteString) ref;
+ java.lang.String s = bs.toStringUtf8();
+ if (bs.isValidUtf8()) {
+ groupName_ = s;
+ }
return s;
} else {
return (java.lang.String) ref;
@@ -2941,6 +3254,7 @@ public final class LlapDaemonProtocolProtos {
}
/**
* <code>optional string group_name = 1;</code>
+ * @return The bytes for groupName.
*/
public com.google.protobuf.ByteString
getGroupNameBytes() {
@@ -2957,69 +3271,77 @@ public final class LlapDaemonProtocolProtos {
}
/**
* <code>optional string group_name = 1;</code>
+ * @param value The groupName to set.
+ * @return This builder for chaining.
*/
public Builder setGroupName(
java.lang.String value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000001;
+ if (value == null) { throw new NullPointerException(); }
groupName_ = value;
+ bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional string group_name = 1;</code>
+ * @return This builder for chaining.
*/
public Builder clearGroupName() {
- bitField0_ = (bitField0_ & ~0x00000001);
groupName_ = getDefaultInstance().getGroupName();
+ bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
return this;
}
/**
* <code>optional string group_name = 1;</code>
+ * @param value The bytes for groupName to set.
+ * @return This builder for chaining.
*/
public Builder setGroupNameBytes(
com.google.protobuf.ByteString value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000001;
+ if (value == null) { throw new NullPointerException(); }
groupName_ = value;
+ bitField0_ |= 0x00000001;
onChanged();
return this;
}
- // repeated string group_vertices = 2;
- private com.google.protobuf.LazyStringList groupVertices_ = com.google.protobuf.LazyStringArrayList.EMPTY;
+ private com.google.protobuf.LazyStringArrayList groupVertices_ =
+ com.google.protobuf.LazyStringArrayList.emptyList();
private void ensureGroupVerticesIsMutable() {
- if (!((bitField0_ & 0x00000002) == 0x00000002)) {
+ if (!groupVertices_.isModifiable()) {
groupVertices_ = new com.google.protobuf.LazyStringArrayList(groupVertices_);
- bitField0_ |= 0x00000002;
- }
+ }
+ bitField0_ |= 0x00000002;
}
/**
* <code>repeated string group_vertices = 2;</code>
+ * @return A list containing the groupVertices.
*/
- public java.util.List<java.lang.String>
+ public com.google.protobuf.ProtocolStringList
getGroupVerticesList() {
- return java.util.Collections.unmodifiableList(groupVertices_);
+ groupVertices_.makeImmutable();
+ return groupVertices_;
}
/**
* <code>repeated string group_vertices = 2;</code>
+ * @return The count of groupVertices.
*/
public int getGroupVerticesCount() {
return groupVertices_.size();
}
/**
* <code>repeated string group_vertices = 2;</code>
+ * @param index The index of the element to return.
+ * @return The groupVertices at the given index.
*/
public java.lang.String getGroupVertices(int index) {
return groupVertices_.get(index);
}
/**
* <code>repeated string group_vertices = 2;</code>
+ * @param index The index of the value to return.
+ * @return The bytes of the groupVertices at the given index.
*/
public com.google.protobuf.ByteString
getGroupVerticesBytes(int index) {
@@ -3027,79 +3349,90 @@ public final class LlapDaemonProtocolProtos {
}
/**
* <code>repeated string group_vertices = 2;</code>
+ * @param index The index to set the value at.
+ * @param value The groupVertices to set.
+ * @return This builder for chaining.
*/
public Builder setGroupVertices(
int index, java.lang.String value) {
- if (value == null) {
- throw new NullPointerException();
- }
- ensureGroupVerticesIsMutable();
+ if (value == null) { throw new NullPointerException(); }
+ ensureGroupVerticesIsMutable();
groupVertices_.set(index, value);
+ bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* <code>repeated string group_vertices = 2;</code>
+ * @param value The groupVertices to add.
+ * @return This builder for chaining.
*/
public Builder addGroupVertices(
java.lang.String value) {
- if (value == null) {
- throw new NullPointerException();
- }
- ensureGroupVerticesIsMutable();
+ if (value == null) { throw new NullPointerException(); }
+ ensureGroupVerticesIsMutable();
groupVertices_.add(value);
+ bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* <code>repeated string group_vertices = 2;</code>
+ * @param values The groupVertices to add.
+ * @return This builder for chaining.
*/
public Builder addAllGroupVertices(
java.lang.Iterable<java.lang.String> values) {
ensureGroupVerticesIsMutable();
- super.addAll(values, groupVertices_);
+ com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ values, groupVertices_);
+ bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* <code>repeated string group_vertices = 2;</code>
+ * @return This builder for chaining.
*/
public Builder clearGroupVertices() {
- groupVertices_ = com.google.protobuf.LazyStringArrayList.EMPTY;
- bitField0_ = (bitField0_ & ~0x00000002);
+ groupVertices_ =
+ com.google.protobuf.LazyStringArrayList.emptyList();
+ bitField0_ = (bitField0_ & ~0x00000002);;
onChanged();
return this;
}
/**
* <code>repeated string group_vertices = 2;</code>
+ * @param value The bytes of the groupVertices to add.
+ * @return This builder for chaining.
*/
public Builder addGroupVerticesBytes(
com.google.protobuf.ByteString value) {
- if (value == null) {
- throw new NullPointerException();
- }
- ensureGroupVerticesIsMutable();
+ if (value == null) { throw new NullPointerException(); }
+ ensureGroupVerticesIsMutable();
groupVertices_.add(value);
+ bitField0_ |= 0x00000002;
onChanged();
return this;
}
- // optional .EntityDescriptorProto merged_input_descriptor = 3;
- private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto mergedInputDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
- private com.google.protobuf.SingleFieldBuilder<
+ private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto mergedInputDescriptor_;
+ private com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> mergedInputDescriptorBuilder_;
/**
* <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code>
+ * @return Whether the mergedInputDescriptor field is set.
*/
public boolean hasMergedInputDescriptor() {
- return ((bitField0_ & 0x00000004) == 0x00000004);
+ return ((bitField0_ & 0x00000004) != 0);
}
/**
* <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code>
+ * @return The mergedInputDescriptor.
*/
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getMergedInputDescriptor() {
if (mergedInputDescriptorBuilder_ == null) {
- return mergedInputDescriptor_;
+ return mergedInputDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : mergedInputDescriptor_;
} else {
return mergedInputDescriptorBuilder_.getMessage();
}
@@ -3113,11 +3446,11 @@ public final class LlapDaemonProtocolProtos {
throw new NullPointerException();
}
mergedInputDescriptor_ = value;
- onChanged();
} else {
mergedInputDescriptorBuilder_.setMessage(value);
}
bitField0_ |= 0x00000004;
+ onChanged();
return this;
}
/**
@@ -3127,11 +3460,11 @@ public final class LlapDaemonProtocolProtos {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder builderForValue) {
if (mergedInputDescriptorBuilder_ == null) {
mergedInputDescriptor_ = builderForValue.build();
- onChanged();
} else {
mergedInputDescriptorBuilder_.setMessage(builderForValue.build());
}
bitField0_ |= 0x00000004;
+ onChanged();
return this;
}
/**
@@ -3139,31 +3472,33 @@ public final class LlapDaemonProtocolProtos {
*/
public Builder mergeMergedInputDescriptor(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto value) {
if (mergedInputDescriptorBuilder_ == null) {
- if (((bitField0_ & 0x00000004) == 0x00000004) &&
- mergedInputDescriptor_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) {
- mergedInputDescriptor_ =
- org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.newBuilder(mergedInputDescriptor_).mergeFrom(value).buildPartial();
+ if (((bitField0_ & 0x00000004) != 0) &&
+ mergedInputDescriptor_ != null &&
+ mergedInputDescriptor_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) {
+ getMergedInputDescriptorBuilder().mergeFrom(value);
} else {
mergedInputDescriptor_ = value;
}
- onChanged();
} else {
mergedInputDescriptorBuilder_.mergeFrom(value);
}
- bitField0_ |= 0x00000004;
+ if (mergedInputDescriptor_ != null) {
+ bitField0_ |= 0x00000004;
+ onChanged();
+ }
return this;
}
/**
* <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code>
*/
public Builder clearMergedInputDescriptor() {
- if (mergedInputDescriptorBuilder_ == null) {
- mergedInputDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
- onChanged();
- } else {
- mergedInputDescriptorBuilder_.clear();
- }
bitField0_ = (bitField0_ & ~0x00000004);
+ mergedInputDescriptor_ = null;
+ if (mergedInputDescriptorBuilder_ != null) {
+ mergedInputDescriptorBuilder_.dispose();
+ mergedInputDescriptorBuilder_ = null;
+ }
+ onChanged();
return this;
}
/**
@@ -3181,72 +3516,130 @@ public final class LlapDaemonProtocolProtos {
if (mergedInputDescriptorBuilder_ != null) {
return mergedInputDescriptorBuilder_.getMessageOrBuilder();
} else {
- return mergedInputDescriptor_;
+ return mergedInputDescriptor_ == null ?
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : mergedInputDescriptor_;
}
}
/**
* <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code>
*/
- private com.google.protobuf.SingleFieldBuilder<
+ private com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder>
getMergedInputDescriptorFieldBuilder() {
if (mergedInputDescriptorBuilder_ == null) {
- mergedInputDescriptorBuilder_ = new com.google.protobuf.SingleFieldBuilder<
+ mergedInputDescriptorBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder>(
- mergedInputDescriptor_,
+ getMergedInputDescriptor(),
getParentForChildren(),
isClean());
mergedInputDescriptor_ = null;
}
return mergedInputDescriptorBuilder_;
}
+ @java.lang.Override
+ public final Builder setUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.setUnknownFields(unknownFields);
+ }
+
+ @java.lang.Override
+ public final Builder mergeUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.mergeUnknownFields(unknownFields);
+ }
+
// @@protoc_insertion_point(builder_scope:GroupInputSpecProto)
}
+ // @@protoc_insertion_point(class_scope:GroupInputSpecProto)
+ private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto DEFAULT_INSTANCE;
static {
- defaultInstance = new GroupInputSpecProto(true);
- defaultInstance.initFields();
+ DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto();
+ }
+
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto getDefaultInstance() {
+ return DEFAULT_INSTANCE;
+ }
+
+ @java.lang.Deprecated public static final com.google.protobuf.Parser<GroupInputSpecProto>
+ PARSER = new com.google.protobuf.AbstractParser<GroupInputSpecProto>() {
+ @java.lang.Override
+ public GroupInputSpecProto parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ Builder builder = newBuilder();
+ try {
+ builder.mergeFrom(input, extensionRegistry);
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(builder.buildPartial());
+ } catch (com.google.protobuf.UninitializedMessageException e) {
+ throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(e)
+ .setUnfinishedMessage(builder.buildPartial());
+ }
+ return builder.buildPartial();
+ }
+ };
+
+ public static com.google.protobuf.Parser<GroupInputSpecProto> parser() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<GroupInputSpecProto> getParserForType() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto getDefaultInstanceForType() {
+ return DEFAULT_INSTANCE;
}
- // @@protoc_insertion_point(class_scope:GroupInputSpecProto)
}
- public interface SignableVertexSpecOrBuilder
- extends com.google.protobuf.MessageOrBuilder {
+ public interface SignableVertexSpecOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:SignableVertexSpec)
+ com.google.protobuf.MessageOrBuilder {
- // optional string user = 1;
/**
* <code>optional string user = 1;</code>
+ * @return Whether the user field is set.
*/
boolean hasUser();
/**
* <code>optional string user = 1;</code>
+ * @return The user.
*/
java.lang.String getUser();
/**
* <code>optional string user = 1;</code>
+ * @return The bytes for user.
*/
com.google.protobuf.ByteString
getUserBytes();
- // optional int64 signatureKeyId = 2;
/**
* <code>optional int64 signatureKeyId = 2;</code>
+ * @return Whether the signatureKeyId field is set.
*/
boolean hasSignatureKeyId();
/**
* <code>optional int64 signatureKeyId = 2;</code>
+ * @return The signatureKeyId.
*/
long getSignatureKeyId();
- // optional .QueryIdentifierProto query_identifier = 3;
/**
* <code>optional .QueryIdentifierProto query_identifier = 3;</code>
+ * @return Whether the queryIdentifier field is set.
*/
boolean hasQueryIdentifier();
/**
* <code>optional .QueryIdentifierProto query_identifier = 3;</code>
+ * @return The queryIdentifier.
*/
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier();
/**
@@ -3254,107 +3647,117 @@ public final class LlapDaemonProtocolProtos {
*/
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder();
- // optional string hive_query_id = 4;
/**
* <code>optional string hive_query_id = 4;</code>
+ * @return Whether the hiveQueryId field is set.
*/
boolean hasHiveQueryId();
/**
* <code>optional string hive_query_id = 4;</code>
+ * @return The hiveQueryId.
*/
java.lang.String getHiveQueryId();
/**
* <code>optional string hive_query_id = 4;</code>
+ * @return The bytes for hiveQueryId.
*/
com.google.protobuf.ByteString
getHiveQueryIdBytes();
- // optional string dag_name = 5;
/**
- * <code>optional string dag_name = 5;</code>
- *
* <pre>
* Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
* </pre>
+ *
+ * <code>optional string dag_name = 5;</code>
+ * @return Whether the dagName field is set.
*/
boolean hasDagName();
/**
- * <code>optional string dag_name = 5;</code>
- *
* <pre>
* Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
* </pre>
+ *
+ * <code>optional string dag_name = 5;</code>
+ * @return The dagName.
*/
java.lang.String getDagName();
/**
- * <code>optional string dag_name = 5;</code>
- *
* <pre>
* Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
* </pre>
+ *
+ * <code>optional string dag_name = 5;</code>
+ * @return The bytes for dagName.
*/
com.google.protobuf.ByteString
getDagNameBytes();
- // optional string vertex_name = 6;
/**
* <code>optional string vertex_name = 6;</code>
+ * @return Whether the vertexName field is set.
*/
boolean hasVertexName();
/**
* <code>optional string vertex_name = 6;</code>
+ * @return The vertexName.
*/
java.lang.String getVertexName();
/**
* <code>optional string vertex_name = 6;</code>
+ * @return The bytes for vertexName.
*/
com.google.protobuf.ByteString
getVertexNameBytes();
- // optional int32 vertex_index = 7;
/**
* <code>optional int32 vertex_index = 7;</code>
+ * @return Whether the vertexIndex field is set.
*/
boolean hasVertexIndex();
/**
* <code>optional int32 vertex_index = 7;</code>
+ * @return The vertexIndex.
*/
int getVertexIndex();
- // optional string token_identifier = 8;
/**
- * <code>optional string token_identifier = 8;</code>
- *
* <pre>
* The core vertex stuff
* </pre>
+ *
+ * <code>optional string token_identifier = 8;</code>
+ * @return Whether the tokenIdentifier field is set.
*/
boolean hasTokenIdentifier();
/**
- * <code>optional string token_identifier = 8;</code>
- *
* <pre>
* The core vertex stuff
* </pre>
+ *
+ * <code>optional string token_identifier = 8;</code>
+ * @return The tokenIdentifier.
*/
java.lang.String getTokenIdentifier();
/**
- * <code>optional string token_identifier = 8;</code>
- *
* <pre>
* The core vertex stuff
* </pre>
+ *
+ * <code>optional string token_identifier = 8;</code>
+ * @return The bytes for tokenIdentifier.
*/
com.google.protobuf.ByteString
getTokenIdentifierBytes();
- // optional .EntityDescriptorProto processor_descriptor = 9;
/**
* <code>optional .EntityDescriptorProto processor_descriptor = 9;</code>
+ * @return Whether the processorDescriptor field is set.
*/
boolean hasProcessorDescriptor();
/**
* <code>optional .EntityDescriptorProto processor_descriptor = 9;</code>
+ * @return The processorDescriptor.
*/
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getProcessorDescriptor();
/**
@@ -3362,7 +3765,6 @@ public final class LlapDaemonProtocolProtos {
*/
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getProcessorDescriptorOrBuilder();
- // repeated .IOSpecProto input_specs = 10;
/**
* <code>repeated .IOSpecProto input_specs = 10;</code>
*/
@@ -3387,7 +3789,6 @@ public final class LlapDaemonProtocolProtos {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder getInputSpecsOrBuilder(
int index);
- // repeated .IOSpecProto output_specs = 11;
/**
* <code>repeated .IOSpecProto output_specs = 11;</code>
*/
@@ -3412,7 +3813,6 @@ public final class LlapDaemonProtocolProtos {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder getOutputSpecsOrBuilder(
int index);
- // repeated .GroupInputSpecProto grouped_input_specs = 12;
/**
* <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code>
*/
@@ -3437,245 +3837,100 @@ public final class LlapDaemonProtocolProtos {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder getGroupedInputSpecsOrBuilder(
int index);
- // optional int32 vertex_parallelism = 13;
/**
- * <code>optional int32 vertex_parallelism = 13;</code>
- *
* <pre>
* An internal field required for Tez.
* </pre>
+ *
+ * <code>optional int32 vertex_parallelism = 13;</code>
+ * @return Whether the vertexParallelism field is set.
*/
boolean hasVertexParallelism();
/**
- * <code>optional int32 vertex_parallelism = 13;</code>
- *
* <pre>
* An internal field required for Tez.
* </pre>
+ *
+ * <code>optional int32 vertex_parallelism = 13;</code>
+ * @return The vertexParallelism.
*/
int getVertexParallelism();
- // optional bool is_external_submission = 14 [default = false];
/**
* <code>optional bool is_external_submission = 14 [default = false];</code>
+ * @return Whether the isExternalSubmission field is set.
*/
boolean hasIsExternalSubmission();
/**
* <code>optional bool is_external_submission = 14 [default = false];</code>
+ * @return The isExternalSubmission.
*/
boolean getIsExternalSubmission();
}
/**
- * Protobuf type {@code SignableVertexSpec}
- *
* <pre>
* The part of SubmitWork that can be signed
* </pre>
+ *
+ * Protobuf type {@code SignableVertexSpec}
*/
public static final class SignableVertexSpec extends
- com.google.protobuf.GeneratedMessage
- implements SignableVertexSpecOrBuilder {
+ com.google.protobuf.GeneratedMessageV3 implements
+ // @@protoc_insertion_point(message_implements:SignableVertexSpec)
+ SignableVertexSpecOrBuilder {
+ private static final long serialVersionUID = 0L;
// Use SignableVertexSpec.newBuilder() to construct.
- private SignableVertexSpec(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ private SignableVertexSpec(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
- this.unknownFields = builder.getUnknownFields();
- }
- private SignableVertexSpec(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
- private static final SignableVertexSpec defaultInstance;
- public static SignableVertexSpec getDefaultInstance() {
- return defaultInstance;
}
-
- public SignableVertexSpec getDefaultInstanceForType() {
- return defaultInstance;
+ private SignableVertexSpec() {
+ user_ = "";
+ hiveQueryId_ = "";
+ dagName_ = "";
+ vertexName_ = "";
+ tokenIdentifier_ = "";
+ inputSpecs_ = java.util.Collections.emptyList();
+ outputSpecs_ = java.util.Collections.emptyList();
+ groupedInputSpecs_ = java.util.Collections.emptyList();
}
- private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
- return this.unknownFields;
- }
- private SignableVertexSpec(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- initFields();
- int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
- try {
- boolean done = false;
- while (!done) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
- done = true;
- break;
- default: {
- if (!parseUnknownField(input, unknownFields,
- extensionRegistry, tag)) {
- done = true;
- }
- break;
- }
- case 10: {
- bitField0_ |= 0x00000001;
- user_ = input.readBytes();
- break;
- }
- case 16: {
- bitField0_ |= 0x00000002;
- signatureKeyId_ = input.readInt64();
- break;
- }
- case 26: {
- org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder subBuilder = null;
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
- subBuilder = queryIdentifier_.toBuilder();
- }
- queryIdentifier_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.PARSER, extensionRegistry);
- if (subBuilder != null) {
- subBuilder.mergeFrom(queryIdentifier_);
- queryIdentifier_ = subBuilder.buildPartial();
- }
- bitField0_ |= 0x00000004;
- break;
- }
- case 34: {
- bitField0_ |= 0x00000008;
- hiveQueryId_ = input.readBytes();
- break;
- }
- case 42: {
- bitField0_ |= 0x00000010;
- dagName_ = input.readBytes();
- break;
- }
- case 50: {
- bitField0_ |= 0x00000020;
- vertexName_ = input.readBytes();
- break;
- }
- case 56: {
- bitField0_ |= 0x00000040;
- vertexIndex_ = input.readInt32();
- break;
- }
- case 66: {
- bitField0_ |= 0x00000080;
- tokenIdentifier_ = input.readBytes();
- break;
- }
- case 74: {
- org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder subBuilder = null;
- if (((bitField0_ & 0x00000100) == 0x00000100)) {
- subBuilder = processorDescriptor_.toBuilder();
- }
- processorDescriptor_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.PARSER, extensionRegistry);
- if (subBuilder != null) {
- subBuilder.mergeFrom(processorDescriptor_);
- processorDescriptor_ = subBuilder.buildPartial();
- }
- bitField0_ |= 0x00000100;
- break;
- }
- case 82: {
- if (!((mutable_bitField0_ & 0x00000200) == 0x00000200)) {
- inputSpecs_ = new java.util.ArrayList<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto>();
- mutable_bitField0_ |= 0x00000200;
- }
- inputSpecs_.add(input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.PARSER, extensionRegistry));
- break;
- }
- case 90: {
- if (!((mutable_bitField0_ & 0x00000400) == 0x00000400)) {
- outputSpecs_ = new java.util.ArrayList<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto>();
- mutable_bitField0_ |= 0x00000400;
- }
- outputSpecs_.add(input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.PARSER, extensionRegistry));
- break;
- }
- case 98: {
- if (!((mutable_bitField0_ & 0x00000800) == 0x00000800)) {
- groupedInputSpecs_ = new java.util.ArrayList<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto>();
- mutable_bitField0_ |= 0x00000800;
- }
- groupedInputSpecs_.add(input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.PARSER, extensionRegistry));
- break;
- }
- case 104: {
- bitField0_ |= 0x00000200;
- vertexParallelism_ = input.readInt32();
- break;
- }
- case 112: {
- bitField0_ |= 0x00000400;
- isExternalSubmission_ = input.readBool();
- break;
- }
- }
- }
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
- throw e.setUnfinishedMessage(this);
- } catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
- e.getMessage()).setUnfinishedMessage(this);
- } finally {
- if (((mutable_bitField0_ & 0x00000200) == 0x00000200)) {
- inputSpecs_ = java.util.Collections.unmodifiableList(inputSpecs_);
- }
- if (((mutable_bitField0_ & 0x00000400) == 0x00000400)) {
- outputSpecs_ = java.util.Collections.unmodifiableList(outputSpecs_);
- }
- if (((mutable_bitField0_ & 0x00000800) == 0x00000800)) {
- groupedInputSpecs_ = java.util.Collections.unmodifiableList(groupedInputSpecs_);
- }
- this.unknownFields = unknownFields.build();
- makeExtensionsImmutable();
- }
+ @SuppressWarnings({"unused"})
+ protected java.lang.Object newInstance(
+ UnusedPrivateParameter unused) {
+ return new SignableVertexSpec();
}
+
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SignableVertexSpec_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ @java.lang.Override
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SignableVertexSpec_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder.class);
}
- public static com.google.protobuf.Parser<SignableVertexSpec> PARSER =
- new com.google.protobuf.AbstractParser<SignableVertexSpec>() {
- public SignableVertexSpec parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return new SignableVertexSpec(input, extensionRegistry);
- }
- };
-
- @java.lang.Override
- public com.google.protobuf.Parser<SignableVertexSpec> getParserForType() {
- return PARSER;
- }
-
private int bitField0_;
- // optional string user = 1;
public static final int USER_FIELD_NUMBER = 1;
- private java.lang.Object user_;
+ @SuppressWarnings("serial")
+ private volatile java.lang.Object user_ = "";
/**
* <code>optional string user = 1;</code>
+ * @return Whether the user field is set.
*/
+ @java.lang.Override
public boolean hasUser() {
- return ((bitField0_ & 0x00000001) == 0x00000001);
+ return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string user = 1;</code>
+ * @return The user.
*/
+ @java.lang.Override
public java.lang.String getUser() {
java.lang.Object ref = user_;
if (ref instanceof java.lang.String) {
@@ -3692,7 +3947,9 @@ public final class LlapDaemonProtocolProtos {
}
/**
* <code>optional string user = 1;</code>
+ * @return The bytes for user.
*/
+ @java.lang.Override
public com.google.protobuf.ByteString
getUserBytes() {
java.lang.Object ref = user_;
@@ -3707,56 +3964,67 @@ public final class LlapDaemonProtocolProtos {
}
}
- // optional int64 signatureKeyId = 2;
public static final int SIGNATUREKEYID_FIELD_NUMBER = 2;
- private long signatureKeyId_;
+ private long signatureKeyId_ = 0L;
/**
* <code>optional int64 signatureKeyId = 2;</code>
+ * @return Whether the signatureKeyId field is set.
*/
+ @java.lang.Override
public boolean hasSignatureKeyId() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
+ return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional int64 signatureKeyId = 2;</code>
+ * @return The signatureKeyId.
*/
+ @java.lang.Override
public long getSignatureKeyId() {
return signatureKeyId_;
}
- // optional .QueryIdentifierProto query_identifier = 3;
public static final int QUERY_IDENTIFIER_FIELD_NUMBER = 3;
private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_;
/**
* <code>optional .QueryIdentifierProto query_identifier = 3;</code>
+ * @return Whether the queryIdentifier field is set.
*/
+ @java.lang.Override
public boolean hasQueryIdentifier() {
- return ((bitField0_ & 0x00000004) == 0x00000004);
+ return ((bitField0_ & 0x00000004) != 0);
}
/**
* <code>optional .QueryIdentifierProto query_identifier = 3;</code>
+ * @return The queryIdentifier.
*/
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() {
- return queryIdentifier_;
+ return queryIdentifier_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_;
}
/**
* <code>optional .QueryIdentifierProto query_identifier = 3;</code>
*/
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder() {
- return queryIdentifier_;
+ return queryIdentifier_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_;
}
- // optional string hive_query_id = 4;
public static final int HIVE_QUERY_ID_FIELD_NUMBER = 4;
- private java.lang.Object hiveQueryId_;
+ @SuppressWarnings("serial")
+ private volatile java.lang.Object hiveQueryId_ = "";
/**
* <code>optional string hive_query_id = 4;</code>
+ * @return Whether the hiveQueryId field is set.
*/
+ @java.lang.Override
public boolean hasHiveQueryId() {
- return ((bitField0_ & 0x00000008) == 0x00000008);
+ return ((bitField0_ & 0x00000008) != 0);
}
/**
* <code>optional string hive_query_id = 4;</code>
+ * @return The hiveQueryId.
*/
+ @java.lang.Override
public java.lang.String getHiveQueryId() {
java.lang.Object ref = hiveQueryId_;
if (ref instanceof java.lang.String) {
@@ -3773,7 +4041,9 @@ public final class LlapDaemonProtocolProtos {
}
/**
* <code>optional string hive_query_id = 4;</code>
+ * @return The bytes for hiveQueryId.
*/
+ @java.lang.Override
public com.google.protobuf.ByteString
getHiveQueryIdBytes() {
java.lang.Object ref = hiveQueryId_;
@@ -3788,26 +4058,30 @@ public final class LlapDaemonProtocolProtos {
}
}
- // optional string dag_name = 5;
public static final int DAG_NAME_FIELD_NUMBER = 5;
- private java.lang.Object dagName_;
+ @SuppressWarnings("serial")
+ private volatile java.lang.Object dagName_ = "";
/**
- * <code>optional string dag_name = 5;</code>
- *
* <pre>
* Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
* </pre>
+ *
+ * <code>optional string dag_name = 5;</code>
+ * @return Whether the dagName field is set.
*/
+ @java.lang.Override
public boolean hasDagName() {
- return ((bitField0_ & 0x00000010) == 0x00000010);
+ return ((bitField0_ & 0x00000010) != 0);
}
/**
- * <code>optional string dag_name = 5;</code>
- *
* <pre>
* Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
* </pre>
+ *
+ * <code>optional string dag_name = 5;</code>
+ * @return The dagName.
*/
+ @java.lang.Override
public java.lang.String getDagName() {
java.lang.Object ref = dagName_;
if (ref instanceof java.lang.String) {
@@ -3823,12 +4097,14 @@ public final class LlapDaemonProtocolProtos {
}
}
/**
- * <code>optional string dag_name = 5;</code>
- *
* <pre>
* Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
* </pre>
+ *
+ * <code>optional string dag_name = 5;</code>
+ * @return The bytes for dagName.
*/
+ @java.lang.Override
public com.google.protobuf.ByteString
getDagNameBytes() {
java.lang.Object ref = dagName_;
@@ -3843,18 +4119,22 @@ public final class LlapDaemonProtocolProtos {
}
}
- // optional string vertex_name = 6;
public static final int VERTEX_NAME_FIELD_NUMBER = 6;
- private java.lang.Object vertexName_;
+ @SuppressWarnings("serial")
+ private volatile java.lang.Object vertexName_ = "";
/**
* <code>optional string vertex_name = 6;</code>
+ * @return Whether the vertexName field is set.
*/
+ @java.lang.Override
public boolean hasVertexName() {
- return ((bitField0_ & 0x00000020) == 0x00000020);
+ return ((bitField0_ & 0x00000020) != 0);
}
/**
* <code>optional string vertex_name = 6;</code>
+ * @return The vertexName.
*/
+ @java.lang.Override
public java.lang.String getVertexName() {
java.lang.Object ref = vertexName_;
if (ref instanceof java.lang.String) {
@@ -3871,7 +4151,9 @@ public final class LlapDaemonProtocolProtos {
}
/**
* <code>optional string vertex_name = 6;</code>
+ * @return The bytes for vertexName.
*/
+ @java.lang.Override
public com.google.protobuf.ByteString
getVertexNameBytes() {
java.lang.Object ref = vertexName_;
@@ -3886,42 +4168,49 @@ public final class LlapDaemonProtocolProtos {
}
}
- // optional int32 vertex_index = 7;
public static final int VERTEX_INDEX_FIELD_NUMBER = 7;
- private int vertexIndex_;
+ private int vertexIndex_ = 0;
/**
* <code>optional int32 vertex_index = 7;</code>
+ * @return Whether the vertexIndex field is set.
*/
+ @java.lang.Override
public boolean hasVertexIndex() {
- return ((bitField0_ & 0x00000040) == 0x00000040);
+ return ((bitField0_ & 0x00000040) != 0);
}
/**
* <code>optional int32 vertex_index = 7;</code>
+ * @return The vertexIndex.
*/
+ @java.lang.Override
public int getVertexIndex() {
return vertexIndex_;
}
- // optional string token_identifier = 8;
public static final int TOKEN_IDENTIFIER_FIELD_NUMBER = 8;
- private java.lang.Object tokenIdentifier_;
+ @SuppressWarnings("serial")
+ private volatile java.lang.Object tokenIdentifier_ = "";
/**
- * <code>optional string token_identifier = 8;</code>
- *
* <pre>
* The core vertex stuff
* </pre>
+ *
+ * <code>optional string token_identifier = 8;</code>
+ * @return Whether the tokenIdentifier field is set.
*/
+ @java.lang.Override
public boolean hasTokenIdentifier() {
- return ((bitField0_ & 0x00000080) == 0x00000080);
+ return ((bitField0_ & 0x00000080) != 0);
}
/**
- * <code>optional string token_identifier = 8;</code>
- *
* <pre>
* The core vertex stuff
* </pre>
+ *
+ * <code>optional string token_identifier = 8;</code>
+ * @return The tokenIdentifier.
*/
+ @java.lang.Override
public java.lang.String getTokenIdentifier() {
java.lang.Object ref = tokenIdentifier_;
if (ref instanceof java.lang.String) {
@@ -3937,12 +4226,14 @@ public final class LlapDaemonProtocolProtos {
}
}
/**
- * <code>optional string token_identifier = 8;</code>
- *
* <pre>
* The core vertex stuff
* </pre>
+ *
+ * <code>optional string token_identifier = 8;</code>
+ * @return The bytes for tokenIdentifier.
*/
+ @java.lang.Override
public com.google.protobuf.ByteString
getTokenIdentifierBytes() {
java.lang.Object ref = tokenIdentifier_;
@@ -3957,40 +4248,46 @@ public final class LlapDaemonProtocolProtos {
}
}
- // optional .EntityDescriptorProto processor_descriptor = 9;
public static final int PROCESSOR_DESCRIPTOR_FIELD_NUMBER = 9;
private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto processorDescriptor_;
/**
* <code>optional .EntityDescriptorProto processor_descriptor = 9;</code>
+ * @return Whether the processorDescriptor field is set.
*/
+ @java.lang.Override
public boolean hasProcessorDescriptor() {
- return ((bitField0_ & 0x00000100) == 0x00000100);
+ return ((bitField0_ & 0x00000100) != 0);
}
/**
* <code>optional .EntityDescriptorProto processor_descriptor = 9;</code>
+ * @return The processorDescriptor.
*/
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getProcessorDescriptor() {
- return processorDescriptor_;
+ return processorDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : processorDescriptor_;
}
/**
* <code>optional .EntityDescriptorProto processor_descriptor = 9;</code>
*/
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getProcessorDescriptorOrBuilder() {
- return processorDescriptor_;
+ return processorDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : processorDescriptor_;
}
- // repeated .IOSpecProto input_specs = 10;
public static final int INPUT_SPECS_FIELD_NUMBER = 10;
+ @SuppressWarnings("serial")
private java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> inputSpecs_;
/**
* <code>repeated .IOSpecProto input_specs = 10;</code>
*/
+ @java.lang.Override
public java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> getInputSpecsList() {
return inputSpecs_;
}
/**
* <code>repeated .IOSpecProto input_specs = 10;</code>
*/
+ @java.lang.Override
public java.util.List<? extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder>
getInputSpecsOrBuilderList() {
return inputSpecs_;
@@ -3998,35 +4295,40 @@ public final class LlapDaemonProtocolProtos {
/**
* <code>repeated .IOSpecProto input_specs = 10;</code>
*/
+ @java.lang.Override
public int getInputSpecsCount() {
return inputSpecs_.size();
}
/**
* <code>repeated .IOSpecProto input_specs = 10;</code>
*/
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getInputSpecs(int index) {
return inputSpecs_.get(index);
}
/**
* <code>repeated .IOSpecProto input_specs = 10;</code>
*/
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder getInputSpecsOrBuilder(
int index) {
return inputSpecs_.get(index);
}
- // repeated .IOSpecProto output_specs = 11;
public static final int OUTPUT_SPECS_FIELD_NUMBER = 11;
+ @SuppressWarnings("serial")
private java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> outputSpecs_;
/**
* <code>repeated .IOSpecProto output_specs = 11;</code>
*/
+ @java.lang.Override
public java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> getOutputSpecsList() {
return outputSpecs_;
}
/**
* <code>repeated .IOSpecProto output_specs = 11;</code>
*/
+ @java.lang.Override
public java.util.List<? extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder>
getOutputSpecsOrBuilderList() {
return outputSpecs_;
@@ -4034,35 +4336,40 @@ public final class LlapDaemonProtocolProtos {
/**
* <code>repeated .IOSpecProto output_specs = 11;</code>
*/
+ @java.lang.Override
public int getOutputSpecsCount() {
return outputSpecs_.size();
}
/**
* <code>repeated .IOSpecProto output_specs = 11;</code>
*/
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getOutputSpecs(int index) {
return outputSpecs_.get(index);
}
/**
* <code>repeated .IOSpecProto output_specs = 11;</code>
*/
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder getOutputSpecsOrBuilder(
int index) {
return outputSpecs_.get(index);
}
- // repeated .GroupInputSpecProto grouped_input_specs = 12;
public static final int GROUPED_INPUT_SPECS_FIELD_NUMBER = 12;
+ @SuppressWarnings("serial")
private java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto> groupedInputSpecs_;
/**
* <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code>
*/
+ @java.lang.Override
public java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto> getGroupedInputSpecsList() {
return groupedInputSpecs_;
}
/**
* <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code>
*/
+ @java.lang.Override
public java.util.List<? extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder>
getGroupedInputSpecsOrBuilderList() {
return groupedInputSpecs_;
@@ -4070,117 +4377,112 @@ public final class LlapDaemonProtocolProtos {
/**
* <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code>
*/
+ @java.lang.Override
public int getGroupedInputSpecsCount() {
return groupedInputSpecs_.size();
}
/**
* <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code>
*/
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto getGroupedInputSpecs(int index) {
return groupedInputSpecs_.get(index);
}
/**
* <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code>
*/
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder getGroupedInputSpecsOrBuilder(
int index) {
return groupedInputSpecs_.get(index);
}
- // optional int32 vertex_parallelism = 13;
public static final int VERTEX_PARALLELISM_FIELD_NUMBER = 13;
- private int vertexParallelism_;
+ private int vertexParallelism_ = 0;
/**
- * <code>optional int32 vertex_parallelism = 13;</code>
- *
* <pre>
* An internal field required for Tez.
* </pre>
+ *
+ * <code>optional int32 vertex_parallelism = 13;</code>
+ * @return Whether the vertexParallelism field is set.
*/
+ @java.lang.Override
public boolean hasVertexParallelism() {
- return ((bitField0_ & 0x00000200) == 0x00000200);
+ return ((bitField0_ & 0x00000200) != 0);
}
/**
- * <code>optional int32 vertex_parallelism = 13;</code>
- *
* <pre>
* An internal field required for Tez.
* </pre>
+ *
+ * <code>optional int32 vertex_parallelism = 13;</code>
+ * @return The vertexParallelism.
*/
+ @java.lang.Override
public int getVertexParallelism() {
return vertexParallelism_;
}
- // optional bool is_external_submission = 14 [default = false];
public static final int IS_EXTERNAL_SUBMISSION_FIELD_NUMBER = 14;
- private boolean isExternalSubmission_;
+ private boolean isExternalSubmission_ = false;
/**
* <code>optional bool is_external_submission = 14 [default = false];</code>
+ * @return Whether the isExternalSubmission field is set.
*/
+ @java.lang.Override
public boolean hasIsExternalSubmission() {
- return ((bitField0_ & 0x00000400) == 0x00000400);
+ return ((bitField0_ & 0x00000400) != 0);
}
/**
* <code>optional bool is_external_submission = 14 [default = false];</code>
+ * @return The isExternalSubmission.
*/
+ @java.lang.Override
public boolean getIsExternalSubmission() {
return isExternalSubmission_;
}
- private void initFields() {
- user_ = "";
- signatureKeyId_ = 0L;
- queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance();
- hiveQueryId_ = "";
- dagName_ = "";
- vertexName_ = "";
- vertexIndex_ = 0;
- tokenIdentifier_ = "";
- processorDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
- inputSpecs_ = java.util.Collections.emptyList();
- outputSpecs_ = java.util.Collections.emptyList();
- groupedInputSpecs_ = java.util.Collections.emptyList();
- vertexParallelism_ = 0;
- isExternalSubmission_ = false;
- }
private byte memoizedIsInitialized = -1;
+ @java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
- if (isInitialized != -1) return isInitialized == 1;
+ if (isInitialized == 1) return true;
+ if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
+ @java.lang.Override
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
- getSerializedSize();
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
- output.writeBytes(1, getUserBytes());
+ if (((bitField0_ & 0x00000001) != 0)) {
+ com.google.protobuf.GeneratedMessageV3.writeString(output, 1, user_);
}
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ if (((bitField0_ & 0x00000002) != 0)) {
output.writeInt64(2, signatureKeyId_);
}
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
- output.writeMessage(3, queryIdentifier_);
+ if (((bitField0_ & 0x00000004) != 0)) {
+ output.writeMessage(3, getQueryIdentifier());
}
- if (((bitField0_ & 0x00000008) == 0x00000008)) {
- output.writeBytes(4, getHiveQueryIdBytes());
+ if (((bitField0_ & 0x00000008) != 0)) {
+ com.google.protobuf.GeneratedMessageV3.writeString(output, 4, hiveQueryId_);
}
- if (((bitField0_ & 0x00000010) == 0x00000010)) {
- output.writeBytes(5, getDagNameBytes());
+ if (((bitField0_ & 0x00000010) != 0)) {
+ com.google.protobuf.GeneratedMessageV3.writeString(output, 5, dagName_);
}
- if (((bitField0_ & 0x00000020) == 0x00000020)) {
- output.writeBytes(6, getVertexNameBytes());
+ if (((bitField0_ & 0x00000020) != 0)) {
+ com.google.protobuf.GeneratedMessageV3.writeString(output, 6, vertexName_);
}
- if (((bitField0_ & 0x00000040) == 0x00000040)) {
+ if (((bitField0_ & 0x00000040) != 0)) {
output.writeInt32(7, vertexIndex_);
}
- if (((bitField0_ & 0x00000080) == 0x00000080)) {
- output.writeBytes(8, getTokenIdentifierBytes());
+ if (((bitField0_ & 0x00000080) != 0)) {
+ com.google.protobuf.GeneratedMessageV3.writeString(output, 8, tokenIdentifier_);
}
- if (((bitField0_ & 0x00000100) == 0x00000100)) {
- output.writeMessage(9, processorDescriptor_);
+ if (((bitField0_ & 0x00000100) != 0)) {
+ output.writeMessage(9, getProcessorDescriptor());
}
for (int i = 0; i < inputSpecs_.size(); i++) {
output.writeMessage(10, inputSpecs_.get(i));
@@ -4191,56 +4493,51 @@ public final class LlapDaemonProtocolProtos {
for (int i = 0; i < groupedInputSpecs_.size(); i++) {
output.writeMessage(12, groupedInputSpecs_.get(i));
}
- if (((bitField0_ & 0x00000200) == 0x00000200)) {
+ if (((bitField0_ & 0x00000200) != 0)) {
output.writeInt32(13, vertexParallelism_);
}
- if (((bitField0_ & 0x00000400) == 0x00000400)) {
+ if (((bitField0_ & 0x00000400) != 0)) {
output.writeBool(14, isExternalSubmission_);
}
getUnknownFields().writeTo(output);
}
- private int memoizedSerializedSize = -1;
+ @java.lang.Override
public int getSerializedSize() {
- int size = memoizedSerializedSize;
+ int size = memoizedSize;
if (size != -1) return size;
size = 0;
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
- .computeBytesSize(1, getUserBytes());
+ if (((bitField0_ & 0x00000001) != 0)) {
+ size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, user_);
}
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ if (((bitField0_ & 0x00000002) != 0)) {
size += com.google.protobuf.CodedOutputStream
.computeInt64Size(2, signatureKeyId_);
}
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
+ if (((bitField0_ & 0x00000004) != 0)) {
size += com.google.protobuf.CodedOutputStream
- .computeMessageSize(3, queryIdentifier_);
+ .computeMessageSize(3, getQueryIdentifier());
}
- if (((bitField0_ & 0x00000008) == 0x00000008)) {
- size += com.google.protobuf.CodedOutputStream
- .computeBytesSize(4, getHiveQueryIdBytes());
+ if (((bitField0_ & 0x00000008) != 0)) {
+ size += com.google.protobuf.GeneratedMessageV3.computeStringSize(4, hiveQueryId_);
}
- if (((bitField0_ & 0x00000010) == 0x00000010)) {
- size += com.google.protobuf.CodedOutputStream
- .computeBytesSize(5, getDagNameBytes());
+ if (((bitField0_ & 0x00000010) != 0)) {
+ size += com.google.protobuf.GeneratedMessageV3.computeStringSize(5, dagName_);
}
- if (((bitField0_ & 0x00000020) == 0x00000020)) {
- size += com.google.protobuf.CodedOutputStream
- .computeBytesSize(6, getVertexNameBytes());
+ if (((bitField0_ & 0x00000020) != 0)) {
+ size += com.google.protobuf.GeneratedMessageV3.computeStringSize(6, vertexName_);
}
- if (((bitField0_ & 0x00000040) == 0x00000040)) {
+ if (((bitField0_ & 0x00000040) != 0)) {
size += com.google.protobuf.CodedOutputStream
.computeInt32Size(7, vertexIndex_);
}
- if (((bitField0_ & 0x00000080) == 0x00000080)) {
- size += com.google.protobuf.CodedOutputStream
- .computeBytesSize(8, getTokenIdentifierBytes());
+ if (((bitField0_ & 0x00000080) != 0)) {
+ size += com.google.protobuf.GeneratedMessageV3.computeStringSize(8, tokenIdentifier_);
}
- if (((bitField0_ & 0x00000100) == 0x00000100)) {
+ if (((bitField0_ & 0x00000100) != 0)) {
size += com.google.protobuf.CodedOutputStream
- .computeMessageSize(9, processorDescriptor_);
+ .computeMessageSize(9, getProcessorDescriptor());
}
for (int i = 0; i < inputSpecs_.size(); i++) {
size += com.google.protobuf.CodedOutputStream
@@ -4254,26 +4551,19 @@ public final class LlapDaemonProtocolProtos {
size += com.google.protobuf.CodedOutputStream
.computeMessageSize(12, groupedInputSpecs_.get(i));
}
- if (((bitField0_ & 0x00000200) == 0x00000200)) {
+ if (((bitField0_ & 0x00000200) != 0)) {
size += com.google.protobuf.CodedOutputStream
.computeInt32Size(13, vertexParallelism_);
}
- if (((bitField0_ & 0x00000400) == 0x00000400)) {
+ if (((bitField0_ & 0x00000400) != 0)) {
size += com.google.protobuf.CodedOutputStream
.computeBoolSize(14, isExternalSubmission_);
}
size += getUnknownFields().getSerializedSize();
- memoizedSerializedSize = size;
+ memoizedSize = size;
return size;
}
- private static final long serialVersionUID = 0L;
- @java.lang.Override
- protected java.lang.Object writeReplace()
- throws java.io.ObjectStreamException {
- return super.writeReplace();
- }
-
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
@@ -4284,88 +4574,86 @@ public final class LlapDaemonProtocolProtos {
}
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec) obj;
- boolean result = true;
- result = result && (hasUser() == other.hasUser());
+ if (hasUser() != other.hasUser()) return false;
if (hasUser()) {
- result = result && getUser()
- .equals(other.getUser());
+ if (!getUser()
+ .equals(other.getUser())) return false;
}
- result = result && (hasSignatureKeyId() == other.hasSignatureKeyId());
+ if (hasSignatureKeyId() != other.hasSignatureKeyId()) return false;
if (hasSignatureKeyId()) {
- result = result && (getSignatureKeyId()
- == other.getSignatureKeyId());
+ if (getSignatureKeyId()
+ != other.getSignatureKeyId()) return false;
}
- result = result && (hasQueryIdentifier() == other.hasQueryIdentifier());
+ if (hasQueryIdentifier() != other.hasQueryIdentifier()) return false;
if (hasQueryIdentifier()) {
- result = result && getQueryIdentifier()
- .equals(other.getQueryIdentifier());
+ if (!getQueryIdentifier()
+ .equals(other.getQueryIdentifier())) return false;
}
- result = result && (hasHiveQueryId() == other.hasHiveQueryId());
+ if (hasHiveQueryId() != other.hasHiveQueryId()) return false;
if (hasHiveQueryId()) {
- result = result && getHiveQueryId()
- .equals(other.getHiveQueryId());
+ if (!getHiveQueryId()
+ .equals(other.getHiveQueryId())) return false;
}
- result = result && (hasDagName() == other.hasDagName());
+ if (hasDagName() != other.hasDagName()) return false;
if (hasDagName()) {
- result = result && getDagName()
- .equals(other.getDagName());
+ if (!getDagName()
+ .equals(other.getDagName())) return false;
}
- result = result && (hasVertexName() == other.hasVertexName());
+ if (hasVertexName() != other.hasVertexName()) return false;
if (hasVertexName()) {
- result = result && getVertexName()
- .equals(other.getVertexName());
+ if (!getVertexName()
+ .equals(other.getVertexName())) return false;
}
- result = result && (hasVertexIndex() == other.hasVertexIndex());
+ if (hasVertexIndex() != other.hasVertexIndex()) return false;
if (hasVertexIndex()) {
- result = result && (getVertexIndex()
- == other.getVertexIndex());
+ if (getVertexIndex()
+ != other.getVertexIndex()) return false;
}
- result = result && (hasTokenIdentifier() == other.hasTokenIdentifier());
+ if (hasTokenIdentifier() != other.hasTokenIdentifier()) return false;
if (hasTokenIdentifier()) {
- result = result && getTokenIdentifier()
- .equals(other.getTokenIdentifier());
+ if (!getTokenIdentifier()
+ .equals(other.getTokenIdentifier())) return false;
}
- result = result && (hasProcessorDescriptor() == other.hasProcessorDescriptor());
+ if (hasProcessorDescriptor() != other.hasProcessorDescriptor()) return false;
if (hasProcessorDescriptor()) {
- result = result && getProcessorDescriptor()
- .equals(other.getProcessorDescriptor());
- }
- result = result && getInputSpecsList()
- .equals(other.getInputSpecsList());
- result = result && getOutputSpecsList()
- .equals(other.getOutputSpecsList());
- result = result && getGroupedInputSpecsList()
- .equals(other.getGroupedInputSpecsList());
- result = result && (hasVertexParallelism() == other.hasVertexParallelism());
+ if (!getProcessorDescriptor()
+ .equals(other.getProcessorDescriptor())) return false;
+ }
+ if (!getInputSpecsList()
+ .equals(other.getInputSpecsList())) return false;
+ if (!getOutputSpecsList()
+ .equals(other.getOutputSpecsList())) return false;
+ if (!getGroupedInputSpecsList()
+ .equals(other.getGroupedInputSpecsList())) return false;
+ if (hasVertexParallelism() != other.hasVertexParallelism()) return false;
if (hasVertexParallelism()) {
- result = result && (getVertexParallelism()
- == other.getVertexParallelism());
+ if (getVertexParallelism()
+ != other.getVertexParallelism()) return false;
}
- result = result && (hasIsExternalSubmission() == other.hasIsExternalSubmission());
+ if (hasIsExternalSubmission() != other.hasIsExternalSubmission()) return false;
if (hasIsExternalSubmission()) {
- result = result && (getIsExternalSubmission()
- == other.getIsExternalSubmission());
+ if (getIsExternalSubmission()
+ != other.getIsExternalSubmission()) return false;
}
- result = result &&
- getUnknownFields().equals(other.getUnknownFields());
- return result;
+ if (!getUnknownFields().equals(other.getUnknownFields())) return false;
+ return true;
}
- private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
- hash = (19 * hash) + getDescriptorForType().hashCode();
+ hash = (19 * hash) + getDescriptor().hashCode();
if (hasUser()) {
hash = (37 * hash) + USER_FIELD_NUMBER;
hash = (53 * hash) + getUser().hashCode();
}
if (hasSignatureKeyId()) {
hash = (37 * hash) + SIGNATUREKEYID_FIELD_NUMBER;
- hash = (53 * hash) + hashLong(getSignatureKeyId());
+ hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ getSignatureKeyId());
}
if (hasQueryIdentifier()) {
hash = (37 * hash) + QUERY_IDENTIFIER_FIELD_NUMBER;
@@ -4413,13 +4701,25 @@ public final class LlapDaemonProtocolProtos {
}
if (hasIsExternalSubmission()) {
hash = (37 * hash) + IS_EXTERNAL_SUBMISSION_FIELD_NUMBER;
- hash = (53 * hash) + hashBoolean(getIsExternalSubmission());
+ hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean(
+ getIsExternalSubmission());
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom(
+ java.nio.ByteBuffer data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom(
+ java.nio.ByteBuffer data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
@@ -4443,65 +4743,82 @@ public final class LlapDaemonProtocolProtos {
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
+
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input);
}
+
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
- public static Builder newBuilder() { return Builder.create(); }
+ @java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder() {
+ return DEFAULT_INSTANCE.toBuilder();
+ }
public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec prototype) {
- return newBuilder().mergeFrom(prototype);
+ return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+ }
+ @java.lang.Override
+ public Builder toBuilder() {
+ return this == DEFAULT_INSTANCE
+ ? new Builder() : new Builder().mergeFrom(this);
}
- public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
- * Protobuf type {@code SignableVertexSpec}
- *
* <pre>
* The part of SubmitWork that can be signed
* </pre>
+ *
+ * Protobuf type {@code SignableVertexSpec}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder {
+ com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ // @@protoc_insertion_point(builder_implements:SignableVertexSpec)
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SignableVertexSpec_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ @java.lang.Override
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SignableVertexSpec_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -4514,12 +4831,13 @@ public final class LlapDaemonProtocolProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ if (com.google.protobuf.GeneratedMessageV3
+ .alwaysUseFieldBuilders) {
getQueryIdentifierFieldBuilder();
getProcessorDescriptorFieldBuilder();
getInputSpecsFieldBuilder();
@@ -4527,76 +4845,65 @@ public final class LlapDaemonProtocolProtos {
getGroupedInputSpecsFieldBuilder();
}
}
- private static Builder create() {
- return new Builder();
- }
-
+ @java.lang.Override
public Builder clear() {
super.clear();
+ bitField0_ = 0;
user_ = "";
- bitField0_ = (bitField0_ & ~0x00000001);
signatureKeyId_ = 0L;
- bitField0_ = (bitField0_ & ~0x00000002);
- if (queryIdentifierBuilder_ == null) {
- queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance();
- } else {
- queryIdentifierBuilder_.clear();
+ queryIdentifier_ = null;
+ if (queryIdentifierBuilder_ != null) {
+ queryIdentifierBuilder_.dispose();
+ queryIdentifierBuilder_ = null;
}
- bitField0_ = (bitField0_ & ~0x00000004);
hiveQueryId_ = "";
- bitField0_ = (bitField0_ & ~0x00000008);
dagName_ = "";
- bitField0_ = (bitField0_ & ~0x00000010);
vertexName_ = "";
- bitField0_ = (bitField0_ & ~0x00000020);
vertexIndex_ = 0;
- bitField0_ = (bitField0_ & ~0x00000040);
tokenIdentifier_ = "";
- bitField0_ = (bitField0_ & ~0x00000080);
- if (processorDescriptorBuilder_ == null) {
- processorDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
- } else {
- processorDescriptorBuilder_.clear();
+ processorDescriptor_ = null;
+ if (processorDescriptorBuilder_ != null) {
+ processorDescriptorBuilder_.dispose();
+ processorDescriptorBuilder_ = null;
}
- bitField0_ = (bitField0_ & ~0x00000100);
if (inputSpecsBuilder_ == null) {
inputSpecs_ = java.util.Collections.emptyList();
- bitField0_ = (bitField0_ & ~0x00000200);
} else {
+ inputSpecs_ = null;
inputSpecsBuilder_.clear();
}
+ bitField0_ = (bitField0_ & ~0x00000200);
if (outputSpecsBuilder_ == null) {
outputSpecs_ = java.util.Collections.emptyList();
- bitField0_ = (bitField0_ & ~0x00000400);
} else {
+ outputSpecs_ = null;
outputSpecsBuilder_.clear();
}
+ bitField0_ = (bitField0_ & ~0x00000400);
if (groupedInputSpecsBuilder_ == null) {
groupedInputSpecs_ = java.util.Collections.emptyList();
- bitField0_ = (bitField0_ & ~0x00000800);
} else {
+ groupedInputSpecs_ = null;
groupedInputSpecsBuilder_.clear();
}
+ bitField0_ = (bitField0_ & ~0x00000800);
vertexParallelism_ = 0;
- bitField0_ = (bitField0_ & ~0x00001000);
isExternalSubmission_ = false;
- bitField0_ = (bitField0_ & ~0x00002000);
return this;
}
- public Builder clone() {
- return create().mergeFrom(buildPartial());
- }
-
+ @java.lang.Override
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SignableVertexSpec_descriptor;
}
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec getDefaultInstanceForType() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance();
}
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec build() {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec result = buildPartial();
if (!result.isInitialized()) {
@@ -4605,56 +4912,18 @@ public final class LlapDaemonProtocolProtos {
return result;
}
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec buildPartial() {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec(this);
- int from_bitField0_ = bitField0_;
- int to_bitField0_ = 0;
- if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
- to_bitField0_ |= 0x00000001;
- }
- result.user_ = user_;
- if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
- to_bitField0_ |= 0x00000002;
- }
- result.signatureKeyId_ = signatureKeyId_;
- if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
- to_bitField0_ |= 0x00000004;
- }
- if (queryIdentifierBuilder_ == null) {
- result.queryIdentifier_ = queryIdentifier_;
- } else {
- result.queryIdentifier_ = queryIdentifierBuilder_.build();
- }
- if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
- to_bitField0_ |= 0x00000008;
- }
- result.hiveQueryId_ = hiveQueryId_;
- if (((from_bitField0_ & 0x00000010) == 0x00000010)) {
- to_bitField0_ |= 0x00000010;
- }
- result.dagName_ = dagName_;
- if (((from_bitField0_ & 0x00000020) == 0x00000020)) {
- to_bitField0_ |= 0x00000020;
- }
- result.vertexName_ = vertexName_;
- if (((from_bitField0_ & 0x00000040) == 0x00000040)) {
- to_bitField0_ |= 0x00000040;
- }
- result.vertexIndex_ = vertexIndex_;
- if (((from_bitField0_ & 0x00000080) == 0x00000080)) {
- to_bitField0_ |= 0x00000080;
- }
- result.tokenIdentifier_ = tokenIdentifier_;
- if (((from_bitField0_ & 0x00000100) == 0x00000100)) {
- to_bitField0_ |= 0x00000100;
- }
- if (processorDescriptorBuilder_ == null) {
- result.processorDescriptor_ = processorDescriptor_;
- } else {
- result.processorDescriptor_ = processorDescriptorBuilder_.build();
- }
+ buildPartialRepeatedFields(result);
+ if (bitField0_ != 0) { buildPartial0(result); }
+ onBuilt();
+ return result;
+ }
+
+ private void buildPartialRepeatedFields(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec result) {
if (inputSpecsBuilder_ == null) {
- if (((bitField0_ & 0x00000200) == 0x00000200)) {
+ if (((bitField0_ & 0x00000200) != 0)) {
inputSpecs_ = java.util.Collections.unmodifiableList(inputSpecs_);
bitField0_ = (bitField0_ & ~0x00000200);
}
@@ -4663,7 +4932,7 @@ public final class LlapDaemonProtocolProtos {
result.inputSpecs_ = inputSpecsBuilder_.build();
}
if (outputSpecsBuilder_ == null) {
- if (((bitField0_ & 0x00000400) == 0x00000400)) {
+ if (((bitField0_ & 0x00000400) != 0)) {
outputSpecs_ = java.util.Collections.unmodifiableList(outputSpecs_);
bitField0_ = (bitField0_ & ~0x00000400);
}
@@ -4672,7 +4941,7 @@ public final class LlapDaemonProtocolProtos {
result.outputSpecs_ = outputSpecsBuilder_.build();
}
if (groupedInputSpecsBuilder_ == null) {
- if (((bitField0_ & 0x00000800) == 0x00000800)) {
+ if (((bitField0_ & 0x00000800) != 0)) {
groupedInputSpecs_ = java.util.Collections.unmodifiableList(groupedInputSpecs_);
bitField0_ = (bitField0_ & ~0x00000800);
}
@@ -4680,19 +4949,95 @@ public final class LlapDaemonProtocolProtos {
} else {
result.groupedInputSpecs_ = groupedInputSpecsBuilder_.build();
}
- if (((from_bitField0_ & 0x00001000) == 0x00001000)) {
+ }
+
+ private void buildPartial0(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec result) {
+ int from_bitField0_ = bitField0_;
+ int to_bitField0_ = 0;
+ if (((from_bitField0_ & 0x00000001) != 0)) {
+ result.user_ = user_;
+ to_bitField0_ |= 0x00000001;
+ }
+ if (((from_bitField0_ & 0x00000002) != 0)) {
+ result.signatureKeyId_ = signatureKeyId_;
+ to_bitField0_ |= 0x00000002;
+ }
+ if (((from_bitField0_ & 0x00000004) != 0)) {
+ result.queryIdentifier_ = queryIdentifierBuilder_ == null
+ ? queryIdentifier_
+ : queryIdentifierBuilder_.build();
+ to_bitField0_ |= 0x00000004;
+ }
+ if (((from_bitField0_ & 0x00000008) != 0)) {
+ result.hiveQueryId_ = hiveQueryId_;
+ to_bitField0_ |= 0x00000008;
+ }
+ if (((from_bitField0_ & 0x00000010) != 0)) {
+ result.dagName_ = dagName_;
+ to_bitField0_ |= 0x00000010;
+ }
+ if (((from_bitField0_ & 0x00000020) != 0)) {
+ result.vertexName_ = vertexName_;
+ to_bitField0_ |= 0x00000020;
+ }
+ if (((from_bitField0_ & 0x00000040) != 0)) {
+ result.vertexIndex_ = vertexIndex_;
+ to_bitField0_ |= 0x00000040;
+ }
+ if (((from_bitField0_ & 0x00000080) != 0)) {
+ result.tokenIdentifier_ = tokenIdentifier_;
+ to_bitField0_ |= 0x00000080;
+ }
+ if (((from_bitField0_ & 0x00000100) != 0)) {
+ result.processorDescriptor_ = processorDescriptorBuilder_ == null
+ ? processorDescriptor_
+ : processorDescriptorBuilder_.build();
+ to_bitField0_ |= 0x00000100;
+ }
+ if (((from_bitField0_ & 0x00001000) != 0)) {
+ result.vertexParallelism_ = vertexParallelism_;
to_bitField0_ |= 0x00000200;
}
- result.vertexParallelism_ = vertexParallelism_;
- if (((from_bitField0_ & 0x00002000) == 0x00002000)) {
+ if (((from_bitField0_ & 0x00002000) != 0)) {
+ result.isExternalSubmission_ = isExternalSubmission_;
to_bitField0_ |= 0x00000400;
}
- result.isExternalSubmission_ = isExternalSubmission_;
- result.bitField0_ = to_bitField0_;
- onBuilt();
- return result;
+ result.bitField0_ |= to_bitField0_;
}
+ @java.lang.Override
+ public Builder clone() {
+ return super.clone();
+ }
+ @java.lang.Override
+ public Builder setField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ java.lang.Object value) {
+ return super.setField(field, value);
+ }
+ @java.lang.Override
+ public Builder clearField(
+ com.google.protobuf.Descriptors.FieldDescriptor field) {
+ return super.clearField(field);
+ }
+ @java.lang.Override
+ public Builder clearOneof(
+ com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ return super.clearOneof(oneof);
+ }
+ @java.lang.Override
+ public Builder setRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ int index, java.lang.Object value) {
+ return super.setRepeatedField(field, index, value);
+ }
+ @java.lang.Override
+ public Builder addRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ java.lang.Object value) {
+ return super.addRepeatedField(field, value);
+ }
+ @java.lang.Override
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec) {
return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec)other);
@@ -4705,8 +5050,8 @@ public final class LlapDaemonProtocolProtos {
public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec other) {
if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance()) return this;
if (other.hasUser()) {
- bitField0_ |= 0x00000001;
user_ = other.user_;
+ bitField0_ |= 0x00000001;
onChanged();
}
if (other.hasSignatureKeyId()) {
@@ -4716,26 +5061,26 @@ public final class LlapDaemonProtocolProtos {
mergeQueryIdentifier(other.getQueryIdentifier());
}
if (other.hasHiveQueryId()) {
- bitField0_ |= 0x00000008;
hiveQueryId_ = other.hiveQueryId_;
+ bitField0_ |= 0x00000008;
onChanged();
}
if (other.hasDagName()) {
- bitField0_ |= 0x00000010;
dagName_ = other.dagName_;
+ bitField0_ |= 0x00000010;
onChanged();
}
if (other.hasVertexName()) {
- bitField0_ |= 0x00000020;
vertexName_ = other.vertexName_;
+ bitField0_ |= 0x00000020;
onChanged();
}
if (other.hasVertexIndex()) {
setVertexIndex(other.getVertexIndex());
}
if (other.hasTokenIdentifier()) {
- bitField0_ |= 0x00000080;
tokenIdentifier_ = other.tokenIdentifier_;
+ bitField0_ |= 0x00000080;
onChanged();
}
if (other.hasProcessorDescriptor()) {
@@ -4760,7 +5105,7 @@ public final class LlapDaemonProtocolProtos {
inputSpecs_ = other.inputSpecs_;
bitField0_ = (bitField0_ & ~0x00000200);
inputSpecsBuilder_ =
- com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ?
+ com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getInputSpecsFieldBuilder() : null;
} else {
inputSpecsBuilder_.addAllMessages(other.inputSpecs_);
@@ -4786,7 +5131,7 @@ public final class LlapDaemonProtocolProtos {
outputSpecs_ = other.outputSpecs_;
bitField0_ = (bitField0_ & ~0x00000400);
outputSpecsBuilder_ =
- com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ?
+ com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getOutputSpecsFieldBuilder() : null;
} else {
outputSpecsBuilder_.addAllMessages(other.outputSpecs_);
@@ -4812,7 +5157,7 @@ public final class LlapDaemonProtocolProtos {
groupedInputSpecs_ = other.groupedInputSpecs_;
bitField0_ = (bitField0_ & ~0x00000800);
groupedInputSpecsBuilder_ =
- com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ?
+ com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getGroupedInputSpecsFieldBuilder() : null;
} else {
groupedInputSpecsBuilder_.addAllMessages(other.groupedInputSpecs_);
@@ -4826,49 +5171,167 @@ public final class LlapDaemonProtocolProtos {
setIsExternalSubmission(other.getIsExternalSubmission());
}
this.mergeUnknownFields(other.getUnknownFields());
+ onChanged();
return this;
}
+ @java.lang.Override
public final boolean isInitialized() {
return true;
}
+ @java.lang.Override
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parsedMessage = null;
+ if (extensionRegistry == null) {
+ throw new java.lang.NullPointerException();
+ }
try {
- parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ case 10: {
+ user_ = input.readBytes();
+ bitField0_ |= 0x00000001;
+ break;
+ } // case 10
+ case 16: {
+ signatureKeyId_ = input.readInt64();
+ bitField0_ |= 0x00000002;
+ break;
+ } // case 16
+ case 26: {
+ input.readMessage(
+ getQueryIdentifierFieldBuilder().getBuilder(),
+ extensionRegistry);
+ bitField0_ |= 0x00000004;
+ break;
+ } // case 26
+ case 34: {
+ hiveQueryId_ = input.readBytes();
+ bitField0_ |= 0x00000008;
+ break;
+ } // case 34
+ case 42: {
+ dagName_ = input.readBytes();
+ bitField0_ |= 0x00000010;
+ break;
+ } // case 42
+ case 50: {
+ vertexName_ = input.readBytes();
+ bitField0_ |= 0x00000020;
+ break;
+ } // case 50
+ case 56: {
+ vertexIndex_ = input.readInt32();
+ bitField0_ |= 0x00000040;
+ break;
+ } // case 56
+ case 66: {
+ tokenIdentifier_ = input.readBytes();
+ bitField0_ |= 0x00000080;
+ break;
+ } // case 66
+ case 74: {
+ input.readMessage(
+ getProcessorDescriptorFieldBuilder().getBuilder(),
+ extensionRegistry);
+ bitField0_ |= 0x00000100;
+ break;
+ } // case 74
+ case 82: {
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto m =
+ input.readMessage(
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.PARSER,
+ extensionRegistry);
+ if (inputSpecsBuilder_ == null) {
+ ensureInputSpecsIsMutable();
+ inputSpecs_.add(m);
+ } else {
+ inputSpecsBuilder_.addMessage(m);
+ }
+ break;
+ } // case 82
+ case 90: {
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto m =
+ input.readMessage(
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.PARSER,
+ extensionRegistry);
+ if (outputSpecsBuilder_ == null) {
+ ensureOutputSpecsIsMutable();
+ outputSpecs_.add(m);
+ } else {
+ outputSpecsBuilder_.addMessage(m);
+ }
+ break;
+ } // case 90
+ case 98: {
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto m =
+ input.readMessage(
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.PARSER,
+ extensionRegistry);
+ if (groupedInputSpecsBuilder_ == null) {
+ ensureGroupedInputSpecsIsMutable();
+ groupedInputSpecs_.add(m);
+ } else {
+ groupedInputSpecsBuilder_.addMessage(m);
+ }
+ break;
+ } // case 98
+ case 104: {
+ vertexParallelism_ = input.readInt32();
+ bitField0_ |= 0x00001000;
+ break;
+ } // case 104
+ case 112: {
+ isExternalSubmission_ = input.readBool();
+ bitField0_ |= 0x00002000;
+ break;
+ } // case 112
+ default: {
+ if (!super.parseUnknownField(input, extensionRegistry, tag)) {
+ done = true; // was an endgroup tag
+ }
+ break;
+ } // default:
+ } // switch (tag)
+ } // while (!done)
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
- parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec) e.getUnfinishedMessage();
- throw e;
+ throw e.unwrapIOException();
} finally {
- if (parsedMessage != null) {
- mergeFrom(parsedMessage);
- }
- }
+ onChanged();
+ } // finally
return this;
}
private int bitField0_;
- // optional string user = 1;
private java.lang.Object user_ = "";
/**
* <code>optional string user = 1;</code>
+ * @return Whether the user field is set.
*/
public boolean hasUser() {
- return ((bitField0_ & 0x00000001) == 0x00000001);
+ return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string user = 1;</code>
+ * @return The user.
*/
public java.lang.String getUser() {
java.lang.Object ref = user_;
if (!(ref instanceof java.lang.String)) {
- java.lang.String s = ((com.google.protobuf.ByteString) ref)
- .toStringUtf8();
- user_ = s;
+ com.google.protobuf.ByteString bs =
+ (com.google.protobuf.ByteString) ref;
+ java.lang.String s = bs.toStringUtf8();
+ if (bs.isValidUtf8()) {
+ user_ = s;
+ }
return s;
} else {
return (java.lang.String) ref;
@@ -4876,6 +5339,7 @@ public final class LlapDaemonProtocolProtos {
}
/**
* <code>optional string user = 1;</code>
+ * @return The bytes for user.
*/
public com.google.protobuf.ByteString
getUserBytes() {
@@ -4892,65 +5356,73 @@ public final class LlapDaemonProtocolProtos {
}
/**
* <code>optional string user = 1;</code>
+ * @param value The user to set.
+ * @return This builder for chaining.
*/
public Builder setUser(
java.lang.String value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000001;
+ if (value == null) { throw new NullPointerException(); }
user_ = value;
+ bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional string user = 1;</code>
+ * @return This builder for chaining.
*/
public Builder clearUser() {
- bitField0_ = (bitField0_ & ~0x00000001);
user_ = getDefaultInstance().getUser();
+ bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
return this;
}
/**
* <code>optional string user = 1;</code>
+ * @param value The bytes for user to set.
+ * @return This builder for chaining.
*/
public Builder setUserBytes(
com.google.protobuf.ByteString value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000001;
+ if (value == null) { throw new NullPointerException(); }
user_ = value;
+ bitField0_ |= 0x00000001;
onChanged();
return this;
}
- // optional int64 signatureKeyId = 2;
private long signatureKeyId_ ;
/**
* <code>optional int64 signatureKeyId = 2;</code>
+ * @return Whether the signatureKeyId field is set.
*/
+ @java.lang.Override
public boolean hasSignatureKeyId() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
+ return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional int64 signatureKeyId = 2;</code>
+ * @return The signatureKeyId.
*/
+ @java.lang.Override
public long getSignatureKeyId() {
return signatureKeyId_;
}
/**
* <code>optional int64 signatureKeyId = 2;</code>
+ * @param value The signatureKeyId to set.
+ * @return This builder for chaining.
*/
public Builder setSignatureKeyId(long value) {
- bitField0_ |= 0x00000002;
+
signatureKeyId_ = value;
+ bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* <code>optional int64 signatureKeyId = 2;</code>
+ * @return This builder for chaining.
*/
public Builder clearSignatureKeyId() {
bitField0_ = (bitField0_ & ~0x00000002);
@@ -4959,22 +5431,23 @@ public final class LlapDaemonProtocolProtos {
return this;
}
- // optional .QueryIdentifierProto query_identifier = 3;
- private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance();
- private com.google.protobuf.SingleFieldBuilder<
+ private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_;
+ private com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> queryIdentifierBuilder_;
/**
* <code>optional .QueryIdentifierProto query_identifier = 3;</code>
+ * @return Whether the queryIdentifier field is set.
*/
public boolean hasQueryIdentifier() {
- return ((bitField0_ & 0x00000004) == 0x00000004);
+ return ((bitField0_ & 0x00000004) != 0);
}
/**
* <code>optional .QueryIdentifierProto query_identifier = 3;</code>
+ * @return The queryIdentifier.
*/
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() {
if (queryIdentifierBuilder_ == null) {
- return queryIdentifier_;
+ return queryIdentifier_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_;
} else {
return queryIdentifierBuilder_.getMessage();
}
@@ -4988,11 +5461,11 @@ public final class LlapDaemonProtocolProtos {
throw new NullPointerException();
}
queryIdentifier_ = value;
- onChanged();
} else {
queryIdentifierBuilder_.setMessage(value);
}
bitField0_ |= 0x00000004;
+ onChanged();
return this;
}
/**
@@ -5002,11 +5475,11 @@ public final class LlapDaemonProtocolProtos {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder builderForValue) {
if (queryIdentifierBuilder_ == null) {
queryIdentifier_ = builderForValue.build();
- onChanged();
} else {
queryIdentifierBuilder_.setMessage(builderForValue.build());
}
bitField0_ |= 0x00000004;
+ onChanged();
return this;
}
/**
@@ -5014,31 +5487,33 @@ public final class LlapDaemonProtocolProtos {
*/
public Builder mergeQueryIdentifier(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto value) {
if (queryIdentifierBuilder_ == null) {
- if (((bitField0_ & 0x00000004) == 0x00000004) &&
- queryIdentifier_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance()) {
- queryIdentifier_ =
- org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.newBuilder(queryIdentifier_).mergeFrom(value).buildPartial();
+ if (((bitField0_ & 0x00000004) != 0) &&
+ queryIdentifier_ != null &&
+ queryIdentifier_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance()) {
+ getQueryIdentifierBuilder().mergeFrom(value);
} else {
queryIdentifier_ = value;
}
- onChanged();
} else {
queryIdentifierBuilder_.mergeFrom(value);
}
- bitField0_ |= 0x00000004;
+ if (queryIdentifier_ != null) {
+ bitField0_ |= 0x00000004;
+ onChanged();
+ }
return this;
}
/**
* <code>optional .QueryIdentifierProto query_identifier = 3;</code>
*/
public Builder clearQueryIdentifier() {
- if (queryIdentifierBuilder_ == null) {
- queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance();
- onChanged();
- } else {
- queryIdentifierBuilder_.clear();
- }
bitField0_ = (bitField0_ & ~0x00000004);
+ queryIdentifier_ = null;
+ if (queryIdentifierBuilder_ != null) {
+ queryIdentifierBuilder_.dispose();
+ queryIdentifierBuilder_ = null;
+ }
+ onChanged();
return this;
}
/**
@@ -5056,19 +5531,20 @@ public final class LlapDaemonProtocolProtos {
if (queryIdentifierBuilder_ != null) {
return queryIdentifierBuilder_.getMessageOrBuilder();
} else {
- return queryIdentifier_;
+ return queryIdentifier_ == null ?
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_;
}
}
/**
* <code>optional .QueryIdentifierProto query_identifier = 3;</code>
*/
- private com.google.protobuf.SingleFieldBuilder<
+ private com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder>
getQueryIdentifierFieldBuilder() {
if (queryIdentifierBuilder_ == null) {
- queryIdentifierBuilder_ = new com.google.protobuf.SingleFieldBuilder<
+ queryIdentifierBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder>(
- queryIdentifier_,
+ getQueryIdentifier(),
getParentForChildren(),
isClean());
queryIdentifier_ = null;
@@ -5076,23 +5552,27 @@ public final class LlapDaemonProtocolProtos {
return queryIdentifierBuilder_;
}
- // optional string hive_query_id = 4;
private java.lang.Object hiveQueryId_ = "";
/**
* <code>optional string hive_query_id = 4;</code>
+ * @return Whether the hiveQueryId field is set.
*/
public boolean hasHiveQueryId() {
- return ((bitField0_ & 0x00000008) == 0x00000008);
+ return ((bitField0_ & 0x00000008) != 0);
}
/**
* <code>optional string hive_query_id = 4;</code>
+ * @return The hiveQueryId.
*/
public java.lang.String getHiveQueryId() {
java.lang.Object ref = hiveQueryId_;
if (!(ref instanceof java.lang.String)) {
- java.lang.String s = ((com.google.protobuf.ByteString) ref)
- .toStringUtf8();
- hiveQueryId_ = s;
+ com.google.protobuf.ByteString bs =
+ (com.google.protobuf.ByteString) ref;
+ java.lang.String s = bs.toStringUtf8();
+ if (bs.isValidUtf8()) {
+ hiveQueryId_ = s;
+ }
return s;
} else {
return (java.lang.String) ref;
@@ -5100,6 +5580,7 @@ public final class LlapDaemonProtocolProtos {
}
/**
* <code>optional string hive_query_id = 4;</code>
+ * @return The bytes for hiveQueryId.
*/
public com.google.protobuf.ByteString
getHiveQueryIdBytes() {
@@ -5116,76 +5597,82 @@ public final class LlapDaemonProtocolProtos {
}
/**
* <code>optional string hive_query_id = 4;</code>
+ * @param value The hiveQueryId to set.
+ * @return This builder for chaining.
*/
public Builder setHiveQueryId(
java.lang.String value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000008;
+ if (value == null) { throw new NullPointerException(); }
hiveQueryId_ = value;
+ bitField0_ |= 0x00000008;
onChanged();
return this;
}
/**
* <code>optional string hive_query_id = 4;</code>
+ * @return This builder for chaining.
*/
public Builder clearHiveQueryId() {
- bitField0_ = (bitField0_ & ~0x00000008);
hiveQueryId_ = getDefaultInstance().getHiveQueryId();
+ bitField0_ = (bitField0_ & ~0x00000008);
onChanged();
return this;
}
/**
* <code>optional string hive_query_id = 4;</code>
+ * @param value The bytes for hiveQueryId to set.
+ * @return This builder for chaining.
*/
public Builder setHiveQueryIdBytes(
com.google.protobuf.ByteString value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000008;
+ if (value == null) { throw new NullPointerException(); }
hiveQueryId_ = value;
+ bitField0_ |= 0x00000008;
onChanged();
return this;
}
- // optional string dag_name = 5;
private java.lang.Object dagName_ = "";
/**
- * <code>optional string dag_name = 5;</code>
- *
* <pre>
* Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
* </pre>
+ *
+ * <code>optional string dag_name = 5;</code>
+ * @return Whether the dagName field is set.
*/
public boolean hasDagName() {
- return ((bitField0_ & 0x00000010) == 0x00000010);
+ return ((bitField0_ & 0x00000010) != 0);
}
/**
- * <code>optional string dag_name = 5;</code>
- *
* <pre>
* Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
* </pre>
+ *
+ * <code>optional string dag_name = 5;</code>
+ * @return The dagName.
*/
public java.lang.String getDagName() {
java.lang.Object ref = dagName_;
if (!(ref instanceof java.lang.String)) {
- java.lang.String s = ((com.google.protobuf.ByteString) ref)
- .toStringUtf8();
- dagName_ = s;
+ com.google.protobuf.ByteString bs =
+ (com.google.protobuf.ByteString) ref;
+ java.lang.String s = bs.toStringUtf8();
+ if (bs.isValidUtf8()) {
+ dagName_ = s;
+ }
return s;
} else {
return (java.lang.String) ref;
}
}
/**
- * <code>optional string dag_name = 5;</code>
- *
* <pre>
* Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
* </pre>
+ *
+ * <code>optional string dag_name = 5;</code>
+ * @return The bytes for dagName.
*/
public com.google.protobuf.ByteString
getDagNameBytes() {
@@ -5201,70 +5688,75 @@ public final class LlapDaemonProtocolProtos {
}
}
/**
- * <code>optional string dag_name = 5;</code>
- *
* <pre>
* Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
* </pre>
+ *
+ * <code>optional string dag_name = 5;</code>
+ * @param value The dagName to set.
+ * @return This builder for chaining.
*/
public Builder setDagName(
java.lang.String value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000010;
+ if (value == null) { throw new NullPointerException(); }
dagName_ = value;
+ bitField0_ |= 0x00000010;
onChanged();
return this;
}
/**
- * <code>optional string dag_name = 5;</code>
- *
* <pre>
* Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
* </pre>
+ *
+ * <code>optional string dag_name = 5;</code>
+ * @return This builder for chaining.
*/
public Builder clearDagName() {
- bitField0_ = (bitField0_ & ~0x00000010);
dagName_ = getDefaultInstance().getDagName();
+ bitField0_ = (bitField0_ & ~0x00000010);
onChanged();
return this;
}
/**
- * <code>optional string dag_name = 5;</code>
- *
* <pre>
* Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
* </pre>
+ *
+ * <code>optional string dag_name = 5;</code>
+ * @param value The bytes for dagName to set.
+ * @return This builder for chaining.
*/
public Builder setDagNameBytes(
com.google.protobuf.ByteString value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000010;
+ if (value == null) { throw new NullPointerException(); }
dagName_ = value;
+ bitField0_ |= 0x00000010;
onChanged();
return this;
}
- // optional string vertex_name = 6;
private java.lang.Object vertexName_ = "";
/**
* <code>optional string vertex_name = 6;</code>
+ * @return Whether the vertexName field is set.
*/
public boolean hasVertexName() {
- return ((bitField0_ & 0x00000020) == 0x00000020);
+ return ((bitField0_ & 0x00000020) != 0);
}
/**
* <code>optional string vertex_name = 6;</code>
+ * @return The vertexName.
*/
public java.lang.String getVertexName() {
java.lang.Object ref = vertexName_;
if (!(ref instanceof java.lang.String)) {
- java.lang.String s = ((com.google.protobuf.ByteString) ref)
- .toStringUtf8();
- vertexName_ = s;
+ com.google.protobuf.ByteString bs =
+ (com.google.protobuf.ByteString) ref;
+ java.lang.String s = bs.toStringUtf8();
+ if (bs.isValidUtf8()) {
+ vertexName_ = s;
+ }
return s;
} else {
return (java.lang.String) ref;
@@ -5272,6 +5764,7 @@ public final class LlapDaemonProtocolProtos {
}
/**
* <code>optional string vertex_name = 6;</code>
+ * @return The bytes for vertexName.
*/
public com.google.protobuf.ByteString
getVertexNameBytes() {
@@ -5288,65 +5781,73 @@ public final class LlapDaemonProtocolProtos {
}
/**
* <code>optional string vertex_name = 6;</code>
+ * @param value The vertexName to set.
+ * @return This builder for chaining.
*/
public Builder setVertexName(
java.lang.String value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000020;
+ if (value == null) { throw new NullPointerException(); }
vertexName_ = value;
+ bitField0_ |= 0x00000020;
onChanged();
return this;
}
/**
* <code>optional string vertex_name = 6;</code>
+ * @return This builder for chaining.
*/
public Builder clearVertexName() {
- bitField0_ = (bitField0_ & ~0x00000020);
vertexName_ = getDefaultInstance().getVertexName();
+ bitField0_ = (bitField0_ & ~0x00000020);
onChanged();
return this;
}
/**
* <code>optional string vertex_name = 6;</code>
+ * @param value The bytes for vertexName to set.
+ * @return This builder for chaining.
*/
public Builder setVertexNameBytes(
com.google.protobuf.ByteString value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000020;
+ if (value == null) { throw new NullPointerException(); }
vertexName_ = value;
+ bitField0_ |= 0x00000020;
onChanged();
return this;
}
- // optional int32 vertex_index = 7;
private int vertexIndex_ ;
/**
* <code>optional int32 vertex_index = 7;</code>
+ * @return Whether the vertexIndex field is set.
*/
+ @java.lang.Override
public boolean hasVertexIndex() {
- return ((bitField0_ & 0x00000040) == 0x00000040);
+ return ((bitField0_ & 0x00000040) != 0);
}
/**
* <code>optional int32 vertex_index = 7;</code>
+ * @return The vertexIndex.
*/
+ @java.lang.Override
public int getVertexIndex() {
return vertexIndex_;
}
/**
* <code>optional int32 vertex_index = 7;</code>
+ * @param value The vertexIndex to set.
+ * @return This builder for chaining.
*/
public Builder setVertexIndex(int value) {
- bitField0_ |= 0x00000040;
+
vertexIndex_ = value;
+ bitField0_ |= 0x00000040;
onChanged();
return this;
}
/**
* <code>optional int32 vertex_index = 7;</code>
+ * @return This builder for chaining.
*/
public Builder clearVertexIndex() {
bitField0_ = (bitField0_ & ~0x00000040);
@@ -5355,42 +5856,47 @@ public final class LlapDaemonProtocolProtos {
return this;
}
- // optional string token_identifier = 8;
private java.lang.Object tokenIdentifier_ = "";
/**
- * <code>optional string token_identifier = 8;</code>
- *
* <pre>
* The core vertex stuff
* </pre>
+ *
+ * <code>optional string token_identifier = 8;</code>
+ * @return Whether the tokenIdentifier field is set.
*/
public boolean hasTokenIdentifier() {
- return ((bitField0_ & 0x00000080) == 0x00000080);
+ return ((bitField0_ & 0x00000080) != 0);
}
/**
- * <code>optional string token_identifier = 8;</code>
- *
* <pre>
* The core vertex stuff
* </pre>
+ *
+ * <code>optional string token_identifier = 8;</code>
+ * @return The tokenIdentifier.
*/
public java.lang.String getTokenIdentifier() {
java.lang.Object ref = tokenIdentifier_;
if (!(ref instanceof java.lang.String)) {
- java.lang.String s = ((com.google.protobuf.ByteString) ref)
- .toStringUtf8();
- tokenIdentifier_ = s;
+ com.google.protobuf.ByteString bs =
+ (com.google.protobuf.ByteString) ref;
+ java.lang.String s = bs.toStringUtf8();
+ if (bs.isValidUtf8()) {
+ tokenIdentifier_ = s;
+ }
return s;
} else {
return (java.lang.String) ref;
}
}
/**
- * <code>optional string token_identifier = 8;</code>
- *
* <pre>
* The core vertex stuff
* </pre>
+ *
+ * <code>optional string token_identifier = 8;</code>
+ * @return The bytes for tokenIdentifier.
*/
public com.google.protobuf.ByteString
getTokenIdentifierBytes() {
@@ -5406,69 +5912,71 @@ public final class LlapDaemonProtocolProtos {
}
}
/**
- * <code>optional string token_identifier = 8;</code>
- *
* <pre>
* The core vertex stuff
* </pre>
+ *
+ * <code>optional string token_identifier = 8;</code>
+ * @param value The tokenIdentifier to set.
+ * @return This builder for chaining.
*/
public Builder setTokenIdentifier(
java.lang.String value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000080;
+ if (value == null) { throw new NullPointerException(); }
tokenIdentifier_ = value;
+ bitField0_ |= 0x00000080;
onChanged();
return this;
}
/**
- * <code>optional string token_identifier = 8;</code>
- *
* <pre>
* The core vertex stuff
* </pre>
+ *
+ * <code>optional string token_identifier = 8;</code>
+ * @return This builder for chaining.
*/
public Builder clearTokenIdentifier() {
- bitField0_ = (bitField0_ & ~0x00000080);
tokenIdentifier_ = getDefaultInstance().getTokenIdentifier();
+ bitField0_ = (bitField0_ & ~0x00000080);
onChanged();
return this;
}
/**
- * <code>optional string token_identifier = 8;</code>
- *
* <pre>
* The core vertex stuff
* </pre>
+ *
+ * <code>optional string token_identifier = 8;</code>
+ * @param value The bytes for tokenIdentifier to set.
+ * @return This builder for chaining.
*/
public Builder setTokenIdentifierBytes(
com.google.protobuf.ByteString value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000080;
+ if (value == null) { throw new NullPointerException(); }
tokenIdentifier_ = value;
+ bitField0_ |= 0x00000080;
onChanged();
return this;
}
- // optional .EntityDescriptorProto processor_descriptor = 9;
- private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto processorDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
- private com.google.protobuf.SingleFieldBuilder<
+ private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto processorDescriptor_;
+ private com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> processorDescriptorBuilder_;
/**
* <code>optional .EntityDescriptorProto processor_descriptor = 9;</code>
+ * @return Whether the processorDescriptor field is set.
*/
public boolean hasProcessorDescriptor() {
- return ((bitField0_ & 0x00000100) == 0x00000100);
+ return ((bitField0_ & 0x00000100) != 0);
}
/**
* <code>optional .EntityDescriptorProto processor_descriptor = 9;</code>
+ * @return The processorDescriptor.
*/
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getProcessorDescriptor() {
if (processorDescriptorBuilder_ == null) {
- return processorDescriptor_;
+ return processorDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : processorDescriptor_;
} else {
return processorDescriptorBuilder_.getMessage();
}
@@ -5482,11 +5990,11 @@ public final class LlapDaemonProtocolProtos {
throw new NullPointerException();
}
processorDescriptor_ = value;
- onChanged();
} else {
processorDescriptorBuilder_.setMessage(value);
}
bitField0_ |= 0x00000100;
+ onChanged();
return this;
}
/**
@@ -5496,11 +6004,11 @@ public final class LlapDaemonProtocolProtos {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder builderForValue) {
if (processorDescriptorBuilder_ == null) {
processorDescriptor_ = builderForValue.build();
- onChanged();
} else {
processorDescriptorBuilder_.setMessage(builderForValue.build());
}
bitField0_ |= 0x00000100;
+ onChanged();
return this;
}
/**
@@ -5508,31 +6016,33 @@ public final class LlapDaemonProtocolProtos {
*/
public Builder mergeProcessorDescriptor(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto value) {
if (processorDescriptorBuilder_ == null) {
- if (((bitField0_ & 0x00000100) == 0x00000100) &&
- processorDescriptor_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) {
- processorDescriptor_ =
- org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.newBuilder(processorDescriptor_).mergeFrom(value).buildPartial();
+ if (((bitField0_ & 0x00000100) != 0) &&
+ processorDescriptor_ != null &&
+ processorDescriptor_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) {
+ getProcessorDescriptorBuilder().mergeFrom(value);
} else {
processorDescriptor_ = value;
}
- onChanged();
} else {
processorDescriptorBuilder_.mergeFrom(value);
}
- bitField0_ |= 0x00000100;
+ if (processorDescriptor_ != null) {
+ bitField0_ |= 0x00000100;
+ onChanged();
+ }
return this;
}
/**
* <code>optional .EntityDescriptorProto processor_descriptor = 9;</code>
*/
public Builder clearProcessorDescriptor() {
- if (processorDescriptorBuilder_ == null) {
- processorDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
- onChanged();
- } else {
- processorDescriptorBuilder_.clear();
- }
bitField0_ = (bitField0_ & ~0x00000100);
+ processorDescriptor_ = null;
+ if (processorDescriptorBuilder_ != null) {
+ processorDescriptorBuilder_.dispose();
+ processorDescriptorBuilder_ = null;
+ }
+ onChanged();
return this;
}
/**
@@ -5550,19 +6060,20 @@ public final class LlapDaemonProtocolProtos {
if (processorDescriptorBuilder_ != null) {
return processorDescriptorBuilder_.getMessageOrBuilder();
} else {
- return processorDescriptor_;
+ return processorDescriptor_ == null ?
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : processorDescriptor_;
}
}
/**
* <code>optional .EntityDescriptorProto processor_descriptor = 9;</code>
*/
- private com.google.protobuf.SingleFieldBuilder<
+ private com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder>
getProcessorDescriptorFieldBuilder() {
if (processorDescriptorBuilder_ == null) {
- processorDescriptorBuilder_ = new com.google.protobuf.SingleFieldBuilder<
+ processorDescriptorBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder>(
- processorDescriptor_,
+ getProcessorDescriptor(),
getParentForChildren(),
isClean());
processorDescriptor_ = null;
@@ -5570,17 +6081,16 @@ public final class LlapDaemonProtocolProtos {
return processorDescriptorBuilder_;
}
- // repeated .IOSpecProto input_specs = 10;
private java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> inputSpecs_ =
java.util.Collections.emptyList();
private void ensureInputSpecsIsMutable() {
- if (!((bitField0_ & 0x00000200) == 0x00000200)) {
+ if (!((bitField0_ & 0x00000200) != 0)) {
inputSpecs_ = new java.util.ArrayList<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto>(inputSpecs_);
bitField0_ |= 0x00000200;
}
}
- private com.google.protobuf.RepeatedFieldBuilder<
+ private com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder> inputSpecsBuilder_;
/**
@@ -5712,7 +6222,8 @@ public final class LlapDaemonProtocolProtos {
java.lang.Iterable<? extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> values) {
if (inputSpecsBuilder_ == null) {
ensureInputSpecsIsMutable();
- super.addAll(values, inputSpecs_);
+ com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ values, inputSpecs_);
onChanged();
} else {
inputSpecsBuilder_.addAllMessages(values);
@@ -5795,14 +6306,14 @@ public final class LlapDaemonProtocolProtos {
getInputSpecsBuilderList() {
return getInputSpecsFieldBuilder().getBuilderList();
}
- private com.google.protobuf.RepeatedFieldBuilder<
+ private com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder>
getInputSpecsFieldBuilder() {
if (inputSpecsBuilder_ == null) {
- inputSpecsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
+ inputSpecsBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder>(
inputSpecs_,
- ((bitField0_ & 0x00000200) == 0x00000200),
+ ((bitField0_ & 0x00000200) != 0),
getParentForChildren(),
isClean());
inputSpecs_ = null;
@@ -5810,17 +6321,16 @@ public final class LlapDaemonProtocolProtos {
return inputSpecsBuilder_;
}
- // repeated .IOSpecProto output_specs = 11;
private java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> outputSpecs_ =
java.util.Collections.emptyList();
private void ensureOutputSpecsIsMutable() {
- if (!((bitField0_ & 0x00000400) == 0x00000400)) {
+ if (!((bitField0_ & 0x00000400) != 0)) {
outputSpecs_ = new java.util.ArrayList<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto>(outputSpecs_);
bitField0_ |= 0x00000400;
}
}
- private com.google.protobuf.RepeatedFieldBuilder<
+ private com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder> outputSpecsBuilder_;
/**
@@ -5952,7 +6462,8 @@ public final class LlapDaemonProtocolProtos {
java.lang.Iterable<? extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> values) {
if (outputSpecsBuilder_ == null) {
ensureOutputSpecsIsMutable();
- super.addAll(values, outputSpecs_);
+ com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ values, outputSpecs_);
onChanged();
} else {
outputSpecsBuilder_.addAllMessages(values);
@@ -6035,14 +6546,14 @@ public final class LlapDaemonProtocolProtos {
getOutputSpecsBuilderList() {
return getOutputSpecsFieldBuilder().getBuilderList();
}
- private com.google.protobuf.RepeatedFieldBuilder<
+ private com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder>
getOutputSpecsFieldBuilder() {
if (outputSpecsBuilder_ == null) {
- outputSpecsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
+ outputSpecsBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder>(
outputSpecs_,
- ((bitField0_ & 0x00000400) == 0x00000400),
+ ((bitField0_ & 0x00000400) != 0),
getParentForChildren(),
isClean());
outputSpecs_ = null;
@@ -6050,17 +6561,16 @@ public final class LlapDaemonProtocolProtos {
return outputSpecsBuilder_;
}
- // repeated .GroupInputSpecProto grouped_input_specs = 12;
private java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto> groupedInputSpecs_ =
java.util.Collections.emptyList();
private void ensureGroupedInputSpecsIsMutable() {
- if (!((bitField0_ & 0x00000800) == 0x00000800)) {
+ if (!((bitField0_ & 0x00000800) != 0)) {
groupedInputSpecs_ = new java.util.ArrayList<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto>(groupedInputSpecs_);
bitField0_ |= 0x00000800;
}
}
- private com.google.protobuf.RepeatedFieldBuilder<
+ private com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder> groupedInputSpecsBuilder_;
/**
@@ -6192,7 +6702,8 @@ public final class LlapDaemonProtocolProtos {
java.lang.Iterable<? extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto> values) {
if (groupedInputSpecsBuilder_ == null) {
ensureGroupedInputSpecsIsMutable();
- super.addAll(values, groupedInputSpecs_);
+ com.google.protobuf.AbstractMessageLite.Builder.addAll(
+ values, groupedInputSpecs_);
onChanged();
} else {
groupedInputSpecsBuilder_.addAllMessages(values);
@@ -6275,14 +6786,14 @@ public final class LlapDaemonProtocolProtos {
getGroupedInputSpecsBuilderList() {
return getGroupedInputSpecsFieldBuilder().getBuilderList();
}
- private com.google.protobuf.RepeatedFieldBuilder<
+ private com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder>
getGroupedInputSpecsFieldBuilder() {
if (groupedInputSpecsBuilder_ == null) {
- groupedInputSpecsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
+ groupedInputSpecsBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder>(
groupedInputSpecs_,
- ((bitField0_ & 0x00000800) == 0x00000800),
+ ((bitField0_ & 0x00000800) != 0),
getParentForChildren(),
isClean());
groupedInputSpecs_ = null;
@@ -6290,47 +6801,54 @@ public final class LlapDaemonProtocolProtos {
return groupedInputSpecsBuilder_;
}
- // optional int32 vertex_parallelism = 13;
private int vertexParallelism_ ;
/**
- * <code>optional int32 vertex_parallelism = 13;</code>
- *
* <pre>
* An internal field required for Tez.
* </pre>
+ *
+ * <code>optional int32 vertex_parallelism = 13;</code>
+ * @return Whether the vertexParallelism field is set.
*/
+ @java.lang.Override
public boolean hasVertexParallelism() {
- return ((bitField0_ & 0x00001000) == 0x00001000);
+ return ((bitField0_ & 0x00001000) != 0);
}
/**
- * <code>optional int32 vertex_parallelism = 13;</code>
- *
* <pre>
* An internal field required for Tez.
* </pre>
+ *
+ * <code>optional int32 vertex_parallelism = 13;</code>
+ * @return The vertexParallelism.
*/
+ @java.lang.Override
public int getVertexParallelism() {
return vertexParallelism_;
}
/**
- * <code>optional int32 vertex_parallelism = 13;</code>
- *
* <pre>
* An internal field required for Tez.
* </pre>
+ *
+ * <code>optional int32 vertex_parallelism = 13;</code>
+ * @param value The vertexParallelism to set.
+ * @return This builder for chaining.
*/
public Builder setVertexParallelism(int value) {
- bitField0_ |= 0x00001000;
+
vertexParallelism_ = value;
+ bitField0_ |= 0x00001000;
onChanged();
return this;
}
/**
- * <code>optional int32 vertex_parallelism = 13;</code>
- *
* <pre>
* An internal field required for Tez.
* </pre>
+ *
+ * <code>optional int32 vertex_parallelism = 13;</code>
+ * @return This builder for chaining.
*/
public Builder clearVertexParallelism() {
bitField0_ = (bitField0_ & ~0x00001000);
@@ -6339,31 +6857,38 @@ public final class LlapDaemonProtocolProtos {
return this;
}
- // optional bool is_external_submission = 14 [default = false];
private boolean isExternalSubmission_ ;
/**
* <code>optional bool is_external_submission = 14 [default = false];</code>
+ * @return Whether the isExternalSubmission field is set.
*/
+ @java.lang.Override
public boolean hasIsExternalSubmission() {
- return ((bitField0_ & 0x00002000) == 0x00002000);
+ return ((bitField0_ & 0x00002000) != 0);
}
/**
* <code>optional bool is_external_submission = 14 [default = false];</code>
+ * @return The isExternalSubmission.
*/
+ @java.lang.Override
public boolean getIsExternalSubmission() {
return isExternalSubmission_;
}
/**
* <code>optional bool is_external_submission = 14 [default = false];</code>
+ * @param value The isExternalSubmission to set.
+ * @return This builder for chaining.
*/
public Builder setIsExternalSubmission(boolean value) {
- bitField0_ |= 0x00002000;
+
isExternalSubmission_ = value;
+ bitField0_ |= 0x00002000;
onChanged();
return this;
}
/**
* <code>optional bool is_external_submission = 14 [default = false];</code>
+ * @return This builder for chaining.
*/
public Builder clearIsExternalSubmission() {
bitField0_ = (bitField0_ & ~0x00002000);
@@ -6371,28 +6896,82 @@ public final class LlapDaemonProtocolProtos {
onChanged();
return this;
}
+ @java.lang.Override
+ public final Builder setUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.setUnknownFields(unknownFields);
+ }
+
+ @java.lang.Override
+ public final Builder mergeUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.mergeUnknownFields(unknownFields);
+ }
+
// @@protoc_insertion_point(builder_scope:SignableVertexSpec)
}
+ // @@protoc_insertion_point(class_scope:SignableVertexSpec)
+ private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec DEFAULT_INSTANCE;
static {
- defaultInstance = new SignableVertexSpec(true);
- defaultInstance.initFields();
+ DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec();
+ }
+
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec getDefaultInstance() {
+ return DEFAULT_INSTANCE;
+ }
+
+ @java.lang.Deprecated public static final com.google.protobuf.Parser<SignableVertexSpec>
+ PARSER = new com.google.protobuf.AbstractParser<SignableVertexSpec>() {
+ @java.lang.Override
+ public SignableVertexSpec parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ Builder builder = newBuilder();
+ try {
+ builder.mergeFrom(input, extensionRegistry);
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(builder.buildPartial());
+ } catch (com.google.protobuf.UninitializedMessageException e) {
+ throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(e)
+ .setUnfinishedMessage(builder.buildPartial());
+ }
+ return builder.buildPartial();
+ }
+ };
+
+ public static com.google.protobuf.Parser<SignableVertexSpec> parser() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<SignableVertexSpec> getParserForType() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec getDefaultInstanceForType() {
+ return DEFAULT_INSTANCE;
}
- // @@protoc_insertion_point(class_scope:SignableVertexSpec)
}
- public interface VertexOrBinaryOrBuilder
- extends com.google.protobuf.MessageOrBuilder {
+ public interface VertexOrBinaryOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:VertexOrBinary)
+ com.google.protobuf.MessageOrBuilder {
- // optional .SignableVertexSpec vertex = 1;
/**
* <code>optional .SignableVertexSpec vertex = 1;</code>
+ * @return Whether the vertex field is set.
*/
boolean hasVertex();
/**
* <code>optional .SignableVertexSpec vertex = 1;</code>
+ * @return The vertex.
*/
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec getVertex();
/**
@@ -6400,234 +6979,161 @@ public final class LlapDaemonProtocolProtos {
*/
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder getVertexOrBuilder();
- // optional bytes vertexBinary = 2;
/**
- * <code>optional bytes vertexBinary = 2;</code>
- *
* <pre>
* SignableVertexSpec
* </pre>
+ *
+ * <code>optional bytes vertexBinary = 2;</code>
+ * @return Whether the vertexBinary field is set.
*/
boolean hasVertexBinary();
/**
- * <code>optional bytes vertexBinary = 2;</code>
- *
* <pre>
* SignableVertexSpec
* </pre>
+ *
+ * <code>optional bytes vertexBinary = 2;</code>
+ * @return The vertexBinary.
*/
com.google.protobuf.ByteString getVertexBinary();
}
/**
- * Protobuf type {@code VertexOrBinary}
- *
* <pre>
* Union
* </pre>
+ *
+ * Protobuf type {@code VertexOrBinary}
*/
public static final class VertexOrBinary extends
- com.google.protobuf.GeneratedMessage
- implements VertexOrBinaryOrBuilder {
+ com.google.protobuf.GeneratedMessageV3 implements
+ // @@protoc_insertion_point(message_implements:VertexOrBinary)
+ VertexOrBinaryOrBuilder {
+ private static final long serialVersionUID = 0L;
// Use VertexOrBinary.newBuilder() to construct.
- private VertexOrBinary(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ private VertexOrBinary(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
- this.unknownFields = builder.getUnknownFields();
- }
- private VertexOrBinary(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
- private static final VertexOrBinary defaultInstance;
- public static VertexOrBinary getDefaultInstance() {
- return defaultInstance;
}
-
- public VertexOrBinary getDefaultInstanceForType() {
- return defaultInstance;
+ private VertexOrBinary() {
+ vertexBinary_ = com.google.protobuf.ByteString.EMPTY;
}
- private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
- return this.unknownFields;
- }
- private VertexOrBinary(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- initFields();
- int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
- try {
- boolean done = false;
- while (!done) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
- done = true;
- break;
- default: {
- if (!parseUnknownField(input, unknownFields,
- extensionRegistry, tag)) {
- done = true;
- }
- break;
- }
- case 10: {
- org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder subBuilder = null;
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
- subBuilder = vertex_.toBuilder();
- }
- vertex_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.PARSER, extensionRegistry);
- if (subBuilder != null) {
- subBuilder.mergeFrom(vertex_);
- vertex_ = subBuilder.buildPartial();
- }
- bitField0_ |= 0x00000001;
- break;
- }
- case 18: {
- bitField0_ |= 0x00000002;
- vertexBinary_ = input.readBytes();
- break;
- }
- }
- }
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
- throw e.setUnfinishedMessage(this);
- } catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
- e.getMessage()).setUnfinishedMessage(this);
- } finally {
- this.unknownFields = unknownFields.build();
- makeExtensionsImmutable();
- }
+ @SuppressWarnings({"unused"})
+ protected java.lang.Object newInstance(
+ UnusedPrivateParameter unused) {
+ return new VertexOrBinary();
}
+
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexOrBinary_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ @java.lang.Override
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexOrBinary_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.Builder.class);
}
- public static com.google.protobuf.Parser<VertexOrBinary> PARSER =
- new com.google.protobuf.AbstractParser<VertexOrBinary>() {
- public VertexOrBinary parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return new VertexOrBinary(input, extensionRegistry);
- }
- };
-
- @java.lang.Override
- public com.google.protobuf.Parser<VertexOrBinary> getParserForType() {
- return PARSER;
- }
-
private int bitField0_;
- // optional .SignableVertexSpec vertex = 1;
public static final int VERTEX_FIELD_NUMBER = 1;
private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec vertex_;
/**
* <code>optional .SignableVertexSpec vertex = 1;</code>
+ * @return Whether the vertex field is set.
*/
+ @java.lang.Override
public boolean hasVertex() {
- return ((bitField0_ & 0x00000001) == 0x00000001);
+ return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional .SignableVertexSpec vertex = 1;</code>
+ * @return The vertex.
*/
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec getVertex() {
- return vertex_;
+ return vertex_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance() : vertex_;
}
/**
* <code>optional .SignableVertexSpec vertex = 1;</code>
*/
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder getVertexOrBuilder() {
- return vertex_;
+ return vertex_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance() : vertex_;
}
- // optional bytes vertexBinary = 2;
public static final int VERTEXBINARY_FIELD_NUMBER = 2;
- private com.google.protobuf.ByteString vertexBinary_;
+ private com.google.protobuf.ByteString vertexBinary_ = com.google.protobuf.ByteString.EMPTY;
/**
- * <code>optional bytes vertexBinary = 2;</code>
- *
* <pre>
* SignableVertexSpec
* </pre>
+ *
+ * <code>optional bytes vertexBinary = 2;</code>
+ * @return Whether the vertexBinary field is set.
*/
+ @java.lang.Override
public boolean hasVertexBinary() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
+ return ((bitField0_ & 0x00000002) != 0);
}
/**
- * <code>optional bytes vertexBinary = 2;</code>
- *
* <pre>
* SignableVertexSpec
* </pre>
+ *
+ * <code>optional bytes vertexBinary = 2;</code>
+ * @return The vertexBinary.
*/
+ @java.lang.Override
public com.google.protobuf.ByteString getVertexBinary() {
return vertexBinary_;
}
- private void initFields() {
- vertex_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance();
- vertexBinary_ = com.google.protobuf.ByteString.EMPTY;
- }
private byte memoizedIsInitialized = -1;
+ @java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
- if (isInitialized != -1) return isInitialized == 1;
+ if (isInitialized == 1) return true;
+ if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
+ @java.lang.Override
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
- getSerializedSize();
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
- output.writeMessage(1, vertex_);
+ if (((bitField0_ & 0x00000001) != 0)) {
+ output.writeMessage(1, getVertex());
}
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ if (((bitField0_ & 0x00000002) != 0)) {
output.writeBytes(2, vertexBinary_);
}
getUnknownFields().writeTo(output);
}
- private int memoizedSerializedSize = -1;
+ @java.lang.Override
public int getSerializedSize() {
- int size = memoizedSerializedSize;
+ int size = memoizedSize;
if (size != -1) return size;
size = 0;
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ if (((bitField0_ & 0x00000001) != 0)) {
size += com.google.protobuf.CodedOutputStream
- .computeMessageSize(1, vertex_);
+ .computeMessageSize(1, getVertex());
}
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ if (((bitField0_ & 0x00000002) != 0)) {
size += com.google.protobuf.CodedOutputStream
.computeBytesSize(2, vertexBinary_);
}
size += getUnknownFields().getSerializedSize();
- memoizedSerializedSize = size;
+ memoizedSize = size;
return size;
}
- private static final long serialVersionUID = 0L;
- @java.lang.Override
- protected java.lang.Object writeReplace()
- throws java.io.ObjectStreamException {
- return super.writeReplace();
- }
-
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
@@ -6638,30 +7144,27 @@ public final class LlapDaemonProtocolProtos {
}
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary) obj;
- boolean result = true;
- result = result && (hasVertex() == other.hasVertex());
+ if (hasVertex() != other.hasVertex()) return false;
if (hasVertex()) {
- result = result && getVertex()
- .equals(other.getVertex());
+ if (!getVertex()
+ .equals(other.getVertex())) return false;
}
- result = result && (hasVertexBinary() == other.hasVertexBinary());
+ if (hasVertexBinary() != other.hasVertexBinary()) return false;
if (hasVertexBinary()) {
- result = result && getVertexBinary()
- .equals(other.getVertexBinary());
+ if (!getVertexBinary()
+ .equals(other.getVertexBinary())) return false;
}
- result = result &&
- getUnknownFields().equals(other.getUnknownFields());
- return result;
+ if (!getUnknownFields().equals(other.getUnknownFields())) return false;
+ return true;
}
- private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
- hash = (19 * hash) + getDescriptorForType().hashCode();
+ hash = (19 * hash) + getDescriptor().hashCode();
if (hasVertex()) {
hash = (37 * hash) + VERTEX_FIELD_NUMBER;
hash = (53 * hash) + getVertex().hashCode();
@@ -6675,6 +7178,17 @@ public final class LlapDaemonProtocolProtos {
return hash;
}
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom(
+ java.nio.ByteBuffer data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom(
+ java.nio.ByteBuffer data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
@@ -6698,65 +7212,82 @@ public final class LlapDaemonProtocolProtos {
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
+
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input);
}
+
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
- public static Builder newBuilder() { return Builder.create(); }
+ @java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder() {
+ return DEFAULT_INSTANCE.toBuilder();
+ }
public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary prototype) {
- return newBuilder().mergeFrom(prototype);
+ return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+ }
+ @java.lang.Override
+ public Builder toBuilder() {
+ return this == DEFAULT_INSTANCE
+ ? new Builder() : new Builder().mergeFrom(this);
}
- public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
- * Protobuf type {@code VertexOrBinary}
- *
* <pre>
* Union
* </pre>
+ *
+ * Protobuf type {@code VertexOrBinary}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinaryOrBuilder {
+ com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ // @@protoc_insertion_point(builder_implements:VertexOrBinary)
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinaryOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexOrBinary_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ @java.lang.Override
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexOrBinary_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -6769,45 +7300,41 @@ public final class LlapDaemonProtocolProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ if (com.google.protobuf.GeneratedMessageV3
+ .alwaysUseFieldBuilders) {
getVertexFieldBuilder();
}
}
- private static Builder create() {
- return new Builder();
- }
-
+ @java.lang.Override
public Builder clear() {
super.clear();
- if (vertexBuilder_ == null) {
- vertex_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance();
- } else {
- vertexBuilder_.clear();
+ bitField0_ = 0;
+ vertex_ = null;
+ if (vertexBuilder_ != null) {
+ vertexBuilder_.dispose();
+ vertexBuilder_ = null;
}
- bitField0_ = (bitField0_ & ~0x00000001);
vertexBinary_ = com.google.protobuf.ByteString.EMPTY;
- bitField0_ = (bitField0_ & ~0x00000002);
return this;
}
- public Builder clone() {
- return create().mergeFrom(buildPartial());
- }
-
+ @java.lang.Override
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexOrBinary_descriptor;
}
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary getDefaultInstanceForType() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.getDefaultInstance();
}
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary build() {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary result = buildPartial();
if (!result.isInitialized()) {
@@ -6816,27 +7343,63 @@ public final class LlapDaemonProtocolProtos {
return result;
}
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary buildPartial() {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary(this);
+ if (bitField0_ != 0) { buildPartial0(result); }
+ onBuilt();
+ return result;
+ }
+
+ private void buildPartial0(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
- if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
+ if (((from_bitField0_ & 0x00000001) != 0)) {
+ result.vertex_ = vertexBuilder_ == null
+ ? vertex_
+ : vertexBuilder_.build();
to_bitField0_ |= 0x00000001;
}
- if (vertexBuilder_ == null) {
- result.vertex_ = vertex_;
- } else {
- result.vertex_ = vertexBuilder_.build();
- }
- if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
+ if (((from_bitField0_ & 0x00000002) != 0)) {
+ result.vertexBinary_ = vertexBinary_;
to_bitField0_ |= 0x00000002;
}
- result.vertexBinary_ = vertexBinary_;
- result.bitField0_ = to_bitField0_;
- onBuilt();
- return result;
+ result.bitField0_ |= to_bitField0_;
}
+ @java.lang.Override
+ public Builder clone() {
+ return super.clone();
+ }
+ @java.lang.Override
+ public Builder setField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ java.lang.Object value) {
+ return super.setField(field, value);
+ }
+ @java.lang.Override
+ public Builder clearField(
+ com.google.protobuf.Descriptors.FieldDescriptor field) {
+ return super.clearField(field);
+ }
+ @java.lang.Override
+ public Builder clearOneof(
+ com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ return super.clearOneof(oneof);
+ }
+ @java.lang.Override
+ public Builder setRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ int index, java.lang.Object value) {
+ return super.setRepeatedField(field, index, value);
+ }
+ @java.lang.Override
+ public Builder addRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ java.lang.Object value) {
+ return super.addRepeatedField(field, value);
+ }
+ @java.lang.Override
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary) {
return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary)other);
@@ -6855,48 +7418,77 @@ public final class LlapDaemonProtocolProtos {
setVertexBinary(other.getVertexBinary());
}
this.mergeUnknownFields(other.getUnknownFields());
+ onChanged();
return this;
}
+ @java.lang.Override
public final boolean isInitialized() {
return true;
}
+ @java.lang.Override
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parsedMessage = null;
+ if (extensionRegistry == null) {
+ throw new java.lang.NullPointerException();
+ }
try {
- parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ case 10: {
+ input.readMessage(
+ getVertexFieldBuilder().getBuilder(),
+ extensionRegistry);
+ bitField0_ |= 0x00000001;
+ break;
+ } // case 10
+ case 18: {
+ vertexBinary_ = input.readBytes();
+ bitField0_ |= 0x00000002;
+ break;
+ } // case 18
+ default: {
+ if (!super.parseUnknownField(input, extensionRegistry, tag)) {
+ done = true; // was an endgroup tag
+ }
+ break;
+ } // default:
+ } // switch (tag)
+ } // while (!done)
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
- parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary) e.getUnfinishedMessage();
- throw e;
+ throw e.unwrapIOException();
} finally {
- if (parsedMessage != null) {
- mergeFrom(parsedMessage);
- }
- }
+ onChanged();
+ } // finally
return this;
}
private int bitField0_;
- // optional .SignableVertexSpec vertex = 1;
- private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec vertex_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance();
- private com.google.protobuf.SingleFieldBuilder<
+ private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec vertex_;
+ private com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder> vertexBuilder_;
/**
* <code>optional .SignableVertexSpec vertex = 1;</code>
+ * @return Whether the vertex field is set.
*/
public boolean hasVertex() {
- return ((bitField0_ & 0x00000001) == 0x00000001);
+ return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional .SignableVertexSpec vertex = 1;</code>
+ * @return The vertex.
*/
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec getVertex() {
if (vertexBuilder_ == null) {
- return vertex_;
+ return vertex_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance() : vertex_;
} else {
return vertexBuilder_.getMessage();
}
@@ -6910,11 +7502,11 @@ public final class LlapDaemonProtocolProtos {
throw new NullPointerException();
}
vertex_ = value;
- onChanged();
} else {
vertexBuilder_.setMessage(value);
}
bitField0_ |= 0x00000001;
+ onChanged();
return this;
}
/**
@@ -6924,11 +7516,11 @@ public final class LlapDaemonProtocolProtos {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder builderForValue) {
if (vertexBuilder_ == null) {
vertex_ = builderForValue.build();
- onChanged();
} else {
vertexBuilder_.setMessage(builderForValue.build());
}
bitField0_ |= 0x00000001;
+ onChanged();
return this;
}
/**
@@ -6936,31 +7528,33 @@ public final class LlapDaemonProtocolProtos {
*/
public Builder mergeVertex(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec value) {
if (vertexBuilder_ == null) {
- if (((bitField0_ & 0x00000001) == 0x00000001) &&
- vertex_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance()) {
- vertex_ =
- org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.newBuilder(vertex_).mergeFrom(value).buildPartial();
+ if (((bitField0_ & 0x00000001) != 0) &&
+ vertex_ != null &&
+ vertex_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance()) {
+ getVertexBuilder().mergeFrom(value);
} else {
vertex_ = value;
}
- onChanged();
} else {
vertexBuilder_.mergeFrom(value);
}
- bitField0_ |= 0x00000001;
+ if (vertex_ != null) {
+ bitField0_ |= 0x00000001;
+ onChanged();
+ }
return this;
}
/**
* <code>optional .SignableVertexSpec vertex = 1;</code>
*/
public Builder clearVertex() {
- if (vertexBuilder_ == null) {
- vertex_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance();
- onChanged();
- } else {
- vertexBuilder_.clear();
- }
bitField0_ = (bitField0_ & ~0x00000001);
+ vertex_ = null;
+ if (vertexBuilder_ != null) {
+ vertexBuilder_.dispose();
+ vertexBuilder_ = null;
+ }
+ onChanged();
return this;
}
/**
@@ -6978,19 +7572,20 @@ public final class LlapDaemonProtocolProtos {
if (vertexBuilder_ != null) {
return vertexBuilder_.getMessageOrBuilder();
} else {
- return vertex_;
+ return vertex_ == null ?
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance() : vertex_;
}
}
/**
* <code>optional .SignableVertexSpec vertex = 1;</code>
*/
- private com.google.protobuf.SingleFieldBuilder<
+ private com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder>
getVertexFieldBuilder() {
if (vertexBuilder_ == null) {
- vertexBuilder_ = new com.google.protobuf.SingleFieldBuilder<
+ vertexBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder>(
- vertex_,
+ getVertex(),
getParentForChildren(),
isClean());
vertex_ = null;
@@ -6998,50 +7593,54 @@ public final class LlapDaemonProtocolProtos {
return vertexBuilder_;
}
- // optional bytes vertexBinary = 2;
private com.google.protobuf.ByteString vertexBinary_ = com.google.protobuf.ByteString.EMPTY;
/**
- * <code>optional bytes vertexBinary = 2;</code>
- *
* <pre>
* SignableVertexSpec
* </pre>
+ *
+ * <code>optional bytes vertexBinary = 2;</code>
+ * @return Whether the vertexBinary field is set.
*/
+ @java.lang.Override
public boolean hasVertexBinary() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
+ return ((bitField0_ & 0x00000002) != 0);
}
/**
- * <code>optional bytes vertexBinary = 2;</code>
- *
* <pre>
* SignableVertexSpec
* </pre>
+ *
+ * <code>optional bytes vertexBinary = 2;</code>
+ * @return The vertexBinary.
*/
+ @java.lang.Override
public com.google.protobuf.ByteString getVertexBinary() {
return vertexBinary_;
}
/**
- * <code>optional bytes vertexBinary = 2;</code>
- *
* <pre>
* SignableVertexSpec
* </pre>
+ *
+ * <code>optional bytes vertexBinary = 2;</code>
+ * @param value The vertexBinary to set.
+ * @return This builder for chaining.
*/
public Builder setVertexBinary(com.google.protobuf.ByteString value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000002;
+ if (value == null) { throw new NullPointerException(); }
vertexBinary_ = value;
+ bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
- * <code>optional bytes vertexBinary = 2;</code>
- *
* <pre>
* SignableVertexSpec
* </pre>
+ *
+ * <code>optional bytes vertexBinary = 2;</code>
+ * @return This builder for chaining.
*/
public Builder clearVertexBinary() {
bitField0_ = (bitField0_ & ~0x00000002);
@@ -7049,78 +7648,137 @@ public final class LlapDaemonProtocolProtos {
onChanged();
return this;
}
+ @java.lang.Override
+ public final Builder setUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.setUnknownFields(unknownFields);
+ }
+
+ @java.lang.Override
+ public final Builder mergeUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.mergeUnknownFields(unknownFields);
+ }
+
// @@protoc_insertion_point(builder_scope:VertexOrBinary)
}
+ // @@protoc_insertion_point(class_scope:VertexOrBinary)
+ private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary DEFAULT_INSTANCE;
static {
- defaultInstance = new VertexOrBinary(true);
- defaultInstance.initFields();
+ DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary();
}
- // @@protoc_insertion_point(class_scope:VertexOrBinary)
- }
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary getDefaultInstance() {
+ return DEFAULT_INSTANCE;
+ }
- public interface FragmentRuntimeInfoOrBuilder
- extends com.google.protobuf.MessageOrBuilder {
+ @java.lang.Deprecated public static final com.google.protobuf.Parser<VertexOrBinary>
+ PARSER = new com.google.protobuf.AbstractParser<VertexOrBinary>() {
+ @java.lang.Override
+ public VertexOrBinary parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ Builder builder = newBuilder();
+ try {
+ builder.mergeFrom(input, extensionRegistry);
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(builder.buildPartial());
+ } catch (com.google.protobuf.UninitializedMessageException e) {
+ throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(e)
+ .setUnfinishedMessage(builder.buildPartial());
+ }
+ return builder.buildPartial();
+ }
+ };
- // optional int32 num_self_and_upstream_tasks = 1;
- /**
- * <code>optional int32 num_self_and_upstream_tasks = 1;</code>
- */
- boolean hasNumSelfAndUpstreamTasks();
- /**
+ public static com.google.protobuf.Parser<VertexOrBinary> parser() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<VertexOrBinary> getParserForType() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary getDefaultInstanceForType() {
+ return DEFAULT_INSTANCE;
+ }
+
+ }
+
+ public interface FragmentRuntimeInfoOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:FragmentRuntimeInfo)
+ com.google.protobuf.MessageOrBuilder {
+
+ /**
+ * <code>optional int32 num_self_and_upstream_tasks = 1;</code>
+ * @return Whether the numSelfAndUpstreamTasks field is set.
+ */
+ boolean hasNumSelfAndUpstreamTasks();
+ /**
* <code>optional int32 num_self_and_upstream_tasks = 1;</code>
+ * @return The numSelfAndUpstreamTasks.
*/
int getNumSelfAndUpstreamTasks();
- // optional int32 num_self_and_upstream_completed_tasks = 2;
/**
* <code>optional int32 num_self_and_upstream_completed_tasks = 2;</code>
+ * @return Whether the numSelfAndUpstreamCompletedTasks field is set.
*/
boolean hasNumSelfAndUpstreamCompletedTasks();
/**
* <code>optional int32 num_self_and_upstream_completed_tasks = 2;</code>
+ * @return The numSelfAndUpstreamCompletedTasks.
*/
int getNumSelfAndUpstreamCompletedTasks();
- // optional int32 within_dag_priority = 3;
/**
* <code>optional int32 within_dag_priority = 3;</code>
+ * @return Whether the withinDagPriority field is set.
*/
boolean hasWithinDagPriority();
/**
* <code>optional int32 within_dag_priority = 3;</code>
+ * @return The withinDagPriority.
*/
int getWithinDagPriority();
- // optional int64 dag_start_time = 4;
/**
* <code>optional int64 dag_start_time = 4;</code>
+ * @return Whether the dagStartTime field is set.
*/
boolean hasDagStartTime();
/**
* <code>optional int64 dag_start_time = 4;</code>
+ * @return The dagStartTime.
*/
long getDagStartTime();
- // optional int64 first_attempt_start_time = 5;
/**
* <code>optional int64 first_attempt_start_time = 5;</code>
+ * @return Whether the firstAttemptStartTime field is set.
*/
boolean hasFirstAttemptStartTime();
/**
* <code>optional int64 first_attempt_start_time = 5;</code>
+ * @return The firstAttemptStartTime.
*/
long getFirstAttemptStartTime();
- // optional int64 current_attempt_start_time = 6;
/**
* <code>optional int64 current_attempt_start_time = 6;</code>
+ * @return Whether the currentAttemptStartTime field is set.
*/
boolean hasCurrentAttemptStartTime();
/**
* <code>optional int64 current_attempt_start_time = 6;</code>
+ * @return The currentAttemptStartTime.
*/
long getCurrentAttemptStartTime();
}
@@ -7128,302 +7786,222 @@ public final class LlapDaemonProtocolProtos {
* Protobuf type {@code FragmentRuntimeInfo}
*/
public static final class FragmentRuntimeInfo extends
- com.google.protobuf.GeneratedMessage
- implements FragmentRuntimeInfoOrBuilder {
+ com.google.protobuf.GeneratedMessageV3 implements
+ // @@protoc_insertion_point(message_implements:FragmentRuntimeInfo)
+ FragmentRuntimeInfoOrBuilder {
+ private static final long serialVersionUID = 0L;
// Use FragmentRuntimeInfo.newBuilder() to construct.
- private FragmentRuntimeInfo(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ private FragmentRuntimeInfo(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
- this.unknownFields = builder.getUnknownFields();
- }
- private FragmentRuntimeInfo(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
- private static final FragmentRuntimeInfo defaultInstance;
- public static FragmentRuntimeInfo getDefaultInstance() {
- return defaultInstance;
}
-
- public FragmentRuntimeInfo getDefaultInstanceForType() {
- return defaultInstance;
+ private FragmentRuntimeInfo() {
}
- private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
- return this.unknownFields;
- }
- private FragmentRuntimeInfo(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- initFields();
- int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
- try {
- boolean done = false;
- while (!done) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
- done = true;
- break;
- default: {
- if (!parseUnknownField(input, unknownFields,
- extensionRegistry, tag)) {
- done = true;
- }
- break;
- }
- case 8: {
- bitField0_ |= 0x00000001;
- numSelfAndUpstreamTasks_ = input.readInt32();
- break;
- }
- case 16: {
- bitField0_ |= 0x00000002;
- numSelfAndUpstreamCompletedTasks_ = input.readInt32();
- break;
- }
- case 24: {
- bitField0_ |= 0x00000004;
- withinDagPriority_ = input.readInt32();
- break;
- }
- case 32: {
- bitField0_ |= 0x00000008;
- dagStartTime_ = input.readInt64();
- break;
- }
- case 40: {
- bitField0_ |= 0x00000010;
- firstAttemptStartTime_ = input.readInt64();
- break;
- }
- case 48: {
- bitField0_ |= 0x00000020;
- currentAttemptStartTime_ = input.readInt64();
- break;
- }
- }
- }
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
- throw e.setUnfinishedMessage(this);
- } catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
- e.getMessage()).setUnfinishedMessage(this);
- } finally {
- this.unknownFields = unknownFields.build();
- makeExtensionsImmutable();
- }
+ @SuppressWarnings({"unused"})
+ protected java.lang.Object newInstance(
+ UnusedPrivateParameter unused) {
+ return new FragmentRuntimeInfo();
}
+
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentRuntimeInfo_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ @java.lang.Override
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentRuntimeInfo_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.Builder.class);
}
- public static com.google.protobuf.Parser<FragmentRuntimeInfo> PARSER =
- new com.google.protobuf.AbstractParser<FragmentRuntimeInfo>() {
- public FragmentRuntimeInfo parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return new FragmentRuntimeInfo(input, extensionRegistry);
- }
- };
-
- @java.lang.Override
- public com.google.protobuf.Parser<FragmentRuntimeInfo> getParserForType() {
- return PARSER;
- }
-
private int bitField0_;
- // optional int32 num_self_and_upstream_tasks = 1;
public static final int NUM_SELF_AND_UPSTREAM_TASKS_FIELD_NUMBER = 1;
- private int numSelfAndUpstreamTasks_;
+ private int numSelfAndUpstreamTasks_ = 0;
/**
* <code>optional int32 num_self_and_upstream_tasks = 1;</code>
+ * @return Whether the numSelfAndUpstreamTasks field is set.
*/
+ @java.lang.Override
public boolean hasNumSelfAndUpstreamTasks() {
- return ((bitField0_ & 0x00000001) == 0x00000001);
+ return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional int32 num_self_and_upstream_tasks = 1;</code>
+ * @return The numSelfAndUpstreamTasks.
*/
+ @java.lang.Override
public int getNumSelfAndUpstreamTasks() {
return numSelfAndUpstreamTasks_;
}
- // optional int32 num_self_and_upstream_completed_tasks = 2;
public static final int NUM_SELF_AND_UPSTREAM_COMPLETED_TASKS_FIELD_NUMBER = 2;
- private int numSelfAndUpstreamCompletedTasks_;
+ private int numSelfAndUpstreamCompletedTasks_ = 0;
/**
* <code>optional int32 num_self_and_upstream_completed_tasks = 2;</code>
+ * @return Whether the numSelfAndUpstreamCompletedTasks field is set.
*/
+ @java.lang.Override
public boolean hasNumSelfAndUpstreamCompletedTasks() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
+ return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional int32 num_self_and_upstream_completed_tasks = 2;</code>
+ * @return The numSelfAndUpstreamCompletedTasks.
*/
+ @java.lang.Override
public int getNumSelfAndUpstreamCompletedTasks() {
return numSelfAndUpstreamCompletedTasks_;
}
- // optional int32 within_dag_priority = 3;
public static final int WITHIN_DAG_PRIORITY_FIELD_NUMBER = 3;
- private int withinDagPriority_;
+ private int withinDagPriority_ = 0;
/**
* <code>optional int32 within_dag_priority = 3;</code>
+ * @return Whether the withinDagPriority field is set.
*/
+ @java.lang.Override
public boolean hasWithinDagPriority() {
- return ((bitField0_ & 0x00000004) == 0x00000004);
+ return ((bitField0_ & 0x00000004) != 0);
}
/**
* <code>optional int32 within_dag_priority = 3;</code>
+ * @return The withinDagPriority.
*/
+ @java.lang.Override
public int getWithinDagPriority() {
return withinDagPriority_;
}
- // optional int64 dag_start_time = 4;
public static final int DAG_START_TIME_FIELD_NUMBER = 4;
- private long dagStartTime_;
+ private long dagStartTime_ = 0L;
/**
* <code>optional int64 dag_start_time = 4;</code>
+ * @return Whether the dagStartTime field is set.
*/
+ @java.lang.Override
public boolean hasDagStartTime() {
- return ((bitField0_ & 0x00000008) == 0x00000008);
+ return ((bitField0_ & 0x00000008) != 0);
}
/**
* <code>optional int64 dag_start_time = 4;</code>
+ * @return The dagStartTime.
*/
+ @java.lang.Override
public long getDagStartTime() {
return dagStartTime_;
}
- // optional int64 first_attempt_start_time = 5;
public static final int FIRST_ATTEMPT_START_TIME_FIELD_NUMBER = 5;
- private long firstAttemptStartTime_;
+ private long firstAttemptStartTime_ = 0L;
/**
* <code>optional int64 first_attempt_start_time = 5;</code>
+ * @return Whether the firstAttemptStartTime field is set.
*/
+ @java.lang.Override
public boolean hasFirstAttemptStartTime() {
- return ((bitField0_ & 0x00000010) == 0x00000010);
+ return ((bitField0_ & 0x00000010) != 0);
}
/**
* <code>optional int64 first_attempt_start_time = 5;</code>
+ * @return The firstAttemptStartTime.
*/
+ @java.lang.Override
public long getFirstAttemptStartTime() {
return firstAttemptStartTime_;
}
- // optional int64 current_attempt_start_time = 6;
public static final int CURRENT_ATTEMPT_START_TIME_FIELD_NUMBER = 6;
- private long currentAttemptStartTime_;
+ private long currentAttemptStartTime_ = 0L;
/**
* <code>optional int64 current_attempt_start_time = 6;</code>
+ * @return Whether the currentAttemptStartTime field is set.
*/
+ @java.lang.Override
public boolean hasCurrentAttemptStartTime() {
- return ((bitField0_ & 0x00000020) == 0x00000020);
+ return ((bitField0_ & 0x00000020) != 0);
}
/**
* <code>optional int64 current_attempt_start_time = 6;</code>
+ * @return The currentAttemptStartTime.
*/
+ @java.lang.Override
public long getCurrentAttemptStartTime() {
return currentAttemptStartTime_;
}
- private void initFields() {
- numSelfAndUpstreamTasks_ = 0;
- numSelfAndUpstreamCompletedTasks_ = 0;
- withinDagPriority_ = 0;
- dagStartTime_ = 0L;
- firstAttemptStartTime_ = 0L;
- currentAttemptStartTime_ = 0L;
- }
private byte memoizedIsInitialized = -1;
+ @java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
- if (isInitialized != -1) return isInitialized == 1;
+ if (isInitialized == 1) return true;
+ if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
+ @java.lang.Override
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
- getSerializedSize();
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ if (((bitField0_ & 0x00000001) != 0)) {
output.writeInt32(1, numSelfAndUpstreamTasks_);
}
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ if (((bitField0_ & 0x00000002) != 0)) {
output.writeInt32(2, numSelfAndUpstreamCompletedTasks_);
}
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
+ if (((bitField0_ & 0x00000004) != 0)) {
output.writeInt32(3, withinDagPriority_);
}
- if (((bitField0_ & 0x00000008) == 0x00000008)) {
+ if (((bitField0_ & 0x00000008) != 0)) {
output.writeInt64(4, dagStartTime_);
}
- if (((bitField0_ & 0x00000010) == 0x00000010)) {
+ if (((bitField0_ & 0x00000010) != 0)) {
output.writeInt64(5, firstAttemptStartTime_);
}
- if (((bitField0_ & 0x00000020) == 0x00000020)) {
+ if (((bitField0_ & 0x00000020) != 0)) {
output.writeInt64(6, currentAttemptStartTime_);
}
getUnknownFields().writeTo(output);
}
- private int memoizedSerializedSize = -1;
+ @java.lang.Override
public int getSerializedSize() {
- int size = memoizedSerializedSize;
+ int size = memoizedSize;
if (size != -1) return size;
size = 0;
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ if (((bitField0_ & 0x00000001) != 0)) {
size += com.google.protobuf.CodedOutputStream
.computeInt32Size(1, numSelfAndUpstreamTasks_);
}
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ if (((bitField0_ & 0x00000002) != 0)) {
size += com.google.protobuf.CodedOutputStream
.computeInt32Size(2, numSelfAndUpstreamCompletedTasks_);
}
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
+ if (((bitField0_ & 0x00000004) != 0)) {
size += com.google.protobuf.CodedOutputStream
.computeInt32Size(3, withinDagPriority_);
}
- if (((bitField0_ & 0x00000008) == 0x00000008)) {
+ if (((bitField0_ & 0x00000008) != 0)) {
size += com.google.protobuf.CodedOutputStream
.computeInt64Size(4, dagStartTime_);
}
- if (((bitField0_ & 0x00000010) == 0x00000010)) {
+ if (((bitField0_ & 0x00000010) != 0)) {
size += com.google.protobuf.CodedOutputStream
.computeInt64Size(5, firstAttemptStartTime_);
}
- if (((bitField0_ & 0x00000020) == 0x00000020)) {
+ if (((bitField0_ & 0x00000020) != 0)) {
size += com.google.protobuf.CodedOutputStream
.computeInt64Size(6, currentAttemptStartTime_);
}
size += getUnknownFields().getSerializedSize();
- memoizedSerializedSize = size;
+ memoizedSize = size;
return size;
}
- private static final long serialVersionUID = 0L;
- @java.lang.Override
- protected java.lang.Object writeReplace()
- throws java.io.ObjectStreamException {
- return super.writeReplace();
- }
-
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
@@ -7434,50 +8012,47 @@ public final class LlapDaemonProtocolProtos {
}
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo) obj;
- boolean result = true;
- result = result && (hasNumSelfAndUpstreamTasks() == other.hasNumSelfAndUpstreamTasks());
+ if (hasNumSelfAndUpstreamTasks() != other.hasNumSelfAndUpstreamTasks()) return false;
if (hasNumSelfAndUpstreamTasks()) {
- result = result && (getNumSelfAndUpstreamTasks()
- == other.getNumSelfAndUpstreamTasks());
+ if (getNumSelfAndUpstreamTasks()
+ != other.getNumSelfAndUpstreamTasks()) return false;
}
- result = result && (hasNumSelfAndUpstreamCompletedTasks() == other.hasNumSelfAndUpstreamCompletedTasks());
+ if (hasNumSelfAndUpstreamCompletedTasks() != other.hasNumSelfAndUpstreamCompletedTasks()) return false;
if (hasNumSelfAndUpstreamCompletedTasks()) {
- result = result && (getNumSelfAndUpstreamCompletedTasks()
- == other.getNumSelfAndUpstreamCompletedTasks());
+ if (getNumSelfAndUpstreamCompletedTasks()
+ != other.getNumSelfAndUpstreamCompletedTasks()) return false;
}
- result = result && (hasWithinDagPriority() == other.hasWithinDagPriority());
+ if (hasWithinDagPriority() != other.hasWithinDagPriority()) return false;
if (hasWithinDagPriority()) {
- result = result && (getWithinDagPriority()
- == other.getWithinDagPriority());
+ if (getWithinDagPriority()
+ != other.getWithinDagPriority()) return false;
}
- result = result && (hasDagStartTime() == other.hasDagStartTime());
+ if (hasDagStartTime() != other.hasDagStartTime()) return false;
if (hasDagStartTime()) {
- result = result && (getDagStartTime()
- == other.getDagStartTime());
+ if (getDagStartTime()
+ != other.getDagStartTime()) return false;
}
- result = result && (hasFirstAttemptStartTime() == other.hasFirstAttemptStartTime());
+ if (hasFirstAttemptStartTime() != other.hasFirstAttemptStartTime()) return false;
if (hasFirstAttemptStartTime()) {
- result = result && (getFirstAttemptStartTime()
- == other.getFirstAttemptStartTime());
+ if (getFirstAttemptStartTime()
+ != other.getFirstAttemptStartTime()) return false;
}
- result = result && (hasCurrentAttemptStartTime() == other.hasCurrentAttemptStartTime());
+ if (hasCurrentAttemptStartTime() != other.hasCurrentAttemptStartTime()) return false;
if (hasCurrentAttemptStartTime()) {
- result = result && (getCurrentAttemptStartTime()
- == other.getCurrentAttemptStartTime());
+ if (getCurrentAttemptStartTime()
+ != other.getCurrentAttemptStartTime()) return false;
}
- result = result &&
- getUnknownFields().equals(other.getUnknownFields());
- return result;
+ if (!getUnknownFields().equals(other.getUnknownFields())) return false;
+ return true;
}
- private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
- hash = (19 * hash) + getDescriptorForType().hashCode();
+ hash = (19 * hash) + getDescriptor().hashCode();
if (hasNumSelfAndUpstreamTasks()) {
hash = (37 * hash) + NUM_SELF_AND_UPSTREAM_TASKS_FIELD_NUMBER;
hash = (53 * hash) + getNumSelfAndUpstreamTasks();
@@ -7492,21 +8067,35 @@ public final class LlapDaemonProtocolProtos {
}
if (hasDagStartTime()) {
hash = (37 * hash) + DAG_START_TIME_FIELD_NUMBER;
- hash = (53 * hash) + hashLong(getDagStartTime());
+ hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ getDagStartTime());
}
if (hasFirstAttemptStartTime()) {
hash = (37 * hash) + FIRST_ATTEMPT_START_TIME_FIELD_NUMBER;
- hash = (53 * hash) + hashLong(getFirstAttemptStartTime());
+ hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ getFirstAttemptStartTime());
}
if (hasCurrentAttemptStartTime()) {
hash = (37 * hash) + CURRENT_ATTEMPT_START_TIME_FIELD_NUMBER;
- hash = (53 * hash) + hashLong(getCurrentAttemptStartTime());
+ hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+ getCurrentAttemptStartTime());
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom(
+ java.nio.ByteBuffer data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom(
+ java.nio.ByteBuffer data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
@@ -7530,46 +8119,61 @@ public final class LlapDaemonProtocolProtos {
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
+
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input);
}
+
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
- public static Builder newBuilder() { return Builder.create(); }
+ @java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder() {
+ return DEFAULT_INSTANCE.toBuilder();
+ }
public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo prototype) {
- return newBuilder().mergeFrom(prototype);
+ return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+ }
+ @java.lang.Override
+ public Builder toBuilder() {
+ return this == DEFAULT_INSTANCE
+ ? new Builder() : new Builder().mergeFrom(this);
}
- public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -7577,14 +8181,16 @@ public final class LlapDaemonProtocolProtos {
* Protobuf type {@code FragmentRuntimeInfo}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfoOrBuilder {
+ com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ // @@protoc_insertion_point(builder_implements:FragmentRuntimeInfo)
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentRuntimeInfo_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ @java.lang.Override
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentRuntimeInfo_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -7593,52 +8199,39 @@ public final class LlapDaemonProtocolProtos {
// Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.newBuilder()
private Builder() {
- maybeForceBuilderInitialization();
+
}
private Builder(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
- maybeForceBuilderInitialization();
- }
- private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
- }
- }
- private static Builder create() {
- return new Builder();
- }
+ }
+ @java.lang.Override
public Builder clear() {
super.clear();
+ bitField0_ = 0;
numSelfAndUpstreamTasks_ = 0;
- bitField0_ = (bitField0_ & ~0x00000001);
numSelfAndUpstreamCompletedTasks_ = 0;
- bitField0_ = (bitField0_ & ~0x00000002);
withinDagPriority_ = 0;
- bitField0_ = (bitField0_ & ~0x00000004);
dagStartTime_ = 0L;
- bitField0_ = (bitField0_ & ~0x00000008);
firstAttemptStartTime_ = 0L;
- bitField0_ = (bitField0_ & ~0x00000010);
currentAttemptStartTime_ = 0L;
- bitField0_ = (bitField0_ & ~0x00000020);
return this;
}
- public Builder clone() {
- return create().mergeFrom(buildPartial());
- }
-
+ @java.lang.Override
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentRuntimeInfo_descriptor;
}
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo getDefaultInstanceForType() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.getDefaultInstance();
}
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo build() {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo result = buildPartial();
if (!result.isInitialized()) {
@@ -7647,39 +8240,77 @@ public final class LlapDaemonProtocolProtos {
return result;
}
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo buildPartial() {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo(this);
+ if (bitField0_ != 0) { buildPartial0(result); }
+ onBuilt();
+ return result;
+ }
+
+ private void buildPartial0(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
- if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
+ if (((from_bitField0_ & 0x00000001) != 0)) {
+ result.numSelfAndUpstreamTasks_ = numSelfAndUpstreamTasks_;
to_bitField0_ |= 0x00000001;
}
- result.numSelfAndUpstreamTasks_ = numSelfAndUpstreamTasks_;
- if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
+ if (((from_bitField0_ & 0x00000002) != 0)) {
+ result.numSelfAndUpstreamCompletedTasks_ = numSelfAndUpstreamCompletedTasks_;
to_bitField0_ |= 0x00000002;
}
- result.numSelfAndUpstreamCompletedTasks_ = numSelfAndUpstreamCompletedTasks_;
- if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
+ if (((from_bitField0_ & 0x00000004) != 0)) {
+ result.withinDagPriority_ = withinDagPriority_;
to_bitField0_ |= 0x00000004;
}
- result.withinDagPriority_ = withinDagPriority_;
- if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
+ if (((from_bitField0_ & 0x00000008) != 0)) {
+ result.dagStartTime_ = dagStartTime_;
to_bitField0_ |= 0x00000008;
}
- result.dagStartTime_ = dagStartTime_;
- if (((from_bitField0_ & 0x00000010) == 0x00000010)) {
+ if (((from_bitField0_ & 0x00000010) != 0)) {
+ result.firstAttemptStartTime_ = firstAttemptStartTime_;
to_bitField0_ |= 0x00000010;
}
- result.firstAttemptStartTime_ = firstAttemptStartTime_;
- if (((from_bitField0_ & 0x00000020) == 0x00000020)) {
+ if (((from_bitField0_ & 0x00000020) != 0)) {
+ result.currentAttemptStartTime_ = currentAttemptStartTime_;
to_bitField0_ |= 0x00000020;
}
- result.currentAttemptStartTime_ = currentAttemptStartTime_;
- result.bitField0_ = to_bitField0_;
- onBuilt();
- return result;
+ result.bitField0_ |= to_bitField0_;
}
+ @java.lang.Override
+ public Builder clone() {
+ return super.clone();
+ }
+ @java.lang.Override
+ public Builder setField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ java.lang.Object value) {
+ return super.setField(field, value);
+ }
+ @java.lang.Override
+ public Builder clearField(
+ com.google.protobuf.Descriptors.FieldDescriptor field) {
+ return super.clearField(field);
+ }
+ @java.lang.Override
+ public Builder clearOneof(
+ com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ return super.clearOneof(oneof);
+ }
+ @java.lang.Override
+ public Builder setRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ int index, java.lang.Object value) {
+ return super.setRepeatedField(field, index, value);
+ }
+ @java.lang.Override
+ public Builder addRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ java.lang.Object value) {
+ return super.addRepeatedField(field, value);
+ }
+ @java.lang.Override
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo) {
return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo)other);
@@ -7710,57 +8341,110 @@ public final class LlapDaemonProtocolProtos {
setCurrentAttemptStartTime(other.getCurrentAttemptStartTime());
}
this.mergeUnknownFields(other.getUnknownFields());
+ onChanged();
return this;
}
+ @java.lang.Override
public final boolean isInitialized() {
return true;
}
+ @java.lang.Override
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parsedMessage = null;
+ if (extensionRegistry == null) {
+ throw new java.lang.NullPointerException();
+ }
try {
- parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ case 8: {
+ numSelfAndUpstreamTasks_ = input.readInt32();
+ bitField0_ |= 0x00000001;
+ break;
+ } // case 8
+ case 16: {
+ numSelfAndUpstreamCompletedTasks_ = input.readInt32();
+ bitField0_ |= 0x00000002;
+ break;
+ } // case 16
+ case 24: {
+ withinDagPriority_ = input.readInt32();
+ bitField0_ |= 0x00000004;
+ break;
+ } // case 24
+ case 32: {
+ dagStartTime_ = input.readInt64();
+ bitField0_ |= 0x00000008;
+ break;
+ } // case 32
+ case 40: {
+ firstAttemptStartTime_ = input.readInt64();
+ bitField0_ |= 0x00000010;
+ break;
+ } // case 40
+ case 48: {
+ currentAttemptStartTime_ = input.readInt64();
+ bitField0_ |= 0x00000020;
+ break;
+ } // case 48
+ default: {
+ if (!super.parseUnknownField(input, extensionRegistry, tag)) {
+ done = true; // was an endgroup tag
+ }
+ break;
+ } // default:
+ } // switch (tag)
+ } // while (!done)
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
- parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo) e.getUnfinishedMessage();
- throw e;
+ throw e.unwrapIOException();
} finally {
- if (parsedMessage != null) {
- mergeFrom(parsedMessage);
- }
- }
+ onChanged();
+ } // finally
return this;
}
private int bitField0_;
- // optional int32 num_self_and_upstream_tasks = 1;
private int numSelfAndUpstreamTasks_ ;
/**
* <code>optional int32 num_self_and_upstream_tasks = 1;</code>
+ * @return Whether the numSelfAndUpstreamTasks field is set.
*/
+ @java.lang.Override
public boolean hasNumSelfAndUpstreamTasks() {
- return ((bitField0_ & 0x00000001) == 0x00000001);
+ return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional int32 num_self_and_upstream_tasks = 1;</code>
+ * @return The numSelfAndUpstreamTasks.
*/
+ @java.lang.Override
public int getNumSelfAndUpstreamTasks() {
return numSelfAndUpstreamTasks_;
}
/**
* <code>optional int32 num_self_and_upstream_tasks = 1;</code>
+ * @param value The numSelfAndUpstreamTasks to set.
+ * @return This builder for chaining.
*/
public Builder setNumSelfAndUpstreamTasks(int value) {
- bitField0_ |= 0x00000001;
+
numSelfAndUpstreamTasks_ = value;
+ bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional int32 num_self_and_upstream_tasks = 1;</code>
+ * @return This builder for chaining.
*/
public Builder clearNumSelfAndUpstreamTasks() {
bitField0_ = (bitField0_ & ~0x00000001);
@@ -7769,31 +8453,38 @@ public final class LlapDaemonProtocolProtos {
return this;
}
- // optional int32 num_self_and_upstream_completed_tasks = 2;
private int numSelfAndUpstreamCompletedTasks_ ;
/**
* <code>optional int32 num_self_and_upstream_completed_tasks = 2;</code>
+ * @return Whether the numSelfAndUpstreamCompletedTasks field is set.
*/
+ @java.lang.Override
public boolean hasNumSelfAndUpstreamCompletedTasks() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
+ return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional int32 num_self_and_upstream_completed_tasks = 2;</code>
+ * @return The numSelfAndUpstreamCompletedTasks.
*/
+ @java.lang.Override
public int getNumSelfAndUpstreamCompletedTasks() {
return numSelfAndUpstreamCompletedTasks_;
}
/**
* <code>optional int32 num_self_and_upstream_completed_tasks = 2;</code>
+ * @param value The numSelfAndUpstreamCompletedTasks to set.
+ * @return This builder for chaining.
*/
public Builder setNumSelfAndUpstreamCompletedTasks(int value) {
- bitField0_ |= 0x00000002;
+
numSelfAndUpstreamCompletedTasks_ = value;
+ bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* <code>optional int32 num_self_and_upstream_completed_tasks = 2;</code>
+ * @return This builder for chaining.
*/
public Builder clearNumSelfAndUpstreamCompletedTasks() {
bitField0_ = (bitField0_ & ~0x00000002);
@@ -7802,31 +8493,38 @@ public final class LlapDaemonProtocolProtos {
return this;
}
- // optional int32 within_dag_priority = 3;
private int withinDagPriority_ ;
/**
* <code>optional int32 within_dag_priority = 3;</code>
+ * @return Whether the withinDagPriority field is set.
*/
+ @java.lang.Override
public boolean hasWithinDagPriority() {
- return ((bitField0_ & 0x00000004) == 0x00000004);
+ return ((bitField0_ & 0x00000004) != 0);
}
/**
* <code>optional int32 within_dag_priority = 3;</code>
+ * @return The withinDagPriority.
*/
+ @java.lang.Override
public int getWithinDagPriority() {
return withinDagPriority_;
}
/**
* <code>optional int32 within_dag_priority = 3;</code>
+ * @param value The withinDagPriority to set.
+ * @return This builder for chaining.
*/
public Builder setWithinDagPriority(int value) {
- bitField0_ |= 0x00000004;
+
withinDagPriority_ = value;
+ bitField0_ |= 0x00000004;
onChanged();
return this;
}
/**
* <code>optional int32 within_dag_priority = 3;</code>
+ * @return This builder for chaining.
*/
public Builder clearWithinDagPriority() {
bitField0_ = (bitField0_ & ~0x00000004);
@@ -7835,31 +8533,38 @@ public final class LlapDaemonProtocolProtos {
return this;
}
- // optional int64 dag_start_time = 4;
private long dagStartTime_ ;
/**
* <code>optional int64 dag_start_time = 4;</code>
+ * @return Whether the dagStartTime field is set.
*/
+ @java.lang.Override
public boolean hasDagStartTime() {
- return ((bitField0_ & 0x00000008) == 0x00000008);
+ return ((bitField0_ & 0x00000008) != 0);
}
/**
* <code>optional int64 dag_start_time = 4;</code>
+ * @return The dagStartTime.
*/
+ @java.lang.Override
public long getDagStartTime() {
return dagStartTime_;
}
/**
* <code>optional int64 dag_start_time = 4;</code>
+ * @param value The dagStartTime to set.
+ * @return This builder for chaining.
*/
public Builder setDagStartTime(long value) {
- bitField0_ |= 0x00000008;
+
dagStartTime_ = value;
+ bitField0_ |= 0x00000008;
onChanged();
return this;
}
/**
* <code>optional int64 dag_start_time = 4;</code>
+ * @return This builder for chaining.
*/
public Builder clearDagStartTime() {
bitField0_ = (bitField0_ & ~0x00000008);
@@ -7868,31 +8573,38 @@ public final class LlapDaemonProtocolProtos {
return this;
}
- // optional int64 first_attempt_start_time = 5;
private long firstAttemptStartTime_ ;
/**
* <code>optional int64 first_attempt_start_time = 5;</code>
+ * @return Whether the firstAttemptStartTime field is set.
*/
+ @java.lang.Override
public boolean hasFirstAttemptStartTime() {
- return ((bitField0_ & 0x00000010) == 0x00000010);
+ return ((bitField0_ & 0x00000010) != 0);
}
/**
* <code>optional int64 first_attempt_start_time = 5;</code>
+ * @return The firstAttemptStartTime.
*/
+ @java.lang.Override
public long getFirstAttemptStartTime() {
return firstAttemptStartTime_;
}
/**
* <code>optional int64 first_attempt_start_time = 5;</code>
+ * @param value The firstAttemptStartTime to set.
+ * @return This builder for chaining.
*/
public Builder setFirstAttemptStartTime(long value) {
- bitField0_ |= 0x00000010;
+
firstAttemptStartTime_ = value;
+ bitField0_ |= 0x00000010;
onChanged();
return this;
}
/**
* <code>optional int64 first_attempt_start_time = 5;</code>
+ * @return This builder for chaining.
*/
public Builder clearFirstAttemptStartTime() {
bitField0_ = (bitField0_ & ~0x00000010);
@@ -7901,31 +8613,38 @@ public final class LlapDaemonProtocolProtos {
return this;
}
- // optional int64 current_attempt_start_time = 6;
private long currentAttemptStartTime_ ;
/**
* <code>optional int64 current_attempt_start_time = 6;</code>
+ * @return Whether the currentAttemptStartTime field is set.
*/
+ @java.lang.Override
public boolean hasCurrentAttemptStartTime() {
- return ((bitField0_ & 0x00000020) == 0x00000020);
+ return ((bitField0_ & 0x00000020) != 0);
}
/**
* <code>optional int64 current_attempt_start_time = 6;</code>
+ * @return The currentAttemptStartTime.
*/
+ @java.lang.Override
public long getCurrentAttemptStartTime() {
return currentAttemptStartTime_;
}
/**
* <code>optional int64 current_attempt_start_time = 6;</code>
+ * @param value The currentAttemptStartTime to set.
+ * @return This builder for chaining.
*/
public Builder setCurrentAttemptStartTime(long value) {
- bitField0_ |= 0x00000020;
+
currentAttemptStartTime_ = value;
+ bitField0_ |= 0x00000020;
onChanged();
return this;
}
/**
* <code>optional int64 current_attempt_start_time = 6;</code>
+ * @return This builder for chaining.
*/
public Builder clearCurrentAttemptStartTime() {
bitField0_ = (bitField0_ & ~0x00000020);
@@ -7933,53 +8652,110 @@ public final class LlapDaemonProtocolProtos {
onChanged();
return this;
}
+ @java.lang.Override
+ public final Builder setUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.setUnknownFields(unknownFields);
+ }
+
+ @java.lang.Override
+ public final Builder mergeUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.mergeUnknownFields(unknownFields);
+ }
+
// @@protoc_insertion_point(builder_scope:FragmentRuntimeInfo)
}
+ // @@protoc_insertion_point(class_scope:FragmentRuntimeInfo)
+ private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo DEFAULT_INSTANCE;
static {
- defaultInstance = new FragmentRuntimeInfo(true);
- defaultInstance.initFields();
+ DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo();
+ }
+
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo getDefaultInstance() {
+ return DEFAULT_INSTANCE;
+ }
+
+ @java.lang.Deprecated public static final com.google.protobuf.Parser<FragmentRuntimeInfo>
+ PARSER = new com.google.protobuf.AbstractParser<FragmentRuntimeInfo>() {
+ @java.lang.Override
+ public FragmentRuntimeInfo parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ Builder builder = newBuilder();
+ try {
+ builder.mergeFrom(input, extensionRegistry);
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(builder.buildPartial());
+ } catch (com.google.protobuf.UninitializedMessageException e) {
+ throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(e)
+ .setUnfinishedMessage(builder.buildPartial());
+ }
+ return builder.buildPartial();
+ }
+ };
+
+ public static com.google.protobuf.Parser<FragmentRuntimeInfo> parser() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<FragmentRuntimeInfo> getParserForType() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo getDefaultInstanceForType() {
+ return DEFAULT_INSTANCE;
}
- // @@protoc_insertion_point(class_scope:FragmentRuntimeInfo)
}
- public interface QueryIdentifierProtoOrBuilder
- extends com.google.protobuf.MessageOrBuilder {
+ public interface QueryIdentifierProtoOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:QueryIdentifierProto)
+ com.google.protobuf.MessageOrBuilder {
- // optional string application_id_string = 1;
/**
* <code>optional string application_id_string = 1;</code>
+ * @return Whether the applicationIdString field is set.
*/
boolean hasApplicationIdString();
/**
* <code>optional string application_id_string = 1;</code>
+ * @return The applicationIdString.
*/
java.lang.String getApplicationIdString();
/**
* <code>optional string application_id_string = 1;</code>
+ * @return The bytes for applicationIdString.
*/
com.google.protobuf.ByteString
getApplicationIdStringBytes();
- // optional int32 dag_index = 2;
/**
* <code>optional int32 dag_index = 2;</code>
+ * @return Whether the dagIndex field is set.
*/
boolean hasDagIndex();
/**
* <code>optional int32 dag_index = 2;</code>
+ * @return The dagIndex.
*/
int getDagIndex();
- // optional int32 app_attempt_number = 3;
/**
* <code>optional int32 app_attempt_number = 3;</code>
+ * @return Whether the appAttemptNumber field is set.
*/
boolean hasAppAttemptNumber();
/**
* <code>optional int32 app_attempt_number = 3;</code>
+ * @return The appAttemptNumber.
*/
int getAppAttemptNumber();
}
@@ -7987,120 +8763,55 @@ public final class LlapDaemonProtocolProtos {
* Protobuf type {@code QueryIdentifierProto}
*/
public static final class QueryIdentifierProto extends
- com.google.protobuf.GeneratedMessage
- implements QueryIdentifierProtoOrBuilder {
+ com.google.protobuf.GeneratedMessageV3 implements
+ // @@protoc_insertion_point(message_implements:QueryIdentifierProto)
+ QueryIdentifierProtoOrBuilder {
+ private static final long serialVersionUID = 0L;
// Use QueryIdentifierProto.newBuilder() to construct.
- private QueryIdentifierProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ private QueryIdentifierProto(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
- this.unknownFields = builder.getUnknownFields();
- }
- private QueryIdentifierProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
- private static final QueryIdentifierProto defaultInstance;
- public static QueryIdentifierProto getDefaultInstance() {
- return defaultInstance;
}
-
- public QueryIdentifierProto getDefaultInstanceForType() {
- return defaultInstance;
+ private QueryIdentifierProto() {
+ applicationIdString_ = "";
}
- private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
- return this.unknownFields;
- }
- private QueryIdentifierProto(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- initFields();
- int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
- try {
- boolean done = false;
- while (!done) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
- done = true;
- break;
- default: {
- if (!parseUnknownField(input, unknownFields,
- extensionRegistry, tag)) {
- done = true;
- }
- break;
- }
- case 10: {
- bitField0_ |= 0x00000001;
- applicationIdString_ = input.readBytes();
- break;
- }
- case 16: {
- bitField0_ |= 0x00000002;
- dagIndex_ = input.readInt32();
- break;
- }
- case 24: {
- bitField0_ |= 0x00000004;
- appAttemptNumber_ = input.readInt32();
- break;
- }
- }
- }
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
- throw e.setUnfinishedMessage(this);
- } catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
- e.getMessage()).setUnfinishedMessage(this);
- } finally {
- this.unknownFields = unknownFields.build();
- makeExtensionsImmutable();
- }
+ @SuppressWarnings({"unused"})
+ protected java.lang.Object newInstance(
+ UnusedPrivateParameter unused) {
+ return new QueryIdentifierProto();
}
+
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ @java.lang.Override
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder.class);
}
- public static com.google.protobuf.Parser<QueryIdentifierProto> PARSER =
- new com.google.protobuf.AbstractParser<QueryIdentifierProto>() {
- public QueryIdentifierProto parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return new QueryIdentifierProto(input, extensionRegistry);
- }
- };
-
- @java.lang.Override
- public com.google.protobuf.Parser<QueryIdentifierProto> getParserForType() {
- return PARSER;
- }
-
private int bitField0_;
- // optional string application_id_string = 1;
public static final int APPLICATION_ID_STRING_FIELD_NUMBER = 1;
- private java.lang.Object applicationIdString_;
+ @SuppressWarnings("serial")
+ private volatile java.lang.Object applicationIdString_ = "";
/**
* <code>optional string application_id_string = 1;</code>
+ * @return Whether the applicationIdString field is set.
*/
+ @java.lang.Override
public boolean hasApplicationIdString() {
- return ((bitField0_ & 0x00000001) == 0x00000001);
+ return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string application_id_string = 1;</code>
+ * @return The applicationIdString.
*/
+ @java.lang.Override
public java.lang.String getApplicationIdString() {
java.lang.Object ref = applicationIdString_;
if (ref instanceof java.lang.String) {
@@ -8117,7 +8828,9 @@ public final class LlapDaemonProtocolProtos {
}
/**
* <code>optional string application_id_string = 1;</code>
+ * @return The bytes for applicationIdString.
*/
+ @java.lang.Override
public com.google.protobuf.ByteString
getApplicationIdStringBytes() {
java.lang.Object ref = applicationIdString_;
@@ -8132,97 +8845,92 @@ public final class LlapDaemonProtocolProtos {
}
}
- // optional int32 dag_index = 2;
public static final int DAG_INDEX_FIELD_NUMBER = 2;
- private int dagIndex_;
+ private int dagIndex_ = 0;
/**
* <code>optional int32 dag_index = 2;</code>
+ * @return Whether the dagIndex field is set.
*/
+ @java.lang.Override
public boolean hasDagIndex() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
+ return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional int32 dag_index = 2;</code>
+ * @return The dagIndex.
*/
+ @java.lang.Override
public int getDagIndex() {
return dagIndex_;
}
- // optional int32 app_attempt_number = 3;
public static final int APP_ATTEMPT_NUMBER_FIELD_NUMBER = 3;
- private int appAttemptNumber_;
+ private int appAttemptNumber_ = 0;
/**
* <code>optional int32 app_attempt_number = 3;</code>
+ * @return Whether the appAttemptNumber field is set.
*/
+ @java.lang.Override
public boolean hasAppAttemptNumber() {
- return ((bitField0_ & 0x00000004) == 0x00000004);
+ return ((bitField0_ & 0x00000004) != 0);
}
/**
* <code>optional int32 app_attempt_number = 3;</code>
+ * @return The appAttemptNumber.
*/
+ @java.lang.Override
public int getAppAttemptNumber() {
return appAttemptNumber_;
}
- private void initFields() {
- applicationIdString_ = "";
- dagIndex_ = 0;
- appAttemptNumber_ = 0;
- }
private byte memoizedIsInitialized = -1;
+ @java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
- if (isInitialized != -1) return isInitialized == 1;
+ if (isInitialized == 1) return true;
+ if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
+ @java.lang.Override
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
- getSerializedSize();
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
- output.writeBytes(1, getApplicationIdStringBytes());
+ if (((bitField0_ & 0x00000001) != 0)) {
+ com.google.protobuf.GeneratedMessageV3.writeString(output, 1, applicationIdString_);
}
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ if (((bitField0_ & 0x00000002) != 0)) {
output.writeInt32(2, dagIndex_);
}
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
+ if (((bitField0_ & 0x00000004) != 0)) {
output.writeInt32(3, appAttemptNumber_);
}
getUnknownFields().writeTo(output);
}
- private int memoizedSerializedSize = -1;
+ @java.lang.Override
public int getSerializedSize() {
- int size = memoizedSerializedSize;
+ int size = memoizedSize;
if (size != -1) return size;
size = 0;
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
- .computeBytesSize(1, getApplicationIdStringBytes());
+ if (((bitField0_ & 0x00000001) != 0)) {
+ size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, applicationIdString_);
}
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ if (((bitField0_ & 0x00000002) != 0)) {
size += com.google.protobuf.CodedOutputStream
.computeInt32Size(2, dagIndex_);
}
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
+ if (((bitField0_ & 0x00000004) != 0)) {
size += com.google.protobuf.CodedOutputStream
.computeInt32Size(3, appAttemptNumber_);
}
size += getUnknownFields().getSerializedSize();
- memoizedSerializedSize = size;
+ memoizedSize = size;
return size;
}
- private static final long serialVersionUID = 0L;
- @java.lang.Override
- protected java.lang.Object writeReplace()
- throws java.io.ObjectStreamException {
- return super.writeReplace();
- }
-
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
@@ -8233,35 +8941,32 @@ public final class LlapDaemonProtocolProtos {
}
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto) obj;
- boolean result = true;
- result = result && (hasApplicationIdString() == other.hasApplicationIdString());
+ if (hasApplicationIdString() != other.hasApplicationIdString()) return false;
if (hasApplicationIdString()) {
- result = result && getApplicationIdString()
- .equals(other.getApplicationIdString());
+ if (!getApplicationIdString()
+ .equals(other.getApplicationIdString())) return false;
}
- result = result && (hasDagIndex() == other.hasDagIndex());
+ if (hasDagIndex() != other.hasDagIndex()) return false;
if (hasDagIndex()) {
- result = result && (getDagIndex()
- == other.getDagIndex());
+ if (getDagIndex()
+ != other.getDagIndex()) return false;
}
- result = result && (hasAppAttemptNumber() == other.hasAppAttemptNumber());
+ if (hasAppAttemptNumber() != other.hasAppAttemptNumber()) return false;
if (hasAppAttemptNumber()) {
- result = result && (getAppAttemptNumber()
- == other.getAppAttemptNumber());
+ if (getAppAttemptNumber()
+ != other.getAppAttemptNumber()) return false;
}
- result = result &&
- getUnknownFields().equals(other.getUnknownFields());
- return result;
+ if (!getUnknownFields().equals(other.getUnknownFields())) return false;
+ return true;
}
- private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
- hash = (19 * hash) + getDescriptorForType().hashCode();
+ hash = (19 * hash) + getDescriptor().hashCode();
if (hasApplicationIdString()) {
hash = (37 * hash) + APPLICATION_ID_STRING_FIELD_NUMBER;
hash = (53 * hash) + getApplicationIdString().hashCode();
@@ -8279,6 +8984,17 @@ public final class LlapDaemonProtocolProtos {
return hash;
}
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(
+ java.nio.ByteBuffer data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(
+ java.nio.ByteBuffer data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
@@ -8302,46 +9018,61 @@ public final class LlapDaemonProtocolProtos {
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
+
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input);
}
+
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
- public static Builder newBuilder() { return Builder.create(); }
+ @java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder() {
+ return DEFAULT_INSTANCE.toBuilder();
+ }
public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto prototype) {
- return newBuilder().mergeFrom(prototype);
+ return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+ }
+ @java.lang.Override
+ public Builder toBuilder() {
+ return this == DEFAULT_INSTANCE
+ ? new Builder() : new Builder().mergeFrom(this);
}
- public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
@@ -8349,14 +9080,16 @@ public final class LlapDaemonProtocolProtos {
* Protobuf type {@code QueryIdentifierProto}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder {
+ com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ // @@protoc_insertion_point(builder_implements:QueryIdentifierProto)
+ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ @java.lang.Override
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -8365,46 +9098,36 @@ public final class LlapDaemonProtocolProtos {
// Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.newBuilder()
private Builder() {
- maybeForceBuilderInitialization();
+
}
private Builder(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
- maybeForceBuilderInitialization();
- }
- private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
- }
- }
- private static Builder create() {
- return new Builder();
- }
+ }
+ @java.lang.Override
public Builder clear() {
super.clear();
+ bitField0_ = 0;
applicationIdString_ = "";
- bitField0_ = (bitField0_ & ~0x00000001);
dagIndex_ = 0;
- bitField0_ = (bitField0_ & ~0x00000002);
appAttemptNumber_ = 0;
- bitField0_ = (bitField0_ & ~0x00000004);
return this;
}
- public Builder clone() {
- return create().mergeFrom(buildPartial());
- }
-
+ @java.lang.Override
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_descriptor;
}
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getDefaultInstanceForType() {
return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance();
}
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto build() {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto result = buildPartial();
if (!result.isInitialized()) {
@@ -8413,27 +9136,65 @@ public final class LlapDaemonProtocolProtos {
return result;
}
+ @java.lang.Override
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto buildPartial() {
org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto(this);
+ if (bitField0_ != 0) { buildPartial0(result); }
+ onBuilt();
... 38296 lines suppressed ...