You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by ay...@apache.org on 2024/03/19 06:53:18 UTC

(hive) branch master updated: HIVE-28123: Add Generated Protobuf code for 3.24.4 Upgrade (#5132). (Indhumathi Muthumurugesh, reviewed by Ayush Saxena, Tanishq Chugh)

This is an automated email from the ASF dual-hosted git repository.

ayushsaxena pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hive.git


The following commit(s) were added to refs/heads/master by this push:
     new 5256d3c2669 HIVE-28123: Add Generated Protobuf code for 3.24.4 Upgrade (#5132). (Indhumathi Muthumurugesh, reviewed by Ayush Saxena, Tanishq Chugh)
5256d3c2669 is described below

commit 5256d3c2669fb523913dd79c8878c3fa94abdbde
Author: Indhumathi27 <34...@users.noreply.github.com>
AuthorDate: Tue Mar 19 12:23:12 2024 +0530

    HIVE-28123: Add Generated Protobuf code for 3.24.4 Upgrade (#5132). (Indhumathi Muthumurugesh, reviewed by Ayush Saxena, Tanishq Chugh)
---
 .../llap/daemon/rpc/LlapDaemonProtocolProtos.java  | 21594 +++++++++++--------
 .../llap/plugin/rpc/LlapPluginProtocolProtos.java  |   749 +-
 .../hadoop/hive/ql/hooks/proto/HiveHookEvents.java |  1908 +-
 .../hadoop/hive/ql/io/protobuf/SampleProtos.java   |  2955 ++-
 4 files changed, 15983 insertions(+), 11223 deletions(-)

diff --git a/llap-common/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java b/llap-common/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java
index b50448053bf..dafa1e7a78d 100644
--- a/llap-common/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java
+++ b/llap-common/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java
@@ -5,8 +5,14 @@ package org.apache.hadoop.hive.llap.daemon.rpc;
 
 public final class LlapDaemonProtocolProtos {
   private LlapDaemonProtocolProtos() {}
+  public static void registerAllExtensions(
+      com.google.protobuf.ExtensionRegistryLite registry) {
+  }
+
   public static void registerAllExtensions(
       com.google.protobuf.ExtensionRegistry registry) {
+    registerAllExtensions(
+        (com.google.protobuf.ExtensionRegistryLite) registry);
   }
   /**
    * Protobuf enum {@code SourceStateProto}
@@ -16,11 +22,11 @@ public final class LlapDaemonProtocolProtos {
     /**
      * <code>S_SUCCEEDED = 1;</code>
      */
-    S_SUCCEEDED(0, 1),
+    S_SUCCEEDED(1),
     /**
      * <code>S_RUNNING = 2;</code>
      */
-    S_RUNNING(1, 2),
+    S_RUNNING(2),
     ;
 
     /**
@@ -33,9 +39,25 @@ public final class LlapDaemonProtocolProtos {
     public static final int S_RUNNING_VALUE = 2;
 
 
-    public final int getNumber() { return value; }
+    public final int getNumber() {
+      return value;
+    }
 
+    /**
+     * @param value The numeric wire value of the corresponding enum entry.
+     * @return The enum associated with the given numeric wire value.
+     * @deprecated Use {@link #forNumber(int)} instead.
+     */
+    @java.lang.Deprecated
     public static SourceStateProto valueOf(int value) {
+      return forNumber(value);
+    }
+
+    /**
+     * @param value The numeric wire value of the corresponding enum entry.
+     * @return The enum associated with the given numeric wire value.
+     */
+    public static SourceStateProto forNumber(int value) {
       switch (value) {
         case 1: return S_SUCCEEDED;
         case 2: return S_RUNNING;
@@ -47,17 +69,17 @@ public final class LlapDaemonProtocolProtos {
         internalGetValueMap() {
       return internalValueMap;
     }
-    private static com.google.protobuf.Internal.EnumLiteMap<SourceStateProto>
-        internalValueMap =
+    private static final com.google.protobuf.Internal.EnumLiteMap<
+        SourceStateProto> internalValueMap =
           new com.google.protobuf.Internal.EnumLiteMap<SourceStateProto>() {
             public SourceStateProto findValueByNumber(int number) {
-              return SourceStateProto.valueOf(number);
+              return SourceStateProto.forNumber(number);
             }
           };
 
     public final com.google.protobuf.Descriptors.EnumValueDescriptor
         getValueDescriptor() {
-      return getDescriptor().getValues().get(index);
+      return getDescriptor().getValues().get(ordinal());
     }
     public final com.google.protobuf.Descriptors.EnumDescriptor
         getDescriptorForType() {
@@ -79,11 +101,9 @@ public final class LlapDaemonProtocolProtos {
       return VALUES[desc.getIndex()];
     }
 
-    private final int index;
     private final int value;
 
-    private SourceStateProto(int index, int value) {
-      this.index = index;
+    private SourceStateProto(int value) {
       this.value = value;
     }
 
@@ -98,15 +118,15 @@ public final class LlapDaemonProtocolProtos {
     /**
      * <code>ACCEPTED = 1;</code>
      */
-    ACCEPTED(0, 1),
+    ACCEPTED(1),
     /**
      * <code>REJECTED = 2;</code>
      */
-    REJECTED(1, 2),
+    REJECTED(2),
     /**
      * <code>EVICTED_OTHER = 3;</code>
      */
-    EVICTED_OTHER(2, 3),
+    EVICTED_OTHER(3),
     ;
 
     /**
@@ -123,9 +143,25 @@ public final class LlapDaemonProtocolProtos {
     public static final int EVICTED_OTHER_VALUE = 3;
 
 
-    public final int getNumber() { return value; }
+    public final int getNumber() {
+      return value;
+    }
 
+    /**
+     * @param value The numeric wire value of the corresponding enum entry.
+     * @return The enum associated with the given numeric wire value.
+     * @deprecated Use {@link #forNumber(int)} instead.
+     */
+    @java.lang.Deprecated
     public static SubmissionStateProto valueOf(int value) {
+      return forNumber(value);
+    }
+
+    /**
+     * @param value The numeric wire value of the corresponding enum entry.
+     * @return The enum associated with the given numeric wire value.
+     */
+    public static SubmissionStateProto forNumber(int value) {
       switch (value) {
         case 1: return ACCEPTED;
         case 2: return REJECTED;
@@ -138,17 +174,17 @@ public final class LlapDaemonProtocolProtos {
         internalGetValueMap() {
       return internalValueMap;
     }
-    private static com.google.protobuf.Internal.EnumLiteMap<SubmissionStateProto>
-        internalValueMap =
+    private static final com.google.protobuf.Internal.EnumLiteMap<
+        SubmissionStateProto> internalValueMap =
           new com.google.protobuf.Internal.EnumLiteMap<SubmissionStateProto>() {
             public SubmissionStateProto findValueByNumber(int number) {
-              return SubmissionStateProto.valueOf(number);
+              return SubmissionStateProto.forNumber(number);
             }
           };
 
     public final com.google.protobuf.Descriptors.EnumValueDescriptor
         getValueDescriptor() {
-      return getDescriptor().getValues().get(index);
+      return getDescriptor().getValues().get(ordinal());
     }
     public final com.google.protobuf.Descriptors.EnumDescriptor
         getDescriptorForType() {
@@ -170,37 +206,38 @@ public final class LlapDaemonProtocolProtos {
       return VALUES[desc.getIndex()];
     }
 
-    private final int index;
     private final int value;
 
-    private SubmissionStateProto(int index, int value) {
-      this.index = index;
+    private SubmissionStateProto(int value) {
       this.value = value;
     }
 
     // @@protoc_insertion_point(enum_scope:SubmissionStateProto)
   }
 
-  public interface UserPayloadProtoOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
+  public interface UserPayloadProtoOrBuilder extends
+      // @@protoc_insertion_point(interface_extends:UserPayloadProto)
+      com.google.protobuf.MessageOrBuilder {
 
-    // optional bytes user_payload = 1;
     /**
      * <code>optional bytes user_payload = 1;</code>
+     * @return Whether the userPayload field is set.
      */
     boolean hasUserPayload();
     /**
      * <code>optional bytes user_payload = 1;</code>
+     * @return The userPayload.
      */
     com.google.protobuf.ByteString getUserPayload();
 
-    // optional int32 version = 2;
     /**
      * <code>optional int32 version = 2;</code>
+     * @return Whether the version field is set.
      */
     boolean hasVersion();
     /**
      * <code>optional int32 version = 2;</code>
+     * @return The version.
      */
     int getVersion();
   }
@@ -208,186 +245,119 @@ public final class LlapDaemonProtocolProtos {
    * Protobuf type {@code UserPayloadProto}
    */
   public static final class UserPayloadProto extends
-      com.google.protobuf.GeneratedMessage
-      implements UserPayloadProtoOrBuilder {
+      com.google.protobuf.GeneratedMessageV3 implements
+      // @@protoc_insertion_point(message_implements:UserPayloadProto)
+      UserPayloadProtoOrBuilder {
+  private static final long serialVersionUID = 0L;
     // Use UserPayloadProto.newBuilder() to construct.
-    private UserPayloadProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+    private UserPayloadProto(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
       super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private UserPayloadProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final UserPayloadProto defaultInstance;
-    public static UserPayloadProto getDefaultInstance() {
-      return defaultInstance;
     }
-
-    public UserPayloadProto getDefaultInstanceForType() {
-      return defaultInstance;
+    private UserPayloadProto() {
+      userPayload_ = com.google.protobuf.ByteString.EMPTY;
     }
 
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
     @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private UserPayloadProto(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              bitField0_ |= 0x00000001;
-              userPayload_ = input.readBytes();
-              break;
-            }
-            case 16: {
-              bitField0_ |= 0x00000002;
-              version_ = input.readInt32();
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
+    @SuppressWarnings({"unused"})
+    protected java.lang.Object newInstance(
+        UnusedPrivateParameter unused) {
+      return new UserPayloadProto();
     }
+
     public static final com.google.protobuf.Descriptors.Descriptor
         getDescriptor() {
       return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_descriptor;
     }
 
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+    @java.lang.Override
+    protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
         internalGetFieldAccessorTable() {
       return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_fieldAccessorTable
           .ensureFieldAccessorsInitialized(
               org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder.class);
     }
 
-    public static com.google.protobuf.Parser<UserPayloadProto> PARSER =
-        new com.google.protobuf.AbstractParser<UserPayloadProto>() {
-      public UserPayloadProto parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new UserPayloadProto(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<UserPayloadProto> getParserForType() {
-      return PARSER;
-    }
-
     private int bitField0_;
-    // optional bytes user_payload = 1;
     public static final int USER_PAYLOAD_FIELD_NUMBER = 1;
-    private com.google.protobuf.ByteString userPayload_;
+    private com.google.protobuf.ByteString userPayload_ = com.google.protobuf.ByteString.EMPTY;
     /**
      * <code>optional bytes user_payload = 1;</code>
+     * @return Whether the userPayload field is set.
      */
+    @java.lang.Override
     public boolean hasUserPayload() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
+      return ((bitField0_ & 0x00000001) != 0);
     }
     /**
      * <code>optional bytes user_payload = 1;</code>
+     * @return The userPayload.
      */
+    @java.lang.Override
     public com.google.protobuf.ByteString getUserPayload() {
       return userPayload_;
     }
 
-    // optional int32 version = 2;
     public static final int VERSION_FIELD_NUMBER = 2;
-    private int version_;
+    private int version_ = 0;
     /**
      * <code>optional int32 version = 2;</code>
+     * @return Whether the version field is set.
      */
+    @java.lang.Override
     public boolean hasVersion() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
+      return ((bitField0_ & 0x00000002) != 0);
     }
     /**
      * <code>optional int32 version = 2;</code>
+     * @return The version.
      */
+    @java.lang.Override
     public int getVersion() {
       return version_;
     }
 
-    private void initFields() {
-      userPayload_ = com.google.protobuf.ByteString.EMPTY;
-      version_ = 0;
-    }
     private byte memoizedIsInitialized = -1;
+    @java.lang.Override
     public final boolean isInitialized() {
       byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
+      if (isInitialized == 1) return true;
+      if (isInitialized == 0) return false;
 
       memoizedIsInitialized = 1;
       return true;
     }
 
+    @java.lang.Override
     public void writeTo(com.google.protobuf.CodedOutputStream output)
                         throws java.io.IOException {
-      getSerializedSize();
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
+      if (((bitField0_ & 0x00000001) != 0)) {
         output.writeBytes(1, userPayload_);
       }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
+      if (((bitField0_ & 0x00000002) != 0)) {
         output.writeInt32(2, version_);
       }
       getUnknownFields().writeTo(output);
     }
 
-    private int memoizedSerializedSize = -1;
+    @java.lang.Override
     public int getSerializedSize() {
-      int size = memoizedSerializedSize;
+      int size = memoizedSize;
       if (size != -1) return size;
 
       size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
+      if (((bitField0_ & 0x00000001) != 0)) {
         size += com.google.protobuf.CodedOutputStream
           .computeBytesSize(1, userPayload_);
       }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
+      if (((bitField0_ & 0x00000002) != 0)) {
         size += com.google.protobuf.CodedOutputStream
           .computeInt32Size(2, version_);
       }
       size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
+      memoizedSize = size;
       return size;
     }
 
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
     @java.lang.Override
     public boolean equals(final java.lang.Object obj) {
       if (obj == this) {
@@ -398,30 +368,27 @@ public final class LlapDaemonProtocolProtos {
       }
       org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto) obj;
 
-      boolean result = true;
-      result = result && (hasUserPayload() == other.hasUserPayload());
+      if (hasUserPayload() != other.hasUserPayload()) return false;
       if (hasUserPayload()) {
-        result = result && getUserPayload()
-            .equals(other.getUserPayload());
+        if (!getUserPayload()
+            .equals(other.getUserPayload())) return false;
       }
-      result = result && (hasVersion() == other.hasVersion());
+      if (hasVersion() != other.hasVersion()) return false;
       if (hasVersion()) {
-        result = result && (getVersion()
-            == other.getVersion());
+        if (getVersion()
+            != other.getVersion()) return false;
       }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
+      if (!getUnknownFields().equals(other.getUnknownFields())) return false;
+      return true;
     }
 
-    private int memoizedHashCode = 0;
     @java.lang.Override
     public int hashCode() {
       if (memoizedHashCode != 0) {
         return memoizedHashCode;
       }
       int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
+      hash = (19 * hash) + getDescriptor().hashCode();
       if (hasUserPayload()) {
         hash = (37 * hash) + USER_PAYLOAD_FIELD_NUMBER;
         hash = (53 * hash) + getUserPayload().hashCode();
@@ -435,6 +402,17 @@ public final class LlapDaemonProtocolProtos {
       return hash;
     }
 
+    public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom(
+        java.nio.ByteBuffer data)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      return PARSER.parseFrom(data);
+    }
+    public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom(
+        java.nio.ByteBuffer data,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      return PARSER.parseFrom(data, extensionRegistry);
+    }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom(
         com.google.protobuf.ByteString data)
         throws com.google.protobuf.InvalidProtocolBufferException {
@@ -458,46 +436,61 @@ public final class LlapDaemonProtocolProtos {
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return PARSER.parseFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input);
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input, extensionRegistry);
     }
+
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseDelimitedFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseDelimitedWithIOException(PARSER, input);
     }
+
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseDelimitedFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom(
         com.google.protobuf.CodedInputStream input)
         throws java.io.IOException {
-      return PARSER.parseFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input);
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom(
         com.google.protobuf.CodedInputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input, extensionRegistry);
     }
 
-    public static Builder newBuilder() { return Builder.create(); }
+    @java.lang.Override
     public Builder newBuilderForType() { return newBuilder(); }
+    public static Builder newBuilder() {
+      return DEFAULT_INSTANCE.toBuilder();
+    }
     public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto prototype) {
-      return newBuilder().mergeFrom(prototype);
+      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+    }
+    @java.lang.Override
+    public Builder toBuilder() {
+      return this == DEFAULT_INSTANCE
+          ? new Builder() : new Builder().mergeFrom(this);
     }
-    public Builder toBuilder() { return newBuilder(this); }
 
     @java.lang.Override
     protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+        com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
       Builder builder = new Builder(parent);
       return builder;
     }
@@ -505,14 +498,16 @@ public final class LlapDaemonProtocolProtos {
      * Protobuf type {@code UserPayloadProto}
      */
     public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder {
+        com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+        // @@protoc_insertion_point(builder_implements:UserPayloadProto)
+        org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder {
       public static final com.google.protobuf.Descriptors.Descriptor
           getDescriptor() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_descriptor;
       }
 
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+      @java.lang.Override
+      protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
           internalGetFieldAccessorTable() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_fieldAccessorTable
             .ensureFieldAccessorsInitialized(
@@ -521,44 +516,35 @@ public final class LlapDaemonProtocolProtos {
 
       // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.newBuilder()
       private Builder() {
-        maybeForceBuilderInitialization();
+
       }
 
       private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+          com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
         super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-        }
-      }
-      private static Builder create() {
-        return new Builder();
-      }
 
+      }
+      @java.lang.Override
       public Builder clear() {
         super.clear();
+        bitField0_ = 0;
         userPayload_ = com.google.protobuf.ByteString.EMPTY;
-        bitField0_ = (bitField0_ & ~0x00000001);
         version_ = 0;
-        bitField0_ = (bitField0_ & ~0x00000002);
         return this;
       }
 
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
+      @java.lang.Override
       public com.google.protobuf.Descriptors.Descriptor
           getDescriptorForType() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_descriptor;
       }
 
+      @java.lang.Override
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getDefaultInstanceForType() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance();
       }
 
+      @java.lang.Override
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto build() {
         org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto result = buildPartial();
         if (!result.isInitialized()) {
@@ -567,23 +553,61 @@ public final class LlapDaemonProtocolProtos {
         return result;
       }
 
+      @java.lang.Override
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto buildPartial() {
         org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto(this);
+        if (bitField0_ != 0) { buildPartial0(result); }
+        onBuilt();
+        return result;
+      }
+
+      private void buildPartial0(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto result) {
         int from_bitField0_ = bitField0_;
         int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
+        if (((from_bitField0_ & 0x00000001) != 0)) {
+          result.userPayload_ = userPayload_;
           to_bitField0_ |= 0x00000001;
         }
-        result.userPayload_ = userPayload_;
-        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
+        if (((from_bitField0_ & 0x00000002) != 0)) {
+          result.version_ = version_;
           to_bitField0_ |= 0x00000002;
         }
-        result.version_ = version_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
+        result.bitField0_ |= to_bitField0_;
       }
 
+      @java.lang.Override
+      public Builder clone() {
+        return super.clone();
+      }
+      @java.lang.Override
+      public Builder setField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          java.lang.Object value) {
+        return super.setField(field, value);
+      }
+      @java.lang.Override
+      public Builder clearField(
+          com.google.protobuf.Descriptors.FieldDescriptor field) {
+        return super.clearField(field);
+      }
+      @java.lang.Override
+      public Builder clearOneof(
+          com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+        return super.clearOneof(oneof);
+      }
+      @java.lang.Override
+      public Builder setRepeatedField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          int index, java.lang.Object value) {
+        return super.setRepeatedField(field, index, value);
+      }
+      @java.lang.Override
+      public Builder addRepeatedField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          java.lang.Object value) {
+        return super.addRepeatedField(field, value);
+      }
+      @java.lang.Override
       public Builder mergeFrom(com.google.protobuf.Message other) {
         if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto) {
           return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto)other);
@@ -602,60 +626,90 @@ public final class LlapDaemonProtocolProtos {
           setVersion(other.getVersion());
         }
         this.mergeUnknownFields(other.getUnknownFields());
+        onChanged();
         return this;
       }
 
+      @java.lang.Override
       public final boolean isInitialized() {
         return true;
       }
 
+      @java.lang.Override
       public Builder mergeFrom(
           com.google.protobuf.CodedInputStream input,
           com.google.protobuf.ExtensionRegistryLite extensionRegistry)
           throws java.io.IOException {
-        org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parsedMessage = null;
+        if (extensionRegistry == null) {
+          throw new java.lang.NullPointerException();
+        }
         try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+          boolean done = false;
+          while (!done) {
+            int tag = input.readTag();
+            switch (tag) {
+              case 0:
+                done = true;
+                break;
+              case 10: {
+                userPayload_ = input.readBytes();
+                bitField0_ |= 0x00000001;
+                break;
+              } // case 10
+              case 16: {
+                version_ = input.readInt32();
+                bitField0_ |= 0x00000002;
+                break;
+              } // case 16
+              default: {
+                if (!super.parseUnknownField(input, extensionRegistry, tag)) {
+                  done = true; // was an endgroup tag
+                }
+                break;
+              } // default:
+            } // switch (tag)
+          } // while (!done)
         } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto) e.getUnfinishedMessage();
-          throw e;
+          throw e.unwrapIOException();
         } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
+          onChanged();
+        } // finally
         return this;
       }
       private int bitField0_;
 
-      // optional bytes user_payload = 1;
       private com.google.protobuf.ByteString userPayload_ = com.google.protobuf.ByteString.EMPTY;
       /**
        * <code>optional bytes user_payload = 1;</code>
+       * @return Whether the userPayload field is set.
        */
+      @java.lang.Override
       public boolean hasUserPayload() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
+        return ((bitField0_ & 0x00000001) != 0);
       }
       /**
        * <code>optional bytes user_payload = 1;</code>
+       * @return The userPayload.
        */
+      @java.lang.Override
       public com.google.protobuf.ByteString getUserPayload() {
         return userPayload_;
       }
       /**
        * <code>optional bytes user_payload = 1;</code>
+       * @param value The userPayload to set.
+       * @return This builder for chaining.
        */
       public Builder setUserPayload(com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000001;
+        if (value == null) { throw new NullPointerException(); }
         userPayload_ = value;
+        bitField0_ |= 0x00000001;
         onChanged();
         return this;
       }
       /**
        * <code>optional bytes user_payload = 1;</code>
+       * @return This builder for chaining.
        */
       public Builder clearUserPayload() {
         bitField0_ = (bitField0_ & ~0x00000001);
@@ -664,31 +718,38 @@ public final class LlapDaemonProtocolProtos {
         return this;
       }
 
-      // optional int32 version = 2;
       private int version_ ;
       /**
        * <code>optional int32 version = 2;</code>
+       * @return Whether the version field is set.
        */
+      @java.lang.Override
       public boolean hasVersion() {
-        return ((bitField0_ & 0x00000002) == 0x00000002);
+        return ((bitField0_ & 0x00000002) != 0);
       }
       /**
        * <code>optional int32 version = 2;</code>
+       * @return The version.
        */
+      @java.lang.Override
       public int getVersion() {
         return version_;
       }
       /**
        * <code>optional int32 version = 2;</code>
+       * @param value The version to set.
+       * @return This builder for chaining.
        */
       public Builder setVersion(int value) {
-        bitField0_ |= 0x00000002;
+
         version_ = value;
+        bitField0_ |= 0x00000002;
         onChanged();
         return this;
       }
       /**
        * <code>optional int32 version = 2;</code>
+       * @return This builder for chaining.
        */
       public Builder clearVersion() {
         bitField0_ = (bitField0_ & ~0x00000002);
@@ -696,43 +757,99 @@ public final class LlapDaemonProtocolProtos {
         onChanged();
         return this;
       }
+      @java.lang.Override
+      public final Builder setUnknownFields(
+          final com.google.protobuf.UnknownFieldSet unknownFields) {
+        return super.setUnknownFields(unknownFields);
+      }
+
+      @java.lang.Override
+      public final Builder mergeUnknownFields(
+          final com.google.protobuf.UnknownFieldSet unknownFields) {
+        return super.mergeUnknownFields(unknownFields);
+      }
+
 
       // @@protoc_insertion_point(builder_scope:UserPayloadProto)
     }
 
+    // @@protoc_insertion_point(class_scope:UserPayloadProto)
+    private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto DEFAULT_INSTANCE;
     static {
-      defaultInstance = new UserPayloadProto(true);
-      defaultInstance.initFields();
+      DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto();
+    }
+
+    public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getDefaultInstance() {
+      return DEFAULT_INSTANCE;
+    }
+
+    @java.lang.Deprecated public static final com.google.protobuf.Parser<UserPayloadProto>
+        PARSER = new com.google.protobuf.AbstractParser<UserPayloadProto>() {
+      @java.lang.Override
+      public UserPayloadProto parsePartialFrom(
+          com.google.protobuf.CodedInputStream input,
+          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+          throws com.google.protobuf.InvalidProtocolBufferException {
+        Builder builder = newBuilder();
+        try {
+          builder.mergeFrom(input, extensionRegistry);
+        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+          throw e.setUnfinishedMessage(builder.buildPartial());
+        } catch (com.google.protobuf.UninitializedMessageException e) {
+          throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
+        } catch (java.io.IOException e) {
+          throw new com.google.protobuf.InvalidProtocolBufferException(e)
+              .setUnfinishedMessage(builder.buildPartial());
+        }
+        return builder.buildPartial();
+      }
+    };
+
+    public static com.google.protobuf.Parser<UserPayloadProto> parser() {
+      return PARSER;
+    }
+
+    @java.lang.Override
+    public com.google.protobuf.Parser<UserPayloadProto> getParserForType() {
+      return PARSER;
+    }
+
+    @java.lang.Override
+    public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getDefaultInstanceForType() {
+      return DEFAULT_INSTANCE;
     }
 
-    // @@protoc_insertion_point(class_scope:UserPayloadProto)
   }
 
-  public interface EntityDescriptorProtoOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
+  public interface EntityDescriptorProtoOrBuilder extends
+      // @@protoc_insertion_point(interface_extends:EntityDescriptorProto)
+      com.google.protobuf.MessageOrBuilder {
 
-    // optional string class_name = 1;
     /**
      * <code>optional string class_name = 1;</code>
+     * @return Whether the className field is set.
      */
     boolean hasClassName();
     /**
      * <code>optional string class_name = 1;</code>
+     * @return The className.
      */
     java.lang.String getClassName();
     /**
      * <code>optional string class_name = 1;</code>
+     * @return The bytes for className.
      */
     com.google.protobuf.ByteString
         getClassNameBytes();
 
-    // optional .UserPayloadProto user_payload = 2;
     /**
      * <code>optional .UserPayloadProto user_payload = 2;</code>
+     * @return Whether the userPayload field is set.
      */
     boolean hasUserPayload();
     /**
      * <code>optional .UserPayloadProto user_payload = 2;</code>
+     * @return The userPayload.
      */
     org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getUserPayload();
     /**
@@ -740,13 +857,14 @@ public final class LlapDaemonProtocolProtos {
      */
     org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder getUserPayloadOrBuilder();
 
-    // optional bytes history_text = 3;
     /**
      * <code>optional bytes history_text = 3;</code>
+     * @return Whether the historyText field is set.
      */
     boolean hasHistoryText();
     /**
      * <code>optional bytes history_text = 3;</code>
+     * @return The historyText.
      */
     com.google.protobuf.ByteString getHistoryText();
   }
@@ -754,128 +872,56 @@ public final class LlapDaemonProtocolProtos {
    * Protobuf type {@code EntityDescriptorProto}
    */
   public static final class EntityDescriptorProto extends
-      com.google.protobuf.GeneratedMessage
-      implements EntityDescriptorProtoOrBuilder {
+      com.google.protobuf.GeneratedMessageV3 implements
+      // @@protoc_insertion_point(message_implements:EntityDescriptorProto)
+      EntityDescriptorProtoOrBuilder {
+  private static final long serialVersionUID = 0L;
     // Use EntityDescriptorProto.newBuilder() to construct.
-    private EntityDescriptorProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+    private EntityDescriptorProto(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
       super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private EntityDescriptorProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final EntityDescriptorProto defaultInstance;
-    public static EntityDescriptorProto getDefaultInstance() {
-      return defaultInstance;
     }
-
-    public EntityDescriptorProto getDefaultInstanceForType() {
-      return defaultInstance;
+    private EntityDescriptorProto() {
+      className_ = "";
+      historyText_ = com.google.protobuf.ByteString.EMPTY;
     }
 
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
     @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private EntityDescriptorProto(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              bitField0_ |= 0x00000001;
-              className_ = input.readBytes();
-              break;
-            }
-            case 18: {
-              org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder subBuilder = null;
-              if (((bitField0_ & 0x00000002) == 0x00000002)) {
-                subBuilder = userPayload_.toBuilder();
-              }
-              userPayload_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.PARSER, extensionRegistry);
-              if (subBuilder != null) {
-                subBuilder.mergeFrom(userPayload_);
-                userPayload_ = subBuilder.buildPartial();
-              }
-              bitField0_ |= 0x00000002;
-              break;
-            }
-            case 26: {
-              bitField0_ |= 0x00000004;
-              historyText_ = input.readBytes();
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
+    @SuppressWarnings({"unused"})
+    protected java.lang.Object newInstance(
+        UnusedPrivateParameter unused) {
+      return new EntityDescriptorProto();
     }
+
     public static final com.google.protobuf.Descriptors.Descriptor
         getDescriptor() {
       return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_descriptor;
     }
 
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+    @java.lang.Override
+    protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
         internalGetFieldAccessorTable() {
       return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_fieldAccessorTable
           .ensureFieldAccessorsInitialized(
               org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder.class);
     }
 
-    public static com.google.protobuf.Parser<EntityDescriptorProto> PARSER =
-        new com.google.protobuf.AbstractParser<EntityDescriptorProto>() {
-      public EntityDescriptorProto parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new EntityDescriptorProto(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<EntityDescriptorProto> getParserForType() {
-      return PARSER;
-    }
-
     private int bitField0_;
-    // optional string class_name = 1;
     public static final int CLASS_NAME_FIELD_NUMBER = 1;
-    private java.lang.Object className_;
+    @SuppressWarnings("serial")
+    private volatile java.lang.Object className_ = "";
     /**
      * <code>optional string class_name = 1;</code>
+     * @return Whether the className field is set.
      */
+    @java.lang.Override
     public boolean hasClassName() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
+      return ((bitField0_ & 0x00000001) != 0);
     }
     /**
      * <code>optional string class_name = 1;</code>
+     * @return The className.
      */
+    @java.lang.Override
     public java.lang.String getClassName() {
       java.lang.Object ref = className_;
       if (ref instanceof java.lang.String) {
@@ -892,7 +938,9 @@ public final class LlapDaemonProtocolProtos {
     }
     /**
      * <code>optional string class_name = 1;</code>
+     * @return The bytes for className.
      */
+    @java.lang.Override
     public com.google.protobuf.ByteString
         getClassNameBytes() {
       java.lang.Object ref = className_;
@@ -907,103 +955,99 @@ public final class LlapDaemonProtocolProtos {
       }
     }
 
-    // optional .UserPayloadProto user_payload = 2;
     public static final int USER_PAYLOAD_FIELD_NUMBER = 2;
     private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto userPayload_;
     /**
      * <code>optional .UserPayloadProto user_payload = 2;</code>
+     * @return Whether the userPayload field is set.
      */
+    @java.lang.Override
     public boolean hasUserPayload() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
+      return ((bitField0_ & 0x00000002) != 0);
     }
     /**
      * <code>optional .UserPayloadProto user_payload = 2;</code>
+     * @return The userPayload.
      */
+    @java.lang.Override
     public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getUserPayload() {
-      return userPayload_;
+      return userPayload_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance() : userPayload_;
     }
     /**
      * <code>optional .UserPayloadProto user_payload = 2;</code>
      */
+    @java.lang.Override
     public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder getUserPayloadOrBuilder() {
-      return userPayload_;
+      return userPayload_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance() : userPayload_;
     }
 
-    // optional bytes history_text = 3;
     public static final int HISTORY_TEXT_FIELD_NUMBER = 3;
-    private com.google.protobuf.ByteString historyText_;
+    private com.google.protobuf.ByteString historyText_ = com.google.protobuf.ByteString.EMPTY;
     /**
      * <code>optional bytes history_text = 3;</code>
+     * @return Whether the historyText field is set.
      */
+    @java.lang.Override
     public boolean hasHistoryText() {
-      return ((bitField0_ & 0x00000004) == 0x00000004);
+      return ((bitField0_ & 0x00000004) != 0);
     }
     /**
      * <code>optional bytes history_text = 3;</code>
+     * @return The historyText.
      */
+    @java.lang.Override
     public com.google.protobuf.ByteString getHistoryText() {
       return historyText_;
     }
 
-    private void initFields() {
-      className_ = "";
-      userPayload_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance();
-      historyText_ = com.google.protobuf.ByteString.EMPTY;
-    }
     private byte memoizedIsInitialized = -1;
+    @java.lang.Override
     public final boolean isInitialized() {
       byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
+      if (isInitialized == 1) return true;
+      if (isInitialized == 0) return false;
 
       memoizedIsInitialized = 1;
       return true;
     }
 
+    @java.lang.Override
     public void writeTo(com.google.protobuf.CodedOutputStream output)
                         throws java.io.IOException {
-      getSerializedSize();
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeBytes(1, getClassNameBytes());
+      if (((bitField0_ & 0x00000001) != 0)) {
+        com.google.protobuf.GeneratedMessageV3.writeString(output, 1, className_);
       }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        output.writeMessage(2, userPayload_);
+      if (((bitField0_ & 0x00000002) != 0)) {
+        output.writeMessage(2, getUserPayload());
       }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
+      if (((bitField0_ & 0x00000004) != 0)) {
         output.writeBytes(3, historyText_);
       }
       getUnknownFields().writeTo(output);
     }
 
-    private int memoizedSerializedSize = -1;
+    @java.lang.Override
     public int getSerializedSize() {
-      int size = memoizedSerializedSize;
+      int size = memoizedSize;
       if (size != -1) return size;
 
       size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(1, getClassNameBytes());
+      if (((bitField0_ & 0x00000001) != 0)) {
+        size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, className_);
       }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
+      if (((bitField0_ & 0x00000002) != 0)) {
         size += com.google.protobuf.CodedOutputStream
-          .computeMessageSize(2, userPayload_);
+          .computeMessageSize(2, getUserPayload());
       }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
+      if (((bitField0_ & 0x00000004) != 0)) {
         size += com.google.protobuf.CodedOutputStream
           .computeBytesSize(3, historyText_);
       }
       size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
+      memoizedSize = size;
       return size;
     }
 
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
     @java.lang.Override
     public boolean equals(final java.lang.Object obj) {
       if (obj == this) {
@@ -1014,35 +1058,32 @@ public final class LlapDaemonProtocolProtos {
       }
       org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto) obj;
 
-      boolean result = true;
-      result = result && (hasClassName() == other.hasClassName());
+      if (hasClassName() != other.hasClassName()) return false;
       if (hasClassName()) {
-        result = result && getClassName()
-            .equals(other.getClassName());
+        if (!getClassName()
+            .equals(other.getClassName())) return false;
       }
-      result = result && (hasUserPayload() == other.hasUserPayload());
+      if (hasUserPayload() != other.hasUserPayload()) return false;
       if (hasUserPayload()) {
-        result = result && getUserPayload()
-            .equals(other.getUserPayload());
+        if (!getUserPayload()
+            .equals(other.getUserPayload())) return false;
       }
-      result = result && (hasHistoryText() == other.hasHistoryText());
+      if (hasHistoryText() != other.hasHistoryText()) return false;
       if (hasHistoryText()) {
-        result = result && getHistoryText()
-            .equals(other.getHistoryText());
+        if (!getHistoryText()
+            .equals(other.getHistoryText())) return false;
       }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
+      if (!getUnknownFields().equals(other.getUnknownFields())) return false;
+      return true;
     }
 
-    private int memoizedHashCode = 0;
     @java.lang.Override
     public int hashCode() {
       if (memoizedHashCode != 0) {
         return memoizedHashCode;
       }
       int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
+      hash = (19 * hash) + getDescriptor().hashCode();
       if (hasClassName()) {
         hash = (37 * hash) + CLASS_NAME_FIELD_NUMBER;
         hash = (53 * hash) + getClassName().hashCode();
@@ -1060,6 +1101,17 @@ public final class LlapDaemonProtocolProtos {
       return hash;
     }
 
+    public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom(
+        java.nio.ByteBuffer data)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      return PARSER.parseFrom(data);
+    }
+    public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom(
+        java.nio.ByteBuffer data,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      return PARSER.parseFrom(data, extensionRegistry);
+    }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom(
         com.google.protobuf.ByteString data)
         throws com.google.protobuf.InvalidProtocolBufferException {
@@ -1083,46 +1135,61 @@ public final class LlapDaemonProtocolProtos {
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return PARSER.parseFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input);
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input, extensionRegistry);
     }
+
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseDelimitedFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseDelimitedWithIOException(PARSER, input);
     }
+
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseDelimitedFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom(
         com.google.protobuf.CodedInputStream input)
         throws java.io.IOException {
-      return PARSER.parseFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input);
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom(
         com.google.protobuf.CodedInputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input, extensionRegistry);
     }
 
-    public static Builder newBuilder() { return Builder.create(); }
+    @java.lang.Override
     public Builder newBuilderForType() { return newBuilder(); }
+    public static Builder newBuilder() {
+      return DEFAULT_INSTANCE.toBuilder();
+    }
     public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto prototype) {
-      return newBuilder().mergeFrom(prototype);
+      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+    }
+    @java.lang.Override
+    public Builder toBuilder() {
+      return this == DEFAULT_INSTANCE
+          ? new Builder() : new Builder().mergeFrom(this);
     }
-    public Builder toBuilder() { return newBuilder(this); }
 
     @java.lang.Override
     protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+        com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
       Builder builder = new Builder(parent);
       return builder;
     }
@@ -1130,14 +1197,16 @@ public final class LlapDaemonProtocolProtos {
      * Protobuf type {@code EntityDescriptorProto}
      */
     public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder {
+        com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+        // @@protoc_insertion_point(builder_implements:EntityDescriptorProto)
+        org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder {
       public static final com.google.protobuf.Descriptors.Descriptor
           getDescriptor() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_descriptor;
       }
 
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+      @java.lang.Override
+      protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
           internalGetFieldAccessorTable() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_fieldAccessorTable
             .ensureFieldAccessorsInitialized(
@@ -1150,47 +1219,42 @@ public final class LlapDaemonProtocolProtos {
       }
 
       private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+          com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
         super(parent);
         maybeForceBuilderInitialization();
       }
       private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+        if (com.google.protobuf.GeneratedMessageV3
+                .alwaysUseFieldBuilders) {
           getUserPayloadFieldBuilder();
         }
       }
-      private static Builder create() {
-        return new Builder();
-      }
-
+      @java.lang.Override
       public Builder clear() {
         super.clear();
+        bitField0_ = 0;
         className_ = "";
-        bitField0_ = (bitField0_ & ~0x00000001);
-        if (userPayloadBuilder_ == null) {
-          userPayload_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance();
-        } else {
-          userPayloadBuilder_.clear();
+        userPayload_ = null;
+        if (userPayloadBuilder_ != null) {
+          userPayloadBuilder_.dispose();
+          userPayloadBuilder_ = null;
         }
-        bitField0_ = (bitField0_ & ~0x00000002);
         historyText_ = com.google.protobuf.ByteString.EMPTY;
-        bitField0_ = (bitField0_ & ~0x00000004);
         return this;
       }
 
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
+      @java.lang.Override
       public com.google.protobuf.Descriptors.Descriptor
           getDescriptorForType() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_descriptor;
       }
 
+      @java.lang.Override
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getDefaultInstanceForType() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
       }
 
+      @java.lang.Override
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto build() {
         org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto result = buildPartial();
         if (!result.isInitialized()) {
@@ -1199,31 +1263,67 @@ public final class LlapDaemonProtocolProtos {
         return result;
       }
 
+      @java.lang.Override
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto buildPartial() {
         org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto(this);
+        if (bitField0_ != 0) { buildPartial0(result); }
+        onBuilt();
+        return result;
+      }
+
+      private void buildPartial0(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto result) {
         int from_bitField0_ = bitField0_;
         int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
+        if (((from_bitField0_ & 0x00000001) != 0)) {
+          result.className_ = className_;
           to_bitField0_ |= 0x00000001;
         }
-        result.className_ = className_;
-        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
+        if (((from_bitField0_ & 0x00000002) != 0)) {
+          result.userPayload_ = userPayloadBuilder_ == null
+              ? userPayload_
+              : userPayloadBuilder_.build();
           to_bitField0_ |= 0x00000002;
         }
-        if (userPayloadBuilder_ == null) {
-          result.userPayload_ = userPayload_;
-        } else {
-          result.userPayload_ = userPayloadBuilder_.build();
-        }
-        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
+        if (((from_bitField0_ & 0x00000004) != 0)) {
+          result.historyText_ = historyText_;
           to_bitField0_ |= 0x00000004;
         }
-        result.historyText_ = historyText_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
+        result.bitField0_ |= to_bitField0_;
       }
 
+      @java.lang.Override
+      public Builder clone() {
+        return super.clone();
+      }
+      @java.lang.Override
+      public Builder setField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          java.lang.Object value) {
+        return super.setField(field, value);
+      }
+      @java.lang.Override
+      public Builder clearField(
+          com.google.protobuf.Descriptors.FieldDescriptor field) {
+        return super.clearField(field);
+      }
+      @java.lang.Override
+      public Builder clearOneof(
+          com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+        return super.clearOneof(oneof);
+      }
+      @java.lang.Override
+      public Builder setRepeatedField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          int index, java.lang.Object value) {
+        return super.setRepeatedField(field, index, value);
+      }
+      @java.lang.Override
+      public Builder addRepeatedField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          java.lang.Object value) {
+        return super.addRepeatedField(field, value);
+      }
+      @java.lang.Override
       public Builder mergeFrom(com.google.protobuf.Message other) {
         if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto) {
           return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto)other);
@@ -1236,8 +1336,8 @@ public final class LlapDaemonProtocolProtos {
       public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto other) {
         if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) return this;
         if (other.hasClassName()) {
-          bitField0_ |= 0x00000001;
           className_ = other.className_;
+          bitField0_ |= 0x00000001;
           onChanged();
         }
         if (other.hasUserPayload()) {
@@ -1247,49 +1347,86 @@ public final class LlapDaemonProtocolProtos {
           setHistoryText(other.getHistoryText());
         }
         this.mergeUnknownFields(other.getUnknownFields());
+        onChanged();
         return this;
       }
 
+      @java.lang.Override
       public final boolean isInitialized() {
         return true;
       }
 
+      @java.lang.Override
       public Builder mergeFrom(
           com.google.protobuf.CodedInputStream input,
           com.google.protobuf.ExtensionRegistryLite extensionRegistry)
           throws java.io.IOException {
-        org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parsedMessage = null;
+        if (extensionRegistry == null) {
+          throw new java.lang.NullPointerException();
+        }
         try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+          boolean done = false;
+          while (!done) {
+            int tag = input.readTag();
+            switch (tag) {
+              case 0:
+                done = true;
+                break;
+              case 10: {
+                className_ = input.readBytes();
+                bitField0_ |= 0x00000001;
+                break;
+              } // case 10
+              case 18: {
+                input.readMessage(
+                    getUserPayloadFieldBuilder().getBuilder(),
+                    extensionRegistry);
+                bitField0_ |= 0x00000002;
+                break;
+              } // case 18
+              case 26: {
+                historyText_ = input.readBytes();
+                bitField0_ |= 0x00000004;
+                break;
+              } // case 26
+              default: {
+                if (!super.parseUnknownField(input, extensionRegistry, tag)) {
+                  done = true; // was an endgroup tag
+                }
+                break;
+              } // default:
+            } // switch (tag)
+          } // while (!done)
         } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto) e.getUnfinishedMessage();
-          throw e;
+          throw e.unwrapIOException();
         } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
+          onChanged();
+        } // finally
         return this;
       }
       private int bitField0_;
 
-      // optional string class_name = 1;
       private java.lang.Object className_ = "";
       /**
        * <code>optional string class_name = 1;</code>
+       * @return Whether the className field is set.
        */
       public boolean hasClassName() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
+        return ((bitField0_ & 0x00000001) != 0);
       }
       /**
        * <code>optional string class_name = 1;</code>
+       * @return The className.
        */
       public java.lang.String getClassName() {
         java.lang.Object ref = className_;
         if (!(ref instanceof java.lang.String)) {
-          java.lang.String s = ((com.google.protobuf.ByteString) ref)
-              .toStringUtf8();
-          className_ = s;
+          com.google.protobuf.ByteString bs =
+              (com.google.protobuf.ByteString) ref;
+          java.lang.String s = bs.toStringUtf8();
+          if (bs.isValidUtf8()) {
+            className_ = s;
+          }
           return s;
         } else {
           return (java.lang.String) ref;
@@ -1297,6 +1434,7 @@ public final class LlapDaemonProtocolProtos {
       }
       /**
        * <code>optional string class_name = 1;</code>
+       * @return The bytes for className.
        */
       public com.google.protobuf.ByteString
           getClassNameBytes() {
@@ -1313,56 +1451,58 @@ public final class LlapDaemonProtocolProtos {
       }
       /**
        * <code>optional string class_name = 1;</code>
+       * @param value The className to set.
+       * @return This builder for chaining.
        */
       public Builder setClassName(
           java.lang.String value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000001;
+        if (value == null) { throw new NullPointerException(); }
         className_ = value;
+        bitField0_ |= 0x00000001;
         onChanged();
         return this;
       }
       /**
        * <code>optional string class_name = 1;</code>
+       * @return This builder for chaining.
        */
       public Builder clearClassName() {
-        bitField0_ = (bitField0_ & ~0x00000001);
         className_ = getDefaultInstance().getClassName();
+        bitField0_ = (bitField0_ & ~0x00000001);
         onChanged();
         return this;
       }
       /**
        * <code>optional string class_name = 1;</code>
+       * @param value The bytes for className to set.
+       * @return This builder for chaining.
        */
       public Builder setClassNameBytes(
           com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000001;
+        if (value == null) { throw new NullPointerException(); }
         className_ = value;
+        bitField0_ |= 0x00000001;
         onChanged();
         return this;
       }
 
-      // optional .UserPayloadProto user_payload = 2;
-      private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto userPayload_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance();
-      private com.google.protobuf.SingleFieldBuilder<
+      private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto userPayload_;
+      private com.google.protobuf.SingleFieldBuilderV3<
           org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder> userPayloadBuilder_;
       /**
        * <code>optional .UserPayloadProto user_payload = 2;</code>
+       * @return Whether the userPayload field is set.
        */
       public boolean hasUserPayload() {
-        return ((bitField0_ & 0x00000002) == 0x00000002);
+        return ((bitField0_ & 0x00000002) != 0);
       }
       /**
        * <code>optional .UserPayloadProto user_payload = 2;</code>
+       * @return The userPayload.
        */
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getUserPayload() {
         if (userPayloadBuilder_ == null) {
-          return userPayload_;
+          return userPayload_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance() : userPayload_;
         } else {
           return userPayloadBuilder_.getMessage();
         }
@@ -1376,11 +1516,11 @@ public final class LlapDaemonProtocolProtos {
             throw new NullPointerException();
           }
           userPayload_ = value;
-          onChanged();
         } else {
           userPayloadBuilder_.setMessage(value);
         }
         bitField0_ |= 0x00000002;
+        onChanged();
         return this;
       }
       /**
@@ -1390,11 +1530,11 @@ public final class LlapDaemonProtocolProtos {
           org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder builderForValue) {
         if (userPayloadBuilder_ == null) {
           userPayload_ = builderForValue.build();
-          onChanged();
         } else {
           userPayloadBuilder_.setMessage(builderForValue.build());
         }
         bitField0_ |= 0x00000002;
+        onChanged();
         return this;
       }
       /**
@@ -1402,31 +1542,33 @@ public final class LlapDaemonProtocolProtos {
        */
       public Builder mergeUserPayload(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto value) {
         if (userPayloadBuilder_ == null) {
-          if (((bitField0_ & 0x00000002) == 0x00000002) &&
-              userPayload_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance()) {
-            userPayload_ =
-              org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.newBuilder(userPayload_).mergeFrom(value).buildPartial();
+          if (((bitField0_ & 0x00000002) != 0) &&
+            userPayload_ != null &&
+            userPayload_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance()) {
+            getUserPayloadBuilder().mergeFrom(value);
           } else {
             userPayload_ = value;
           }
-          onChanged();
         } else {
           userPayloadBuilder_.mergeFrom(value);
         }
-        bitField0_ |= 0x00000002;
-        return this;
+        if (userPayload_ != null) {
+          bitField0_ |= 0x00000002;
+          onChanged();
+        }
+        return this;
       }
       /**
        * <code>optional .UserPayloadProto user_payload = 2;</code>
        */
       public Builder clearUserPayload() {
-        if (userPayloadBuilder_ == null) {
-          userPayload_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance();
-          onChanged();
-        } else {
-          userPayloadBuilder_.clear();
-        }
         bitField0_ = (bitField0_ & ~0x00000002);
+        userPayload_ = null;
+        if (userPayloadBuilder_ != null) {
+          userPayloadBuilder_.dispose();
+          userPayloadBuilder_ = null;
+        }
+        onChanged();
         return this;
       }
       /**
@@ -1444,19 +1586,20 @@ public final class LlapDaemonProtocolProtos {
         if (userPayloadBuilder_ != null) {
           return userPayloadBuilder_.getMessageOrBuilder();
         } else {
-          return userPayload_;
+          return userPayload_ == null ?
+              org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance() : userPayload_;
         }
       }
       /**
        * <code>optional .UserPayloadProto user_payload = 2;</code>
        */
-      private com.google.protobuf.SingleFieldBuilder<
+      private com.google.protobuf.SingleFieldBuilderV3<
           org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder> 
           getUserPayloadFieldBuilder() {
         if (userPayloadBuilder_ == null) {
-          userPayloadBuilder_ = new com.google.protobuf.SingleFieldBuilder<
+          userPayloadBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
               org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder>(
-                  userPayload_,
+                  getUserPayload(),
                   getParentForChildren(),
                   isClean());
           userPayload_ = null;
@@ -1464,34 +1607,38 @@ public final class LlapDaemonProtocolProtos {
         return userPayloadBuilder_;
       }
 
-      // optional bytes history_text = 3;
       private com.google.protobuf.ByteString historyText_ = com.google.protobuf.ByteString.EMPTY;
       /**
        * <code>optional bytes history_text = 3;</code>
+       * @return Whether the historyText field is set.
        */
+      @java.lang.Override
       public boolean hasHistoryText() {
-        return ((bitField0_ & 0x00000004) == 0x00000004);
+        return ((bitField0_ & 0x00000004) != 0);
       }
       /**
        * <code>optional bytes history_text = 3;</code>
+       * @return The historyText.
        */
+      @java.lang.Override
       public com.google.protobuf.ByteString getHistoryText() {
         return historyText_;
       }
       /**
        * <code>optional bytes history_text = 3;</code>
+       * @param value The historyText to set.
+       * @return This builder for chaining.
        */
       public Builder setHistoryText(com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000004;
+        if (value == null) { throw new NullPointerException(); }
         historyText_ = value;
+        bitField0_ |= 0x00000004;
         onChanged();
         return this;
       }
       /**
        * <code>optional bytes history_text = 3;</code>
+       * @return This builder for chaining.
        */
       public Builder clearHistoryText() {
         bitField0_ = (bitField0_ & ~0x00000004);
@@ -1499,43 +1646,99 @@ public final class LlapDaemonProtocolProtos {
         onChanged();
         return this;
       }
+      @java.lang.Override
+      public final Builder setUnknownFields(
+          final com.google.protobuf.UnknownFieldSet unknownFields) {
+        return super.setUnknownFields(unknownFields);
+      }
+
+      @java.lang.Override
+      public final Builder mergeUnknownFields(
+          final com.google.protobuf.UnknownFieldSet unknownFields) {
+        return super.mergeUnknownFields(unknownFields);
+      }
+
 
       // @@protoc_insertion_point(builder_scope:EntityDescriptorProto)
     }
 
+    // @@protoc_insertion_point(class_scope:EntityDescriptorProto)
+    private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto DEFAULT_INSTANCE;
     static {
-      defaultInstance = new EntityDescriptorProto(true);
-      defaultInstance.initFields();
+      DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto();
+    }
+
+    public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getDefaultInstance() {
+      return DEFAULT_INSTANCE;
+    }
+
+    @java.lang.Deprecated public static final com.google.protobuf.Parser<EntityDescriptorProto>
+        PARSER = new com.google.protobuf.AbstractParser<EntityDescriptorProto>() {
+      @java.lang.Override
+      public EntityDescriptorProto parsePartialFrom(
+          com.google.protobuf.CodedInputStream input,
+          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+          throws com.google.protobuf.InvalidProtocolBufferException {
+        Builder builder = newBuilder();
+        try {
+          builder.mergeFrom(input, extensionRegistry);
+        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+          throw e.setUnfinishedMessage(builder.buildPartial());
+        } catch (com.google.protobuf.UninitializedMessageException e) {
+          throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
+        } catch (java.io.IOException e) {
+          throw new com.google.protobuf.InvalidProtocolBufferException(e)
+              .setUnfinishedMessage(builder.buildPartial());
+        }
+        return builder.buildPartial();
+      }
+    };
+
+    public static com.google.protobuf.Parser<EntityDescriptorProto> parser() {
+      return PARSER;
+    }
+
+    @java.lang.Override
+    public com.google.protobuf.Parser<EntityDescriptorProto> getParserForType() {
+      return PARSER;
+    }
+
+    @java.lang.Override
+    public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getDefaultInstanceForType() {
+      return DEFAULT_INSTANCE;
     }
 
-    // @@protoc_insertion_point(class_scope:EntityDescriptorProto)
   }
 
-  public interface IOSpecProtoOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
+  public interface IOSpecProtoOrBuilder extends
+      // @@protoc_insertion_point(interface_extends:IOSpecProto)
+      com.google.protobuf.MessageOrBuilder {
 
-    // optional string connected_vertex_name = 1;
     /**
      * <code>optional string connected_vertex_name = 1;</code>
+     * @return Whether the connectedVertexName field is set.
      */
     boolean hasConnectedVertexName();
     /**
      * <code>optional string connected_vertex_name = 1;</code>
+     * @return The connectedVertexName.
      */
     java.lang.String getConnectedVertexName();
     /**
      * <code>optional string connected_vertex_name = 1;</code>
+     * @return The bytes for connectedVertexName.
      */
     com.google.protobuf.ByteString
         getConnectedVertexNameBytes();
 
-    // optional .EntityDescriptorProto io_descriptor = 2;
     /**
      * <code>optional .EntityDescriptorProto io_descriptor = 2;</code>
+     * @return Whether the ioDescriptor field is set.
      */
     boolean hasIoDescriptor();
     /**
      * <code>optional .EntityDescriptorProto io_descriptor = 2;</code>
+     * @return The ioDescriptor.
      */
     org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getIoDescriptor();
     /**
@@ -1543,13 +1746,14 @@ public final class LlapDaemonProtocolProtos {
      */
     org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getIoDescriptorOrBuilder();
 
-    // optional int32 physical_edge_count = 3;
     /**
      * <code>optional int32 physical_edge_count = 3;</code>
+     * @return Whether the physicalEdgeCount field is set.
      */
     boolean hasPhysicalEdgeCount();
     /**
      * <code>optional int32 physical_edge_count = 3;</code>
+     * @return The physicalEdgeCount.
      */
     int getPhysicalEdgeCount();
   }
@@ -1557,128 +1761,55 @@ public final class LlapDaemonProtocolProtos {
    * Protobuf type {@code IOSpecProto}
    */
   public static final class IOSpecProto extends
-      com.google.protobuf.GeneratedMessage
-      implements IOSpecProtoOrBuilder {
+      com.google.protobuf.GeneratedMessageV3 implements
+      // @@protoc_insertion_point(message_implements:IOSpecProto)
+      IOSpecProtoOrBuilder {
+  private static final long serialVersionUID = 0L;
     // Use IOSpecProto.newBuilder() to construct.
-    private IOSpecProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+    private IOSpecProto(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
       super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private IOSpecProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final IOSpecProto defaultInstance;
-    public static IOSpecProto getDefaultInstance() {
-      return defaultInstance;
     }
-
-    public IOSpecProto getDefaultInstanceForType() {
-      return defaultInstance;
+    private IOSpecProto() {
+      connectedVertexName_ = "";
     }
 
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
     @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private IOSpecProto(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              bitField0_ |= 0x00000001;
-              connectedVertexName_ = input.readBytes();
-              break;
-            }
-            case 18: {
-              org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder subBuilder = null;
-              if (((bitField0_ & 0x00000002) == 0x00000002)) {
-                subBuilder = ioDescriptor_.toBuilder();
-              }
-              ioDescriptor_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.PARSER, extensionRegistry);
-              if (subBuilder != null) {
-                subBuilder.mergeFrom(ioDescriptor_);
-                ioDescriptor_ = subBuilder.buildPartial();
-              }
-              bitField0_ |= 0x00000002;
-              break;
-            }
-            case 24: {
-              bitField0_ |= 0x00000004;
-              physicalEdgeCount_ = input.readInt32();
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
+    @SuppressWarnings({"unused"})
+    protected java.lang.Object newInstance(
+        UnusedPrivateParameter unused) {
+      return new IOSpecProto();
     }
+
     public static final com.google.protobuf.Descriptors.Descriptor
         getDescriptor() {
       return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_descriptor;
     }
 
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+    @java.lang.Override
+    protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
         internalGetFieldAccessorTable() {
       return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_fieldAccessorTable
           .ensureFieldAccessorsInitialized(
               org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder.class);
     }
 
-    public static com.google.protobuf.Parser<IOSpecProto> PARSER =
-        new com.google.protobuf.AbstractParser<IOSpecProto>() {
-      public IOSpecProto parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new IOSpecProto(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<IOSpecProto> getParserForType() {
-      return PARSER;
-    }
-
     private int bitField0_;
-    // optional string connected_vertex_name = 1;
     public static final int CONNECTED_VERTEX_NAME_FIELD_NUMBER = 1;
-    private java.lang.Object connectedVertexName_;
+    @SuppressWarnings("serial")
+    private volatile java.lang.Object connectedVertexName_ = "";
     /**
      * <code>optional string connected_vertex_name = 1;</code>
+     * @return Whether the connectedVertexName field is set.
      */
+    @java.lang.Override
     public boolean hasConnectedVertexName() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
+      return ((bitField0_ & 0x00000001) != 0);
     }
     /**
      * <code>optional string connected_vertex_name = 1;</code>
+     * @return The connectedVertexName.
      */
+    @java.lang.Override
     public java.lang.String getConnectedVertexName() {
       java.lang.Object ref = connectedVertexName_;
       if (ref instanceof java.lang.String) {
@@ -1695,7 +1826,9 @@ public final class LlapDaemonProtocolProtos {
     }
     /**
      * <code>optional string connected_vertex_name = 1;</code>
+     * @return The bytes for connectedVertexName.
      */
+    @java.lang.Override
     public com.google.protobuf.ByteString
         getConnectedVertexNameBytes() {
       java.lang.Object ref = connectedVertexName_;
@@ -1710,103 +1843,99 @@ public final class LlapDaemonProtocolProtos {
       }
     }
 
-    // optional .EntityDescriptorProto io_descriptor = 2;
     public static final int IO_DESCRIPTOR_FIELD_NUMBER = 2;
     private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto ioDescriptor_;
     /**
      * <code>optional .EntityDescriptorProto io_descriptor = 2;</code>
+     * @return Whether the ioDescriptor field is set.
      */
+    @java.lang.Override
     public boolean hasIoDescriptor() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
+      return ((bitField0_ & 0x00000002) != 0);
     }
     /**
      * <code>optional .EntityDescriptorProto io_descriptor = 2;</code>
+     * @return The ioDescriptor.
      */
+    @java.lang.Override
     public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getIoDescriptor() {
-      return ioDescriptor_;
+      return ioDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : ioDescriptor_;
     }
     /**
      * <code>optional .EntityDescriptorProto io_descriptor = 2;</code>
      */
+    @java.lang.Override
     public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getIoDescriptorOrBuilder() {
-      return ioDescriptor_;
+      return ioDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : ioDescriptor_;
     }
 
-    // optional int32 physical_edge_count = 3;
     public static final int PHYSICAL_EDGE_COUNT_FIELD_NUMBER = 3;
-    private int physicalEdgeCount_;
+    private int physicalEdgeCount_ = 0;
     /**
      * <code>optional int32 physical_edge_count = 3;</code>
+     * @return Whether the physicalEdgeCount field is set.
      */
+    @java.lang.Override
     public boolean hasPhysicalEdgeCount() {
-      return ((bitField0_ & 0x00000004) == 0x00000004);
+      return ((bitField0_ & 0x00000004) != 0);
     }
     /**
      * <code>optional int32 physical_edge_count = 3;</code>
+     * @return The physicalEdgeCount.
      */
+    @java.lang.Override
     public int getPhysicalEdgeCount() {
       return physicalEdgeCount_;
     }
 
-    private void initFields() {
-      connectedVertexName_ = "";
-      ioDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
-      physicalEdgeCount_ = 0;
-    }
     private byte memoizedIsInitialized = -1;
+    @java.lang.Override
     public final boolean isInitialized() {
       byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
+      if (isInitialized == 1) return true;
+      if (isInitialized == 0) return false;
 
       memoizedIsInitialized = 1;
       return true;
     }
 
+    @java.lang.Override
     public void writeTo(com.google.protobuf.CodedOutputStream output)
                         throws java.io.IOException {
-      getSerializedSize();
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeBytes(1, getConnectedVertexNameBytes());
+      if (((bitField0_ & 0x00000001) != 0)) {
+        com.google.protobuf.GeneratedMessageV3.writeString(output, 1, connectedVertexName_);
       }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        output.writeMessage(2, ioDescriptor_);
+      if (((bitField0_ & 0x00000002) != 0)) {
+        output.writeMessage(2, getIoDescriptor());
       }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
+      if (((bitField0_ & 0x00000004) != 0)) {
         output.writeInt32(3, physicalEdgeCount_);
       }
       getUnknownFields().writeTo(output);
     }
 
-    private int memoizedSerializedSize = -1;
+    @java.lang.Override
     public int getSerializedSize() {
-      int size = memoizedSerializedSize;
+      int size = memoizedSize;
       if (size != -1) return size;
 
       size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(1, getConnectedVertexNameBytes());
+      if (((bitField0_ & 0x00000001) != 0)) {
+        size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, connectedVertexName_);
       }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
+      if (((bitField0_ & 0x00000002) != 0)) {
         size += com.google.protobuf.CodedOutputStream
-          .computeMessageSize(2, ioDescriptor_);
+          .computeMessageSize(2, getIoDescriptor());
       }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
+      if (((bitField0_ & 0x00000004) != 0)) {
         size += com.google.protobuf.CodedOutputStream
           .computeInt32Size(3, physicalEdgeCount_);
       }
       size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
+      memoizedSize = size;
       return size;
     }
 
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
     @java.lang.Override
     public boolean equals(final java.lang.Object obj) {
       if (obj == this) {
@@ -1817,35 +1946,32 @@ public final class LlapDaemonProtocolProtos {
       }
       org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto) obj;
 
-      boolean result = true;
-      result = result && (hasConnectedVertexName() == other.hasConnectedVertexName());
+      if (hasConnectedVertexName() != other.hasConnectedVertexName()) return false;
       if (hasConnectedVertexName()) {
-        result = result && getConnectedVertexName()
-            .equals(other.getConnectedVertexName());
+        if (!getConnectedVertexName()
+            .equals(other.getConnectedVertexName())) return false;
       }
-      result = result && (hasIoDescriptor() == other.hasIoDescriptor());
+      if (hasIoDescriptor() != other.hasIoDescriptor()) return false;
       if (hasIoDescriptor()) {
-        result = result && getIoDescriptor()
-            .equals(other.getIoDescriptor());
+        if (!getIoDescriptor()
+            .equals(other.getIoDescriptor())) return false;
       }
-      result = result && (hasPhysicalEdgeCount() == other.hasPhysicalEdgeCount());
+      if (hasPhysicalEdgeCount() != other.hasPhysicalEdgeCount()) return false;
       if (hasPhysicalEdgeCount()) {
-        result = result && (getPhysicalEdgeCount()
-            == other.getPhysicalEdgeCount());
+        if (getPhysicalEdgeCount()
+            != other.getPhysicalEdgeCount()) return false;
       }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
+      if (!getUnknownFields().equals(other.getUnknownFields())) return false;
+      return true;
     }
 
-    private int memoizedHashCode = 0;
     @java.lang.Override
     public int hashCode() {
       if (memoizedHashCode != 0) {
         return memoizedHashCode;
       }
       int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
+      hash = (19 * hash) + getDescriptor().hashCode();
       if (hasConnectedVertexName()) {
         hash = (37 * hash) + CONNECTED_VERTEX_NAME_FIELD_NUMBER;
         hash = (53 * hash) + getConnectedVertexName().hashCode();
@@ -1863,6 +1989,17 @@ public final class LlapDaemonProtocolProtos {
       return hash;
     }
 
+    public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom(
+        java.nio.ByteBuffer data)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      return PARSER.parseFrom(data);
+    }
+    public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom(
+        java.nio.ByteBuffer data,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      return PARSER.parseFrom(data, extensionRegistry);
+    }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom(
         com.google.protobuf.ByteString data)
         throws com.google.protobuf.InvalidProtocolBufferException {
@@ -1886,46 +2023,61 @@ public final class LlapDaemonProtocolProtos {
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return PARSER.parseFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input);
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input, extensionRegistry);
     }
+
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseDelimitedFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseDelimitedWithIOException(PARSER, input);
     }
+
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseDelimitedFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom(
         com.google.protobuf.CodedInputStream input)
         throws java.io.IOException {
-      return PARSER.parseFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input);
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom(
         com.google.protobuf.CodedInputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input, extensionRegistry);
     }
 
-    public static Builder newBuilder() { return Builder.create(); }
+    @java.lang.Override
     public Builder newBuilderForType() { return newBuilder(); }
+    public static Builder newBuilder() {
+      return DEFAULT_INSTANCE.toBuilder();
+    }
     public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto prototype) {
-      return newBuilder().mergeFrom(prototype);
+      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+    }
+    @java.lang.Override
+    public Builder toBuilder() {
+      return this == DEFAULT_INSTANCE
+          ? new Builder() : new Builder().mergeFrom(this);
     }
-    public Builder toBuilder() { return newBuilder(this); }
 
     @java.lang.Override
     protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+        com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
       Builder builder = new Builder(parent);
       return builder;
     }
@@ -1933,14 +2085,16 @@ public final class LlapDaemonProtocolProtos {
      * Protobuf type {@code IOSpecProto}
      */
     public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder {
+        com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+        // @@protoc_insertion_point(builder_implements:IOSpecProto)
+        org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder {
       public static final com.google.protobuf.Descriptors.Descriptor
           getDescriptor() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_descriptor;
       }
 
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+      @java.lang.Override
+      protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
           internalGetFieldAccessorTable() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_fieldAccessorTable
             .ensureFieldAccessorsInitialized(
@@ -1953,47 +2107,42 @@ public final class LlapDaemonProtocolProtos {
       }
 
       private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+          com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
         super(parent);
         maybeForceBuilderInitialization();
       }
       private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+        if (com.google.protobuf.GeneratedMessageV3
+                .alwaysUseFieldBuilders) {
           getIoDescriptorFieldBuilder();
         }
       }
-      private static Builder create() {
-        return new Builder();
-      }
-
+      @java.lang.Override
       public Builder clear() {
         super.clear();
+        bitField0_ = 0;
         connectedVertexName_ = "";
-        bitField0_ = (bitField0_ & ~0x00000001);
-        if (ioDescriptorBuilder_ == null) {
-          ioDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
-        } else {
-          ioDescriptorBuilder_.clear();
+        ioDescriptor_ = null;
+        if (ioDescriptorBuilder_ != null) {
+          ioDescriptorBuilder_.dispose();
+          ioDescriptorBuilder_ = null;
         }
-        bitField0_ = (bitField0_ & ~0x00000002);
         physicalEdgeCount_ = 0;
-        bitField0_ = (bitField0_ & ~0x00000004);
         return this;
       }
 
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
+      @java.lang.Override
       public com.google.protobuf.Descriptors.Descriptor
           getDescriptorForType() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_descriptor;
       }
 
+      @java.lang.Override
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getDefaultInstanceForType() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.getDefaultInstance();
       }
 
+      @java.lang.Override
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto build() {
         org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto result = buildPartial();
         if (!result.isInitialized()) {
@@ -2002,31 +2151,67 @@ public final class LlapDaemonProtocolProtos {
         return result;
       }
 
+      @java.lang.Override
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto buildPartial() {
         org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto(this);
+        if (bitField0_ != 0) { buildPartial0(result); }
+        onBuilt();
+        return result;
+      }
+
+      private void buildPartial0(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto result) {
         int from_bitField0_ = bitField0_;
         int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
+        if (((from_bitField0_ & 0x00000001) != 0)) {
+          result.connectedVertexName_ = connectedVertexName_;
           to_bitField0_ |= 0x00000001;
         }
-        result.connectedVertexName_ = connectedVertexName_;
-        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
+        if (((from_bitField0_ & 0x00000002) != 0)) {
+          result.ioDescriptor_ = ioDescriptorBuilder_ == null
+              ? ioDescriptor_
+              : ioDescriptorBuilder_.build();
           to_bitField0_ |= 0x00000002;
         }
-        if (ioDescriptorBuilder_ == null) {
-          result.ioDescriptor_ = ioDescriptor_;
-        } else {
-          result.ioDescriptor_ = ioDescriptorBuilder_.build();
-        }
-        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
+        if (((from_bitField0_ & 0x00000004) != 0)) {
+          result.physicalEdgeCount_ = physicalEdgeCount_;
           to_bitField0_ |= 0x00000004;
         }
-        result.physicalEdgeCount_ = physicalEdgeCount_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
+        result.bitField0_ |= to_bitField0_;
       }
 
+      @java.lang.Override
+      public Builder clone() {
+        return super.clone();
+      }
+      @java.lang.Override
+      public Builder setField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          java.lang.Object value) {
+        return super.setField(field, value);
+      }
+      @java.lang.Override
+      public Builder clearField(
+          com.google.protobuf.Descriptors.FieldDescriptor field) {
+        return super.clearField(field);
+      }
+      @java.lang.Override
+      public Builder clearOneof(
+          com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+        return super.clearOneof(oneof);
+      }
+      @java.lang.Override
+      public Builder setRepeatedField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          int index, java.lang.Object value) {
+        return super.setRepeatedField(field, index, value);
+      }
+      @java.lang.Override
+      public Builder addRepeatedField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          java.lang.Object value) {
+        return super.addRepeatedField(field, value);
+      }
+      @java.lang.Override
       public Builder mergeFrom(com.google.protobuf.Message other) {
         if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto) {
           return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto)other);
@@ -2039,8 +2224,8 @@ public final class LlapDaemonProtocolProtos {
       public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto other) {
         if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.getDefaultInstance()) return this;
         if (other.hasConnectedVertexName()) {
-          bitField0_ |= 0x00000001;
           connectedVertexName_ = other.connectedVertexName_;
+          bitField0_ |= 0x00000001;
           onChanged();
         }
         if (other.hasIoDescriptor()) {
@@ -2050,49 +2235,86 @@ public final class LlapDaemonProtocolProtos {
           setPhysicalEdgeCount(other.getPhysicalEdgeCount());
         }
         this.mergeUnknownFields(other.getUnknownFields());
+        onChanged();
         return this;
       }
 
+      @java.lang.Override
       public final boolean isInitialized() {
         return true;
       }
 
+      @java.lang.Override
       public Builder mergeFrom(
           com.google.protobuf.CodedInputStream input,
           com.google.protobuf.ExtensionRegistryLite extensionRegistry)
           throws java.io.IOException {
-        org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parsedMessage = null;
+        if (extensionRegistry == null) {
+          throw new java.lang.NullPointerException();
+        }
         try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+          boolean done = false;
+          while (!done) {
+            int tag = input.readTag();
+            switch (tag) {
+              case 0:
+                done = true;
+                break;
+              case 10: {
+                connectedVertexName_ = input.readBytes();
+                bitField0_ |= 0x00000001;
+                break;
+              } // case 10
+              case 18: {
+                input.readMessage(
+                    getIoDescriptorFieldBuilder().getBuilder(),
+                    extensionRegistry);
+                bitField0_ |= 0x00000002;
+                break;
+              } // case 18
+              case 24: {
+                physicalEdgeCount_ = input.readInt32();
+                bitField0_ |= 0x00000004;
+                break;
+              } // case 24
+              default: {
+                if (!super.parseUnknownField(input, extensionRegistry, tag)) {
+                  done = true; // was an endgroup tag
+                }
+                break;
+              } // default:
+            } // switch (tag)
+          } // while (!done)
         } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto) e.getUnfinishedMessage();
-          throw e;
+          throw e.unwrapIOException();
         } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
+          onChanged();
+        } // finally
         return this;
       }
       private int bitField0_;
 
-      // optional string connected_vertex_name = 1;
       private java.lang.Object connectedVertexName_ = "";
       /**
        * <code>optional string connected_vertex_name = 1;</code>
+       * @return Whether the connectedVertexName field is set.
        */
       public boolean hasConnectedVertexName() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
+        return ((bitField0_ & 0x00000001) != 0);
       }
       /**
        * <code>optional string connected_vertex_name = 1;</code>
+       * @return The connectedVertexName.
        */
       public java.lang.String getConnectedVertexName() {
         java.lang.Object ref = connectedVertexName_;
         if (!(ref instanceof java.lang.String)) {
-          java.lang.String s = ((com.google.protobuf.ByteString) ref)
-              .toStringUtf8();
-          connectedVertexName_ = s;
+          com.google.protobuf.ByteString bs =
+              (com.google.protobuf.ByteString) ref;
+          java.lang.String s = bs.toStringUtf8();
+          if (bs.isValidUtf8()) {
+            connectedVertexName_ = s;
+          }
           return s;
         } else {
           return (java.lang.String) ref;
@@ -2100,6 +2322,7 @@ public final class LlapDaemonProtocolProtos {
       }
       /**
        * <code>optional string connected_vertex_name = 1;</code>
+       * @return The bytes for connectedVertexName.
        */
       public com.google.protobuf.ByteString
           getConnectedVertexNameBytes() {
@@ -2116,56 +2339,58 @@ public final class LlapDaemonProtocolProtos {
       }
       /**
        * <code>optional string connected_vertex_name = 1;</code>
+       * @param value The connectedVertexName to set.
+       * @return This builder for chaining.
        */
       public Builder setConnectedVertexName(
           java.lang.String value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000001;
+        if (value == null) { throw new NullPointerException(); }
         connectedVertexName_ = value;
+        bitField0_ |= 0x00000001;
         onChanged();
         return this;
       }
       /**
        * <code>optional string connected_vertex_name = 1;</code>
+       * @return This builder for chaining.
        */
       public Builder clearConnectedVertexName() {
-        bitField0_ = (bitField0_ & ~0x00000001);
         connectedVertexName_ = getDefaultInstance().getConnectedVertexName();
+        bitField0_ = (bitField0_ & ~0x00000001);
         onChanged();
         return this;
       }
       /**
        * <code>optional string connected_vertex_name = 1;</code>
+       * @param value The bytes for connectedVertexName to set.
+       * @return This builder for chaining.
        */
       public Builder setConnectedVertexNameBytes(
           com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000001;
+        if (value == null) { throw new NullPointerException(); }
         connectedVertexName_ = value;
+        bitField0_ |= 0x00000001;
         onChanged();
         return this;
       }
 
-      // optional .EntityDescriptorProto io_descriptor = 2;
-      private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto ioDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
-      private com.google.protobuf.SingleFieldBuilder<
+      private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto ioDescriptor_;
+      private com.google.protobuf.SingleFieldBuilderV3<
           org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> ioDescriptorBuilder_;
       /**
        * <code>optional .EntityDescriptorProto io_descriptor = 2;</code>
+       * @return Whether the ioDescriptor field is set.
        */
       public boolean hasIoDescriptor() {
-        return ((bitField0_ & 0x00000002) == 0x00000002);
+        return ((bitField0_ & 0x00000002) != 0);
       }
       /**
        * <code>optional .EntityDescriptorProto io_descriptor = 2;</code>
+       * @return The ioDescriptor.
        */
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getIoDescriptor() {
         if (ioDescriptorBuilder_ == null) {
-          return ioDescriptor_;
+          return ioDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : ioDescriptor_;
         } else {
           return ioDescriptorBuilder_.getMessage();
         }
@@ -2179,11 +2404,11 @@ public final class LlapDaemonProtocolProtos {
             throw new NullPointerException();
           }
           ioDescriptor_ = value;
-          onChanged();
         } else {
           ioDescriptorBuilder_.setMessage(value);
         }
         bitField0_ |= 0x00000002;
+        onChanged();
         return this;
       }
       /**
@@ -2193,11 +2418,11 @@ public final class LlapDaemonProtocolProtos {
           org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder builderForValue) {
         if (ioDescriptorBuilder_ == null) {
           ioDescriptor_ = builderForValue.build();
-          onChanged();
         } else {
           ioDescriptorBuilder_.setMessage(builderForValue.build());
         }
         bitField0_ |= 0x00000002;
+        onChanged();
         return this;
       }
       /**
@@ -2205,31 +2430,33 @@ public final class LlapDaemonProtocolProtos {
        */
       public Builder mergeIoDescriptor(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto value) {
         if (ioDescriptorBuilder_ == null) {
-          if (((bitField0_ & 0x00000002) == 0x00000002) &&
-              ioDescriptor_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) {
-            ioDescriptor_ =
-              org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.newBuilder(ioDescriptor_).mergeFrom(value).buildPartial();
+          if (((bitField0_ & 0x00000002) != 0) &&
+            ioDescriptor_ != null &&
+            ioDescriptor_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) {
+            getIoDescriptorBuilder().mergeFrom(value);
           } else {
             ioDescriptor_ = value;
           }
-          onChanged();
         } else {
           ioDescriptorBuilder_.mergeFrom(value);
         }
-        bitField0_ |= 0x00000002;
+        if (ioDescriptor_ != null) {
+          bitField0_ |= 0x00000002;
+          onChanged();
+        }
         return this;
       }
       /**
        * <code>optional .EntityDescriptorProto io_descriptor = 2;</code>
        */
       public Builder clearIoDescriptor() {
-        if (ioDescriptorBuilder_ == null) {
-          ioDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
-          onChanged();
-        } else {
-          ioDescriptorBuilder_.clear();
-        }
         bitField0_ = (bitField0_ & ~0x00000002);
+        ioDescriptor_ = null;
+        if (ioDescriptorBuilder_ != null) {
+          ioDescriptorBuilder_.dispose();
+          ioDescriptorBuilder_ = null;
+        }
+        onChanged();
         return this;
       }
       /**
@@ -2247,19 +2474,20 @@ public final class LlapDaemonProtocolProtos {
         if (ioDescriptorBuilder_ != null) {
           return ioDescriptorBuilder_.getMessageOrBuilder();
         } else {
-          return ioDescriptor_;
+          return ioDescriptor_ == null ?
+              org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : ioDescriptor_;
         }
       }
       /**
        * <code>optional .EntityDescriptorProto io_descriptor = 2;</code>
        */
-      private com.google.protobuf.SingleFieldBuilder<
+      private com.google.protobuf.SingleFieldBuilderV3<
           org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> 
           getIoDescriptorFieldBuilder() {
         if (ioDescriptorBuilder_ == null) {
-          ioDescriptorBuilder_ = new com.google.protobuf.SingleFieldBuilder<
+          ioDescriptorBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
               org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder>(
-                  ioDescriptor_,
+                  getIoDescriptor(),
                   getParentForChildren(),
                   isClean());
           ioDescriptor_ = null;
@@ -2267,31 +2495,38 @@ public final class LlapDaemonProtocolProtos {
         return ioDescriptorBuilder_;
       }
 
-      // optional int32 physical_edge_count = 3;
       private int physicalEdgeCount_ ;
       /**
        * <code>optional int32 physical_edge_count = 3;</code>
+       * @return Whether the physicalEdgeCount field is set.
        */
+      @java.lang.Override
       public boolean hasPhysicalEdgeCount() {
-        return ((bitField0_ & 0x00000004) == 0x00000004);
+        return ((bitField0_ & 0x00000004) != 0);
       }
       /**
        * <code>optional int32 physical_edge_count = 3;</code>
+       * @return The physicalEdgeCount.
        */
+      @java.lang.Override
       public int getPhysicalEdgeCount() {
         return physicalEdgeCount_;
       }
       /**
        * <code>optional int32 physical_edge_count = 3;</code>
+       * @param value The physicalEdgeCount to set.
+       * @return This builder for chaining.
        */
       public Builder setPhysicalEdgeCount(int value) {
-        bitField0_ |= 0x00000004;
+
         physicalEdgeCount_ = value;
+        bitField0_ |= 0x00000004;
         onChanged();
         return this;
       }
       /**
        * <code>optional int32 physical_edge_count = 3;</code>
+       * @return This builder for chaining.
        */
       public Builder clearPhysicalEdgeCount() {
         bitField0_ = (bitField0_ & ~0x00000004);
@@ -2299,63 +2534,124 @@ public final class LlapDaemonProtocolProtos {
         onChanged();
         return this;
       }
+      @java.lang.Override
+      public final Builder setUnknownFields(
+          final com.google.protobuf.UnknownFieldSet unknownFields) {
+        return super.setUnknownFields(unknownFields);
+      }
+
+      @java.lang.Override
+      public final Builder mergeUnknownFields(
+          final com.google.protobuf.UnknownFieldSet unknownFields) {
+        return super.mergeUnknownFields(unknownFields);
+      }
+
 
       // @@protoc_insertion_point(builder_scope:IOSpecProto)
     }
 
+    // @@protoc_insertion_point(class_scope:IOSpecProto)
+    private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto DEFAULT_INSTANCE;
     static {
-      defaultInstance = new IOSpecProto(true);
-      defaultInstance.initFields();
+      DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto();
+    }
+
+    public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getDefaultInstance() {
+      return DEFAULT_INSTANCE;
+    }
+
+    @java.lang.Deprecated public static final com.google.protobuf.Parser<IOSpecProto>
+        PARSER = new com.google.protobuf.AbstractParser<IOSpecProto>() {
+      @java.lang.Override
+      public IOSpecProto parsePartialFrom(
+          com.google.protobuf.CodedInputStream input,
+          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+          throws com.google.protobuf.InvalidProtocolBufferException {
+        Builder builder = newBuilder();
+        try {
+          builder.mergeFrom(input, extensionRegistry);
+        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+          throw e.setUnfinishedMessage(builder.buildPartial());
+        } catch (com.google.protobuf.UninitializedMessageException e) {
+          throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
+        } catch (java.io.IOException e) {
+          throw new com.google.protobuf.InvalidProtocolBufferException(e)
+              .setUnfinishedMessage(builder.buildPartial());
+        }
+        return builder.buildPartial();
+      }
+    };
+
+    public static com.google.protobuf.Parser<IOSpecProto> parser() {
+      return PARSER;
+    }
+
+    @java.lang.Override
+    public com.google.protobuf.Parser<IOSpecProto> getParserForType() {
+      return PARSER;
+    }
+
+    @java.lang.Override
+    public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getDefaultInstanceForType() {
+      return DEFAULT_INSTANCE;
     }
 
-    // @@protoc_insertion_point(class_scope:IOSpecProto)
   }
 
-  public interface GroupInputSpecProtoOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
+  public interface GroupInputSpecProtoOrBuilder extends
+      // @@protoc_insertion_point(interface_extends:GroupInputSpecProto)
+      com.google.protobuf.MessageOrBuilder {
 
-    // optional string group_name = 1;
     /**
      * <code>optional string group_name = 1;</code>
+     * @return Whether the groupName field is set.
      */
     boolean hasGroupName();
     /**
      * <code>optional string group_name = 1;</code>
+     * @return The groupName.
      */
     java.lang.String getGroupName();
     /**
      * <code>optional string group_name = 1;</code>
+     * @return The bytes for groupName.
      */
     com.google.protobuf.ByteString
         getGroupNameBytes();
 
-    // repeated string group_vertices = 2;
     /**
      * <code>repeated string group_vertices = 2;</code>
+     * @return A list containing the groupVertices.
      */
     java.util.List<java.lang.String>
-    getGroupVerticesList();
+        getGroupVerticesList();
     /**
      * <code>repeated string group_vertices = 2;</code>
+     * @return The count of groupVertices.
      */
     int getGroupVerticesCount();
     /**
      * <code>repeated string group_vertices = 2;</code>
+     * @param index The index of the element to return.
+     * @return The groupVertices at the given index.
      */
     java.lang.String getGroupVertices(int index);
     /**
      * <code>repeated string group_vertices = 2;</code>
+     * @param index The index of the value to return.
+     * @return The bytes of the groupVertices at the given index.
      */
     com.google.protobuf.ByteString
         getGroupVerticesBytes(int index);
 
-    // optional .EntityDescriptorProto merged_input_descriptor = 3;
     /**
      * <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code>
+     * @return Whether the mergedInputDescriptor field is set.
      */
     boolean hasMergedInputDescriptor();
     /**
      * <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code>
+     * @return The mergedInputDescriptor.
      */
     org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getMergedInputDescriptor();
     /**
@@ -2367,134 +2663,57 @@ public final class LlapDaemonProtocolProtos {
    * Protobuf type {@code GroupInputSpecProto}
    */
   public static final class GroupInputSpecProto extends
-      com.google.protobuf.GeneratedMessage
-      implements GroupInputSpecProtoOrBuilder {
+      com.google.protobuf.GeneratedMessageV3 implements
+      // @@protoc_insertion_point(message_implements:GroupInputSpecProto)
+      GroupInputSpecProtoOrBuilder {
+  private static final long serialVersionUID = 0L;
     // Use GroupInputSpecProto.newBuilder() to construct.
-    private GroupInputSpecProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+    private GroupInputSpecProto(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
       super(builder);
-      this.unknownFields = builder.getUnknownFields();
     }
-    private GroupInputSpecProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final GroupInputSpecProto defaultInstance;
-    public static GroupInputSpecProto getDefaultInstance() {
-      return defaultInstance;
-    }
-
-    public GroupInputSpecProto getDefaultInstanceForType() {
-      return defaultInstance;
+    private GroupInputSpecProto() {
+      groupName_ = "";
+      groupVertices_ =
+          com.google.protobuf.LazyStringArrayList.emptyList();
     }
 
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
     @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private GroupInputSpecProto(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              bitField0_ |= 0x00000001;
-              groupName_ = input.readBytes();
-              break;
-            }
-            case 18: {
-              if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
-                groupVertices_ = new com.google.protobuf.LazyStringArrayList();
-                mutable_bitField0_ |= 0x00000002;
-              }
-              groupVertices_.add(input.readBytes());
-              break;
-            }
-            case 26: {
-              org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder subBuilder = null;
-              if (((bitField0_ & 0x00000002) == 0x00000002)) {
-                subBuilder = mergedInputDescriptor_.toBuilder();
-              }
-              mergedInputDescriptor_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.PARSER, extensionRegistry);
-              if (subBuilder != null) {
-                subBuilder.mergeFrom(mergedInputDescriptor_);
-                mergedInputDescriptor_ = subBuilder.buildPartial();
-              }
-              bitField0_ |= 0x00000002;
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
-          groupVertices_ = new com.google.protobuf.UnmodifiableLazyStringList(groupVertices_);
-        }
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
+    @SuppressWarnings({"unused"})
+    protected java.lang.Object newInstance(
+        UnusedPrivateParameter unused) {
+      return new GroupInputSpecProto();
     }
+
     public static final com.google.protobuf.Descriptors.Descriptor
         getDescriptor() {
       return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GroupInputSpecProto_descriptor;
     }
 
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+    @java.lang.Override
+    protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
         internalGetFieldAccessorTable() {
       return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GroupInputSpecProto_fieldAccessorTable
           .ensureFieldAccessorsInitialized(
               org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder.class);
     }
 
-    public static com.google.protobuf.Parser<GroupInputSpecProto> PARSER =
-        new com.google.protobuf.AbstractParser<GroupInputSpecProto>() {
-      public GroupInputSpecProto parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new GroupInputSpecProto(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<GroupInputSpecProto> getParserForType() {
-      return PARSER;
-    }
-
     private int bitField0_;
-    // optional string group_name = 1;
     public static final int GROUP_NAME_FIELD_NUMBER = 1;
-    private java.lang.Object groupName_;
+    @SuppressWarnings("serial")
+    private volatile java.lang.Object groupName_ = "";
     /**
      * <code>optional string group_name = 1;</code>
+     * @return Whether the groupName field is set.
      */
+    @java.lang.Override
     public boolean hasGroupName() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
+      return ((bitField0_ & 0x00000001) != 0);
     }
     /**
      * <code>optional string group_name = 1;</code>
+     * @return The groupName.
      */
+    @java.lang.Override
     public java.lang.String getGroupName() {
       java.lang.Object ref = groupName_;
       if (ref instanceof java.lang.String) {
@@ -2511,7 +2730,9 @@ public final class LlapDaemonProtocolProtos {
     }
     /**
      * <code>optional string group_name = 1;</code>
+     * @return The bytes for groupName.
      */
+    @java.lang.Override
     public com.google.protobuf.ByteString
         getGroupNameBytes() {
       java.lang.Object ref = groupName_;
@@ -2526,122 +2747,121 @@ public final class LlapDaemonProtocolProtos {
       }
     }
 
-    // repeated string group_vertices = 2;
     public static final int GROUP_VERTICES_FIELD_NUMBER = 2;
-    private com.google.protobuf.LazyStringList groupVertices_;
+    @SuppressWarnings("serial")
+    private com.google.protobuf.LazyStringArrayList groupVertices_ =
+        com.google.protobuf.LazyStringArrayList.emptyList();
     /**
      * <code>repeated string group_vertices = 2;</code>
+     * @return A list containing the groupVertices.
      */
-    public java.util.List<java.lang.String>
+    public com.google.protobuf.ProtocolStringList
         getGroupVerticesList() {
       return groupVertices_;
     }
     /**
      * <code>repeated string group_vertices = 2;</code>
+     * @return The count of groupVertices.
      */
     public int getGroupVerticesCount() {
       return groupVertices_.size();
     }
     /**
      * <code>repeated string group_vertices = 2;</code>
+     * @param index The index of the element to return.
+     * @return The groupVertices at the given index.
      */
     public java.lang.String getGroupVertices(int index) {
       return groupVertices_.get(index);
     }
     /**
      * <code>repeated string group_vertices = 2;</code>
+     * @param index The index of the value to return.
+     * @return The bytes of the groupVertices at the given index.
      */
     public com.google.protobuf.ByteString
         getGroupVerticesBytes(int index) {
       return groupVertices_.getByteString(index);
     }
 
-    // optional .EntityDescriptorProto merged_input_descriptor = 3;
     public static final int MERGED_INPUT_DESCRIPTOR_FIELD_NUMBER = 3;
     private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto mergedInputDescriptor_;
     /**
      * <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code>
+     * @return Whether the mergedInputDescriptor field is set.
      */
+    @java.lang.Override
     public boolean hasMergedInputDescriptor() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
+      return ((bitField0_ & 0x00000002) != 0);
     }
     /**
      * <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code>
+     * @return The mergedInputDescriptor.
      */
+    @java.lang.Override
     public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getMergedInputDescriptor() {
-      return mergedInputDescriptor_;
+      return mergedInputDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : mergedInputDescriptor_;
     }
     /**
      * <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code>
      */
+    @java.lang.Override
     public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getMergedInputDescriptorOrBuilder() {
-      return mergedInputDescriptor_;
+      return mergedInputDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : mergedInputDescriptor_;
     }
 
-    private void initFields() {
-      groupName_ = "";
-      groupVertices_ = com.google.protobuf.LazyStringArrayList.EMPTY;
-      mergedInputDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
-    }
     private byte memoizedIsInitialized = -1;
+    @java.lang.Override
     public final boolean isInitialized() {
       byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
+      if (isInitialized == 1) return true;
+      if (isInitialized == 0) return false;
 
       memoizedIsInitialized = 1;
       return true;
     }
 
+    @java.lang.Override
     public void writeTo(com.google.protobuf.CodedOutputStream output)
                         throws java.io.IOException {
-      getSerializedSize();
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeBytes(1, getGroupNameBytes());
+      if (((bitField0_ & 0x00000001) != 0)) {
+        com.google.protobuf.GeneratedMessageV3.writeString(output, 1, groupName_);
       }
       for (int i = 0; i < groupVertices_.size(); i++) {
-        output.writeBytes(2, groupVertices_.getByteString(i));
+        com.google.protobuf.GeneratedMessageV3.writeString(output, 2, groupVertices_.getRaw(i));
       }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        output.writeMessage(3, mergedInputDescriptor_);
+      if (((bitField0_ & 0x00000002) != 0)) {
+        output.writeMessage(3, getMergedInputDescriptor());
       }
       getUnknownFields().writeTo(output);
     }
 
-    private int memoizedSerializedSize = -1;
+    @java.lang.Override
     public int getSerializedSize() {
-      int size = memoizedSerializedSize;
+      int size = memoizedSize;
       if (size != -1) return size;
 
       size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(1, getGroupNameBytes());
+      if (((bitField0_ & 0x00000001) != 0)) {
+        size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, groupName_);
       }
       {
         int dataSize = 0;
         for (int i = 0; i < groupVertices_.size(); i++) {
-          dataSize += com.google.protobuf.CodedOutputStream
-            .computeBytesSizeNoTag(groupVertices_.getByteString(i));
+          dataSize += computeStringSizeNoTag(groupVertices_.getRaw(i));
         }
         size += dataSize;
         size += 1 * getGroupVerticesList().size();
       }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
+      if (((bitField0_ & 0x00000002) != 0)) {
         size += com.google.protobuf.CodedOutputStream
-          .computeMessageSize(3, mergedInputDescriptor_);
+          .computeMessageSize(3, getMergedInputDescriptor());
       }
       size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
+      memoizedSize = size;
       return size;
     }
 
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
     @java.lang.Override
     public boolean equals(final java.lang.Object obj) {
       if (obj == this) {
@@ -2652,32 +2872,29 @@ public final class LlapDaemonProtocolProtos {
       }
       org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto) obj;
 
-      boolean result = true;
-      result = result && (hasGroupName() == other.hasGroupName());
+      if (hasGroupName() != other.hasGroupName()) return false;
       if (hasGroupName()) {
-        result = result && getGroupName()
-            .equals(other.getGroupName());
+        if (!getGroupName()
+            .equals(other.getGroupName())) return false;
       }
-      result = result && getGroupVerticesList()
-          .equals(other.getGroupVerticesList());
-      result = result && (hasMergedInputDescriptor() == other.hasMergedInputDescriptor());
+      if (!getGroupVerticesList()
+          .equals(other.getGroupVerticesList())) return false;
+      if (hasMergedInputDescriptor() != other.hasMergedInputDescriptor()) return false;
       if (hasMergedInputDescriptor()) {
-        result = result && getMergedInputDescriptor()
-            .equals(other.getMergedInputDescriptor());
+        if (!getMergedInputDescriptor()
+            .equals(other.getMergedInputDescriptor())) return false;
       }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
+      if (!getUnknownFields().equals(other.getUnknownFields())) return false;
+      return true;
     }
 
-    private int memoizedHashCode = 0;
     @java.lang.Override
     public int hashCode() {
       if (memoizedHashCode != 0) {
         return memoizedHashCode;
       }
       int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
+      hash = (19 * hash) + getDescriptor().hashCode();
       if (hasGroupName()) {
         hash = (37 * hash) + GROUP_NAME_FIELD_NUMBER;
         hash = (53 * hash) + getGroupName().hashCode();
@@ -2695,6 +2912,17 @@ public final class LlapDaemonProtocolProtos {
       return hash;
     }
 
+    public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom(
+        java.nio.ByteBuffer data)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      return PARSER.parseFrom(data);
+    }
+    public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom(
+        java.nio.ByteBuffer data,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      return PARSER.parseFrom(data, extensionRegistry);
+    }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom(
         com.google.protobuf.ByteString data)
         throws com.google.protobuf.InvalidProtocolBufferException {
@@ -2718,46 +2946,61 @@ public final class LlapDaemonProtocolProtos {
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return PARSER.parseFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input);
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input, extensionRegistry);
     }
+
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseDelimitedFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseDelimitedWithIOException(PARSER, input);
     }
+
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseDelimitedFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom(
         com.google.protobuf.CodedInputStream input)
         throws java.io.IOException {
-      return PARSER.parseFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input);
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom(
         com.google.protobuf.CodedInputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input, extensionRegistry);
     }
 
-    public static Builder newBuilder() { return Builder.create(); }
+    @java.lang.Override
     public Builder newBuilderForType() { return newBuilder(); }
+    public static Builder newBuilder() {
+      return DEFAULT_INSTANCE.toBuilder();
+    }
     public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto prototype) {
-      return newBuilder().mergeFrom(prototype);
+      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+    }
+    @java.lang.Override
+    public Builder toBuilder() {
+      return this == DEFAULT_INSTANCE
+          ? new Builder() : new Builder().mergeFrom(this);
     }
-    public Builder toBuilder() { return newBuilder(this); }
 
     @java.lang.Override
     protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+        com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
       Builder builder = new Builder(parent);
       return builder;
     }
@@ -2765,14 +3008,16 @@ public final class LlapDaemonProtocolProtos {
      * Protobuf type {@code GroupInputSpecProto}
      */
     public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder {
+        com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+        // @@protoc_insertion_point(builder_implements:GroupInputSpecProto)
+        org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder {
       public static final com.google.protobuf.Descriptors.Descriptor
           getDescriptor() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GroupInputSpecProto_descriptor;
       }
 
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+      @java.lang.Override
+      protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
           internalGetFieldAccessorTable() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GroupInputSpecProto_fieldAccessorTable
             .ensureFieldAccessorsInitialized(
@@ -2785,47 +3030,43 @@ public final class LlapDaemonProtocolProtos {
       }
 
       private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+          com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
         super(parent);
         maybeForceBuilderInitialization();
       }
       private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+        if (com.google.protobuf.GeneratedMessageV3
+                .alwaysUseFieldBuilders) {
           getMergedInputDescriptorFieldBuilder();
         }
       }
-      private static Builder create() {
-        return new Builder();
-      }
-
+      @java.lang.Override
       public Builder clear() {
         super.clear();
+        bitField0_ = 0;
         groupName_ = "";
-        bitField0_ = (bitField0_ & ~0x00000001);
-        groupVertices_ = com.google.protobuf.LazyStringArrayList.EMPTY;
-        bitField0_ = (bitField0_ & ~0x00000002);
-        if (mergedInputDescriptorBuilder_ == null) {
-          mergedInputDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
-        } else {
-          mergedInputDescriptorBuilder_.clear();
+        groupVertices_ =
+            com.google.protobuf.LazyStringArrayList.emptyList();
+        mergedInputDescriptor_ = null;
+        if (mergedInputDescriptorBuilder_ != null) {
+          mergedInputDescriptorBuilder_.dispose();
+          mergedInputDescriptorBuilder_ = null;
         }
-        bitField0_ = (bitField0_ & ~0x00000004);
         return this;
       }
 
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
+      @java.lang.Override
       public com.google.protobuf.Descriptors.Descriptor
           getDescriptorForType() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GroupInputSpecProto_descriptor;
       }
 
+      @java.lang.Override
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto getDefaultInstanceForType() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.getDefaultInstance();
       }
 
+      @java.lang.Override
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto build() {
         org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto result = buildPartial();
         if (!result.isInitialized()) {
@@ -2834,33 +3075,67 @@ public final class LlapDaemonProtocolProtos {
         return result;
       }
 
+      @java.lang.Override
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto buildPartial() {
         org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto(this);
+        if (bitField0_ != 0) { buildPartial0(result); }
+        onBuilt();
+        return result;
+      }
+
+      private void buildPartial0(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto result) {
         int from_bitField0_ = bitField0_;
         int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
+        if (((from_bitField0_ & 0x00000001) != 0)) {
+          result.groupName_ = groupName_;
           to_bitField0_ |= 0x00000001;
         }
-        result.groupName_ = groupName_;
-        if (((bitField0_ & 0x00000002) == 0x00000002)) {
-          groupVertices_ = new com.google.protobuf.UnmodifiableLazyStringList(
-              groupVertices_);
-          bitField0_ = (bitField0_ & ~0x00000002);
+        if (((from_bitField0_ & 0x00000002) != 0)) {
+          groupVertices_.makeImmutable();
+          result.groupVertices_ = groupVertices_;
         }
-        result.groupVertices_ = groupVertices_;
-        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
+        if (((from_bitField0_ & 0x00000004) != 0)) {
+          result.mergedInputDescriptor_ = mergedInputDescriptorBuilder_ == null
+              ? mergedInputDescriptor_
+              : mergedInputDescriptorBuilder_.build();
           to_bitField0_ |= 0x00000002;
         }
-        if (mergedInputDescriptorBuilder_ == null) {
-          result.mergedInputDescriptor_ = mergedInputDescriptor_;
-        } else {
-          result.mergedInputDescriptor_ = mergedInputDescriptorBuilder_.build();
-        }
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
+        result.bitField0_ |= to_bitField0_;
       }
 
+      @java.lang.Override
+      public Builder clone() {
+        return super.clone();
+      }
+      @java.lang.Override
+      public Builder setField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          java.lang.Object value) {
+        return super.setField(field, value);
+      }
+      @java.lang.Override
+      public Builder clearField(
+          com.google.protobuf.Descriptors.FieldDescriptor field) {
+        return super.clearField(field);
+      }
+      @java.lang.Override
+      public Builder clearOneof(
+          com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+        return super.clearOneof(oneof);
+      }
+      @java.lang.Override
+      public Builder setRepeatedField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          int index, java.lang.Object value) {
+        return super.setRepeatedField(field, index, value);
+      }
+      @java.lang.Override
+      public Builder addRepeatedField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          java.lang.Object value) {
+        return super.addRepeatedField(field, value);
+      }
+      @java.lang.Override
       public Builder mergeFrom(com.google.protobuf.Message other) {
         if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto) {
           return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto)other);
@@ -2873,14 +3148,14 @@ public final class LlapDaemonProtocolProtos {
       public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto other) {
         if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.getDefaultInstance()) return this;
         if (other.hasGroupName()) {
-          bitField0_ |= 0x00000001;
           groupName_ = other.groupName_;
+          bitField0_ |= 0x00000001;
           onChanged();
         }
         if (!other.groupVertices_.isEmpty()) {
           if (groupVertices_.isEmpty()) {
             groupVertices_ = other.groupVertices_;
-            bitField0_ = (bitField0_ & ~0x00000002);
+            bitField0_ |= 0x00000002;
           } else {
             ensureGroupVerticesIsMutable();
             groupVertices_.addAll(other.groupVertices_);
@@ -2891,49 +3166,87 @@ public final class LlapDaemonProtocolProtos {
           mergeMergedInputDescriptor(other.getMergedInputDescriptor());
         }
         this.mergeUnknownFields(other.getUnknownFields());
+        onChanged();
         return this;
       }
 
+      @java.lang.Override
       public final boolean isInitialized() {
         return true;
       }
 
+      @java.lang.Override
       public Builder mergeFrom(
           com.google.protobuf.CodedInputStream input,
           com.google.protobuf.ExtensionRegistryLite extensionRegistry)
           throws java.io.IOException {
-        org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parsedMessage = null;
+        if (extensionRegistry == null) {
+          throw new java.lang.NullPointerException();
+        }
         try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+          boolean done = false;
+          while (!done) {
+            int tag = input.readTag();
+            switch (tag) {
+              case 0:
+                done = true;
+                break;
+              case 10: {
+                groupName_ = input.readBytes();
+                bitField0_ |= 0x00000001;
+                break;
+              } // case 10
+              case 18: {
+                com.google.protobuf.ByteString bs = input.readBytes();
+                ensureGroupVerticesIsMutable();
+                groupVertices_.add(bs);
+                break;
+              } // case 18
+              case 26: {
+                input.readMessage(
+                    getMergedInputDescriptorFieldBuilder().getBuilder(),
+                    extensionRegistry);
+                bitField0_ |= 0x00000004;
+                break;
+              } // case 26
+              default: {
+                if (!super.parseUnknownField(input, extensionRegistry, tag)) {
+                  done = true; // was an endgroup tag
+                }
+                break;
+              } // default:
+            } // switch (tag)
+          } // while (!done)
         } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto) e.getUnfinishedMessage();
-          throw e;
+          throw e.unwrapIOException();
         } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
+          onChanged();
+        } // finally
         return this;
       }
       private int bitField0_;
 
-      // optional string group_name = 1;
       private java.lang.Object groupName_ = "";
       /**
        * <code>optional string group_name = 1;</code>
+       * @return Whether the groupName field is set.
        */
       public boolean hasGroupName() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
+        return ((bitField0_ & 0x00000001) != 0);
       }
       /**
        * <code>optional string group_name = 1;</code>
+       * @return The groupName.
        */
       public java.lang.String getGroupName() {
         java.lang.Object ref = groupName_;
         if (!(ref instanceof java.lang.String)) {
-          java.lang.String s = ((com.google.protobuf.ByteString) ref)
-              .toStringUtf8();
-          groupName_ = s;
+          com.google.protobuf.ByteString bs =
+              (com.google.protobuf.ByteString) ref;
+          java.lang.String s = bs.toStringUtf8();
+          if (bs.isValidUtf8()) {
+            groupName_ = s;
+          }
           return s;
         } else {
           return (java.lang.String) ref;
@@ -2941,6 +3254,7 @@ public final class LlapDaemonProtocolProtos {
       }
       /**
        * <code>optional string group_name = 1;</code>
+       * @return The bytes for groupName.
        */
       public com.google.protobuf.ByteString
           getGroupNameBytes() {
@@ -2957,69 +3271,77 @@ public final class LlapDaemonProtocolProtos {
       }
       /**
        * <code>optional string group_name = 1;</code>
+       * @param value The groupName to set.
+       * @return This builder for chaining.
        */
       public Builder setGroupName(
           java.lang.String value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000001;
+        if (value == null) { throw new NullPointerException(); }
         groupName_ = value;
+        bitField0_ |= 0x00000001;
         onChanged();
         return this;
       }
       /**
        * <code>optional string group_name = 1;</code>
+       * @return This builder for chaining.
        */
       public Builder clearGroupName() {
-        bitField0_ = (bitField0_ & ~0x00000001);
         groupName_ = getDefaultInstance().getGroupName();
+        bitField0_ = (bitField0_ & ~0x00000001);
         onChanged();
         return this;
       }
       /**
        * <code>optional string group_name = 1;</code>
+       * @param value The bytes for groupName to set.
+       * @return This builder for chaining.
        */
       public Builder setGroupNameBytes(
           com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000001;
+        if (value == null) { throw new NullPointerException(); }
         groupName_ = value;
+        bitField0_ |= 0x00000001;
         onChanged();
         return this;
       }
 
-      // repeated string group_vertices = 2;
-      private com.google.protobuf.LazyStringList groupVertices_ = com.google.protobuf.LazyStringArrayList.EMPTY;
+      private com.google.protobuf.LazyStringArrayList groupVertices_ =
+          com.google.protobuf.LazyStringArrayList.emptyList();
       private void ensureGroupVerticesIsMutable() {
-        if (!((bitField0_ & 0x00000002) == 0x00000002)) {
+        if (!groupVertices_.isModifiable()) {
           groupVertices_ = new com.google.protobuf.LazyStringArrayList(groupVertices_);
-          bitField0_ |= 0x00000002;
-         }
+        }
+        bitField0_ |= 0x00000002;
       }
       /**
        * <code>repeated string group_vertices = 2;</code>
+       * @return A list containing the groupVertices.
        */
-      public java.util.List<java.lang.String>
+      public com.google.protobuf.ProtocolStringList
           getGroupVerticesList() {
-        return java.util.Collections.unmodifiableList(groupVertices_);
+        groupVertices_.makeImmutable();
+        return groupVertices_;
       }
       /**
        * <code>repeated string group_vertices = 2;</code>
+       * @return The count of groupVertices.
        */
       public int getGroupVerticesCount() {
         return groupVertices_.size();
       }
       /**
        * <code>repeated string group_vertices = 2;</code>
+       * @param index The index of the element to return.
+       * @return The groupVertices at the given index.
        */
       public java.lang.String getGroupVertices(int index) {
         return groupVertices_.get(index);
       }
       /**
        * <code>repeated string group_vertices = 2;</code>
+       * @param index The index of the value to return.
+       * @return The bytes of the groupVertices at the given index.
        */
       public com.google.protobuf.ByteString
           getGroupVerticesBytes(int index) {
@@ -3027,79 +3349,90 @@ public final class LlapDaemonProtocolProtos {
       }
       /**
        * <code>repeated string group_vertices = 2;</code>
+       * @param index The index to set the value at.
+       * @param value The groupVertices to set.
+       * @return This builder for chaining.
        */
       public Builder setGroupVertices(
           int index, java.lang.String value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  ensureGroupVerticesIsMutable();
+        if (value == null) { throw new NullPointerException(); }
+        ensureGroupVerticesIsMutable();
         groupVertices_.set(index, value);
+        bitField0_ |= 0x00000002;
         onChanged();
         return this;
       }
       /**
        * <code>repeated string group_vertices = 2;</code>
+       * @param value The groupVertices to add.
+       * @return This builder for chaining.
        */
       public Builder addGroupVertices(
           java.lang.String value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  ensureGroupVerticesIsMutable();
+        if (value == null) { throw new NullPointerException(); }
+        ensureGroupVerticesIsMutable();
         groupVertices_.add(value);
+        bitField0_ |= 0x00000002;
         onChanged();
         return this;
       }
       /**
        * <code>repeated string group_vertices = 2;</code>
+       * @param values The groupVertices to add.
+       * @return This builder for chaining.
        */
       public Builder addAllGroupVertices(
           java.lang.Iterable<java.lang.String> values) {
         ensureGroupVerticesIsMutable();
-        super.addAll(values, groupVertices_);
+        com.google.protobuf.AbstractMessageLite.Builder.addAll(
+            values, groupVertices_);
+        bitField0_ |= 0x00000002;
         onChanged();
         return this;
       }
       /**
        * <code>repeated string group_vertices = 2;</code>
+       * @return This builder for chaining.
        */
       public Builder clearGroupVertices() {
-        groupVertices_ = com.google.protobuf.LazyStringArrayList.EMPTY;
-        bitField0_ = (bitField0_ & ~0x00000002);
+        groupVertices_ =
+          com.google.protobuf.LazyStringArrayList.emptyList();
+        bitField0_ = (bitField0_ & ~0x00000002);;
         onChanged();
         return this;
       }
       /**
        * <code>repeated string group_vertices = 2;</code>
+       * @param value The bytes of the groupVertices to add.
+       * @return This builder for chaining.
        */
       public Builder addGroupVerticesBytes(
           com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  ensureGroupVerticesIsMutable();
+        if (value == null) { throw new NullPointerException(); }
+        ensureGroupVerticesIsMutable();
         groupVertices_.add(value);
+        bitField0_ |= 0x00000002;
         onChanged();
         return this;
       }
 
-      // optional .EntityDescriptorProto merged_input_descriptor = 3;
-      private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto mergedInputDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
-      private com.google.protobuf.SingleFieldBuilder<
+      private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto mergedInputDescriptor_;
+      private com.google.protobuf.SingleFieldBuilderV3<
           org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> mergedInputDescriptorBuilder_;
       /**
        * <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code>
+       * @return Whether the mergedInputDescriptor field is set.
        */
       public boolean hasMergedInputDescriptor() {
-        return ((bitField0_ & 0x00000004) == 0x00000004);
+        return ((bitField0_ & 0x00000004) != 0);
       }
       /**
        * <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code>
+       * @return The mergedInputDescriptor.
        */
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getMergedInputDescriptor() {
         if (mergedInputDescriptorBuilder_ == null) {
-          return mergedInputDescriptor_;
+          return mergedInputDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : mergedInputDescriptor_;
         } else {
           return mergedInputDescriptorBuilder_.getMessage();
         }
@@ -3113,11 +3446,11 @@ public final class LlapDaemonProtocolProtos {
             throw new NullPointerException();
           }
           mergedInputDescriptor_ = value;
-          onChanged();
         } else {
           mergedInputDescriptorBuilder_.setMessage(value);
         }
         bitField0_ |= 0x00000004;
+        onChanged();
         return this;
       }
       /**
@@ -3127,11 +3460,11 @@ public final class LlapDaemonProtocolProtos {
           org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder builderForValue) {
         if (mergedInputDescriptorBuilder_ == null) {
           mergedInputDescriptor_ = builderForValue.build();
-          onChanged();
         } else {
           mergedInputDescriptorBuilder_.setMessage(builderForValue.build());
         }
         bitField0_ |= 0x00000004;
+        onChanged();
         return this;
       }
       /**
@@ -3139,31 +3472,33 @@ public final class LlapDaemonProtocolProtos {
        */
       public Builder mergeMergedInputDescriptor(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto value) {
         if (mergedInputDescriptorBuilder_ == null) {
-          if (((bitField0_ & 0x00000004) == 0x00000004) &&
-              mergedInputDescriptor_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) {
-            mergedInputDescriptor_ =
-              org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.newBuilder(mergedInputDescriptor_).mergeFrom(value).buildPartial();
+          if (((bitField0_ & 0x00000004) != 0) &&
+            mergedInputDescriptor_ != null &&
+            mergedInputDescriptor_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) {
+            getMergedInputDescriptorBuilder().mergeFrom(value);
           } else {
             mergedInputDescriptor_ = value;
           }
-          onChanged();
         } else {
           mergedInputDescriptorBuilder_.mergeFrom(value);
         }
-        bitField0_ |= 0x00000004;
+        if (mergedInputDescriptor_ != null) {
+          bitField0_ |= 0x00000004;
+          onChanged();
+        }
         return this;
       }
       /**
        * <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code>
        */
       public Builder clearMergedInputDescriptor() {
-        if (mergedInputDescriptorBuilder_ == null) {
-          mergedInputDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
-          onChanged();
-        } else {
-          mergedInputDescriptorBuilder_.clear();
-        }
         bitField0_ = (bitField0_ & ~0x00000004);
+        mergedInputDescriptor_ = null;
+        if (mergedInputDescriptorBuilder_ != null) {
+          mergedInputDescriptorBuilder_.dispose();
+          mergedInputDescriptorBuilder_ = null;
+        }
+        onChanged();
         return this;
       }
       /**
@@ -3181,72 +3516,130 @@ public final class LlapDaemonProtocolProtos {
         if (mergedInputDescriptorBuilder_ != null) {
           return mergedInputDescriptorBuilder_.getMessageOrBuilder();
         } else {
-          return mergedInputDescriptor_;
+          return mergedInputDescriptor_ == null ?
+              org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : mergedInputDescriptor_;
         }
       }
       /**
        * <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code>
        */
-      private com.google.protobuf.SingleFieldBuilder<
+      private com.google.protobuf.SingleFieldBuilderV3<
           org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> 
           getMergedInputDescriptorFieldBuilder() {
         if (mergedInputDescriptorBuilder_ == null) {
-          mergedInputDescriptorBuilder_ = new com.google.protobuf.SingleFieldBuilder<
+          mergedInputDescriptorBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
               org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder>(
-                  mergedInputDescriptor_,
+                  getMergedInputDescriptor(),
                   getParentForChildren(),
                   isClean());
           mergedInputDescriptor_ = null;
         }
         return mergedInputDescriptorBuilder_;
       }
+      @java.lang.Override
+      public final Builder setUnknownFields(
+          final com.google.protobuf.UnknownFieldSet unknownFields) {
+        return super.setUnknownFields(unknownFields);
+      }
+
+      @java.lang.Override
+      public final Builder mergeUnknownFields(
+          final com.google.protobuf.UnknownFieldSet unknownFields) {
+        return super.mergeUnknownFields(unknownFields);
+      }
+
 
       // @@protoc_insertion_point(builder_scope:GroupInputSpecProto)
     }
 
+    // @@protoc_insertion_point(class_scope:GroupInputSpecProto)
+    private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto DEFAULT_INSTANCE;
     static {
-      defaultInstance = new GroupInputSpecProto(true);
-      defaultInstance.initFields();
+      DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto();
+    }
+
+    public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto getDefaultInstance() {
+      return DEFAULT_INSTANCE;
+    }
+
+    @java.lang.Deprecated public static final com.google.protobuf.Parser<GroupInputSpecProto>
+        PARSER = new com.google.protobuf.AbstractParser<GroupInputSpecProto>() {
+      @java.lang.Override
+      public GroupInputSpecProto parsePartialFrom(
+          com.google.protobuf.CodedInputStream input,
+          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+          throws com.google.protobuf.InvalidProtocolBufferException {
+        Builder builder = newBuilder();
+        try {
+          builder.mergeFrom(input, extensionRegistry);
+        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+          throw e.setUnfinishedMessage(builder.buildPartial());
+        } catch (com.google.protobuf.UninitializedMessageException e) {
+          throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
+        } catch (java.io.IOException e) {
+          throw new com.google.protobuf.InvalidProtocolBufferException(e)
+              .setUnfinishedMessage(builder.buildPartial());
+        }
+        return builder.buildPartial();
+      }
+    };
+
+    public static com.google.protobuf.Parser<GroupInputSpecProto> parser() {
+      return PARSER;
+    }
+
+    @java.lang.Override
+    public com.google.protobuf.Parser<GroupInputSpecProto> getParserForType() {
+      return PARSER;
+    }
+
+    @java.lang.Override
+    public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto getDefaultInstanceForType() {
+      return DEFAULT_INSTANCE;
     }
 
-    // @@protoc_insertion_point(class_scope:GroupInputSpecProto)
   }
 
-  public interface SignableVertexSpecOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
+  public interface SignableVertexSpecOrBuilder extends
+      // @@protoc_insertion_point(interface_extends:SignableVertexSpec)
+      com.google.protobuf.MessageOrBuilder {
 
-    // optional string user = 1;
     /**
      * <code>optional string user = 1;</code>
+     * @return Whether the user field is set.
      */
     boolean hasUser();
     /**
      * <code>optional string user = 1;</code>
+     * @return The user.
      */
     java.lang.String getUser();
     /**
      * <code>optional string user = 1;</code>
+     * @return The bytes for user.
      */
     com.google.protobuf.ByteString
         getUserBytes();
 
-    // optional int64 signatureKeyId = 2;
     /**
      * <code>optional int64 signatureKeyId = 2;</code>
+     * @return Whether the signatureKeyId field is set.
      */
     boolean hasSignatureKeyId();
     /**
      * <code>optional int64 signatureKeyId = 2;</code>
+     * @return The signatureKeyId.
      */
     long getSignatureKeyId();
 
-    // optional .QueryIdentifierProto query_identifier = 3;
     /**
      * <code>optional .QueryIdentifierProto query_identifier = 3;</code>
+     * @return Whether the queryIdentifier field is set.
      */
     boolean hasQueryIdentifier();
     /**
      * <code>optional .QueryIdentifierProto query_identifier = 3;</code>
+     * @return The queryIdentifier.
      */
     org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier();
     /**
@@ -3254,107 +3647,117 @@ public final class LlapDaemonProtocolProtos {
      */
     org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder();
 
-    // optional string hive_query_id = 4;
     /**
      * <code>optional string hive_query_id = 4;</code>
+     * @return Whether the hiveQueryId field is set.
      */
     boolean hasHiveQueryId();
     /**
      * <code>optional string hive_query_id = 4;</code>
+     * @return The hiveQueryId.
      */
     java.lang.String getHiveQueryId();
     /**
      * <code>optional string hive_query_id = 4;</code>
+     * @return The bytes for hiveQueryId.
      */
     com.google.protobuf.ByteString
         getHiveQueryIdBytes();
 
-    // optional string dag_name = 5;
     /**
-     * <code>optional string dag_name = 5;</code>
-     *
      * <pre>
      * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
      * </pre>
+     *
+     * <code>optional string dag_name = 5;</code>
+     * @return Whether the dagName field is set.
      */
     boolean hasDagName();
     /**
-     * <code>optional string dag_name = 5;</code>
-     *
      * <pre>
      * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
      * </pre>
+     *
+     * <code>optional string dag_name = 5;</code>
+     * @return The dagName.
      */
     java.lang.String getDagName();
     /**
-     * <code>optional string dag_name = 5;</code>
-     *
      * <pre>
      * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
      * </pre>
+     *
+     * <code>optional string dag_name = 5;</code>
+     * @return The bytes for dagName.
      */
     com.google.protobuf.ByteString
         getDagNameBytes();
 
-    // optional string vertex_name = 6;
     /**
      * <code>optional string vertex_name = 6;</code>
+     * @return Whether the vertexName field is set.
      */
     boolean hasVertexName();
     /**
      * <code>optional string vertex_name = 6;</code>
+     * @return The vertexName.
      */
     java.lang.String getVertexName();
     /**
      * <code>optional string vertex_name = 6;</code>
+     * @return The bytes for vertexName.
      */
     com.google.protobuf.ByteString
         getVertexNameBytes();
 
-    // optional int32 vertex_index = 7;
     /**
      * <code>optional int32 vertex_index = 7;</code>
+     * @return Whether the vertexIndex field is set.
      */
     boolean hasVertexIndex();
     /**
      * <code>optional int32 vertex_index = 7;</code>
+     * @return The vertexIndex.
      */
     int getVertexIndex();
 
-    // optional string token_identifier = 8;
     /**
-     * <code>optional string token_identifier = 8;</code>
-     *
      * <pre>
      * The core vertex stuff 
      * </pre>
+     *
+     * <code>optional string token_identifier = 8;</code>
+     * @return Whether the tokenIdentifier field is set.
      */
     boolean hasTokenIdentifier();
     /**
-     * <code>optional string token_identifier = 8;</code>
-     *
      * <pre>
      * The core vertex stuff 
      * </pre>
+     *
+     * <code>optional string token_identifier = 8;</code>
+     * @return The tokenIdentifier.
      */
     java.lang.String getTokenIdentifier();
     /**
-     * <code>optional string token_identifier = 8;</code>
-     *
      * <pre>
      * The core vertex stuff 
      * </pre>
+     *
+     * <code>optional string token_identifier = 8;</code>
+     * @return The bytes for tokenIdentifier.
      */
     com.google.protobuf.ByteString
         getTokenIdentifierBytes();
 
-    // optional .EntityDescriptorProto processor_descriptor = 9;
     /**
      * <code>optional .EntityDescriptorProto processor_descriptor = 9;</code>
+     * @return Whether the processorDescriptor field is set.
      */
     boolean hasProcessorDescriptor();
     /**
      * <code>optional .EntityDescriptorProto processor_descriptor = 9;</code>
+     * @return The processorDescriptor.
      */
     org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getProcessorDescriptor();
     /**
@@ -3362,7 +3765,6 @@ public final class LlapDaemonProtocolProtos {
      */
     org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getProcessorDescriptorOrBuilder();
 
-    // repeated .IOSpecProto input_specs = 10;
     /**
      * <code>repeated .IOSpecProto input_specs = 10;</code>
      */
@@ -3387,7 +3789,6 @@ public final class LlapDaemonProtocolProtos {
     org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder getInputSpecsOrBuilder(
         int index);
 
-    // repeated .IOSpecProto output_specs = 11;
     /**
      * <code>repeated .IOSpecProto output_specs = 11;</code>
      */
@@ -3412,7 +3813,6 @@ public final class LlapDaemonProtocolProtos {
     org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder getOutputSpecsOrBuilder(
         int index);
 
-    // repeated .GroupInputSpecProto grouped_input_specs = 12;
     /**
      * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code>
      */
@@ -3437,245 +3837,100 @@ public final class LlapDaemonProtocolProtos {
     org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder getGroupedInputSpecsOrBuilder(
         int index);
 
-    // optional int32 vertex_parallelism = 13;
     /**
-     * <code>optional int32 vertex_parallelism = 13;</code>
-     *
      * <pre>
      * An internal field required for Tez.
      * </pre>
+     *
+     * <code>optional int32 vertex_parallelism = 13;</code>
+     * @return Whether the vertexParallelism field is set.
      */
     boolean hasVertexParallelism();
     /**
-     * <code>optional int32 vertex_parallelism = 13;</code>
-     *
      * <pre>
      * An internal field required for Tez.
      * </pre>
+     *
+     * <code>optional int32 vertex_parallelism = 13;</code>
+     * @return The vertexParallelism.
      */
     int getVertexParallelism();
 
-    // optional bool is_external_submission = 14 [default = false];
     /**
      * <code>optional bool is_external_submission = 14 [default = false];</code>
+     * @return Whether the isExternalSubmission field is set.
      */
     boolean hasIsExternalSubmission();
     /**
      * <code>optional bool is_external_submission = 14 [default = false];</code>
+     * @return The isExternalSubmission.
      */
     boolean getIsExternalSubmission();
   }
   /**
-   * Protobuf type {@code SignableVertexSpec}
-   *
    * <pre>
    * The part of SubmitWork that can be signed 
    * </pre>
+   *
+   * Protobuf type {@code SignableVertexSpec}
    */
   public static final class SignableVertexSpec extends
-      com.google.protobuf.GeneratedMessage
-      implements SignableVertexSpecOrBuilder {
+      com.google.protobuf.GeneratedMessageV3 implements
+      // @@protoc_insertion_point(message_implements:SignableVertexSpec)
+      SignableVertexSpecOrBuilder {
+  private static final long serialVersionUID = 0L;
     // Use SignableVertexSpec.newBuilder() to construct.
-    private SignableVertexSpec(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+    private SignableVertexSpec(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
       super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private SignableVertexSpec(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final SignableVertexSpec defaultInstance;
-    public static SignableVertexSpec getDefaultInstance() {
-      return defaultInstance;
     }
-
-    public SignableVertexSpec getDefaultInstanceForType() {
-      return defaultInstance;
+    private SignableVertexSpec() {
+      user_ = "";
+      hiveQueryId_ = "";
+      dagName_ = "";
+      vertexName_ = "";
+      tokenIdentifier_ = "";
+      inputSpecs_ = java.util.Collections.emptyList();
+      outputSpecs_ = java.util.Collections.emptyList();
+      groupedInputSpecs_ = java.util.Collections.emptyList();
     }
 
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
     @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private SignableVertexSpec(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              bitField0_ |= 0x00000001;
-              user_ = input.readBytes();
-              break;
-            }
-            case 16: {
-              bitField0_ |= 0x00000002;
-              signatureKeyId_ = input.readInt64();
-              break;
-            }
-            case 26: {
-              org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder subBuilder = null;
-              if (((bitField0_ & 0x00000004) == 0x00000004)) {
-                subBuilder = queryIdentifier_.toBuilder();
-              }
-              queryIdentifier_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.PARSER, extensionRegistry);
-              if (subBuilder != null) {
-                subBuilder.mergeFrom(queryIdentifier_);
-                queryIdentifier_ = subBuilder.buildPartial();
-              }
-              bitField0_ |= 0x00000004;
-              break;
-            }
-            case 34: {
-              bitField0_ |= 0x00000008;
-              hiveQueryId_ = input.readBytes();
-              break;
-            }
-            case 42: {
-              bitField0_ |= 0x00000010;
-              dagName_ = input.readBytes();
-              break;
-            }
-            case 50: {
-              bitField0_ |= 0x00000020;
-              vertexName_ = input.readBytes();
-              break;
-            }
-            case 56: {
-              bitField0_ |= 0x00000040;
-              vertexIndex_ = input.readInt32();
-              break;
-            }
-            case 66: {
-              bitField0_ |= 0x00000080;
-              tokenIdentifier_ = input.readBytes();
-              break;
-            }
-            case 74: {
-              org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder subBuilder = null;
-              if (((bitField0_ & 0x00000100) == 0x00000100)) {
-                subBuilder = processorDescriptor_.toBuilder();
-              }
-              processorDescriptor_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.PARSER, extensionRegistry);
-              if (subBuilder != null) {
-                subBuilder.mergeFrom(processorDescriptor_);
-                processorDescriptor_ = subBuilder.buildPartial();
-              }
-              bitField0_ |= 0x00000100;
-              break;
-            }
-            case 82: {
-              if (!((mutable_bitField0_ & 0x00000200) == 0x00000200)) {
-                inputSpecs_ = new java.util.ArrayList<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto>();
-                mutable_bitField0_ |= 0x00000200;
-              }
-              inputSpecs_.add(input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.PARSER, extensionRegistry));
-              break;
-            }
-            case 90: {
-              if (!((mutable_bitField0_ & 0x00000400) == 0x00000400)) {
-                outputSpecs_ = new java.util.ArrayList<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto>();
-                mutable_bitField0_ |= 0x00000400;
-              }
-              outputSpecs_.add(input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.PARSER, extensionRegistry));
-              break;
-            }
-            case 98: {
-              if (!((mutable_bitField0_ & 0x00000800) == 0x00000800)) {
-                groupedInputSpecs_ = new java.util.ArrayList<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto>();
-                mutable_bitField0_ |= 0x00000800;
-              }
-              groupedInputSpecs_.add(input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.PARSER, extensionRegistry));
-              break;
-            }
-            case 104: {
-              bitField0_ |= 0x00000200;
-              vertexParallelism_ = input.readInt32();
-              break;
-            }
-            case 112: {
-              bitField0_ |= 0x00000400;
-              isExternalSubmission_ = input.readBool();
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        if (((mutable_bitField0_ & 0x00000200) == 0x00000200)) {
-          inputSpecs_ = java.util.Collections.unmodifiableList(inputSpecs_);
-        }
-        if (((mutable_bitField0_ & 0x00000400) == 0x00000400)) {
-          outputSpecs_ = java.util.Collections.unmodifiableList(outputSpecs_);
-        }
-        if (((mutable_bitField0_ & 0x00000800) == 0x00000800)) {
-          groupedInputSpecs_ = java.util.Collections.unmodifiableList(groupedInputSpecs_);
-        }
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
+    @SuppressWarnings({"unused"})
+    protected java.lang.Object newInstance(
+        UnusedPrivateParameter unused) {
+      return new SignableVertexSpec();
     }
+
     public static final com.google.protobuf.Descriptors.Descriptor
         getDescriptor() {
       return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SignableVertexSpec_descriptor;
     }
 
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+    @java.lang.Override
+    protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
         internalGetFieldAccessorTable() {
       return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SignableVertexSpec_fieldAccessorTable
           .ensureFieldAccessorsInitialized(
               org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder.class);
     }
 
-    public static com.google.protobuf.Parser<SignableVertexSpec> PARSER =
-        new com.google.protobuf.AbstractParser<SignableVertexSpec>() {
-      public SignableVertexSpec parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new SignableVertexSpec(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<SignableVertexSpec> getParserForType() {
-      return PARSER;
-    }
-
     private int bitField0_;
-    // optional string user = 1;
     public static final int USER_FIELD_NUMBER = 1;
-    private java.lang.Object user_;
+    @SuppressWarnings("serial")
+    private volatile java.lang.Object user_ = "";
     /**
      * <code>optional string user = 1;</code>
+     * @return Whether the user field is set.
      */
+    @java.lang.Override
     public boolean hasUser() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
+      return ((bitField0_ & 0x00000001) != 0);
     }
     /**
      * <code>optional string user = 1;</code>
+     * @return The user.
      */
+    @java.lang.Override
     public java.lang.String getUser() {
       java.lang.Object ref = user_;
       if (ref instanceof java.lang.String) {
@@ -3692,7 +3947,9 @@ public final class LlapDaemonProtocolProtos {
     }
     /**
      * <code>optional string user = 1;</code>
+     * @return The bytes for user.
      */
+    @java.lang.Override
     public com.google.protobuf.ByteString
         getUserBytes() {
       java.lang.Object ref = user_;
@@ -3707,56 +3964,67 @@ public final class LlapDaemonProtocolProtos {
       }
     }
 
-    // optional int64 signatureKeyId = 2;
     public static final int SIGNATUREKEYID_FIELD_NUMBER = 2;
-    private long signatureKeyId_;
+    private long signatureKeyId_ = 0L;
     /**
      * <code>optional int64 signatureKeyId = 2;</code>
+     * @return Whether the signatureKeyId field is set.
      */
+    @java.lang.Override
     public boolean hasSignatureKeyId() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
+      return ((bitField0_ & 0x00000002) != 0);
     }
     /**
      * <code>optional int64 signatureKeyId = 2;</code>
+     * @return The signatureKeyId.
      */
+    @java.lang.Override
     public long getSignatureKeyId() {
       return signatureKeyId_;
     }
 
-    // optional .QueryIdentifierProto query_identifier = 3;
     public static final int QUERY_IDENTIFIER_FIELD_NUMBER = 3;
     private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_;
     /**
      * <code>optional .QueryIdentifierProto query_identifier = 3;</code>
+     * @return Whether the queryIdentifier field is set.
      */
+    @java.lang.Override
     public boolean hasQueryIdentifier() {
-      return ((bitField0_ & 0x00000004) == 0x00000004);
+      return ((bitField0_ & 0x00000004) != 0);
     }
     /**
      * <code>optional .QueryIdentifierProto query_identifier = 3;</code>
+     * @return The queryIdentifier.
      */
+    @java.lang.Override
     public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() {
-      return queryIdentifier_;
+      return queryIdentifier_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_;
     }
     /**
      * <code>optional .QueryIdentifierProto query_identifier = 3;</code>
      */
+    @java.lang.Override
     public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder() {
-      return queryIdentifier_;
+      return queryIdentifier_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_;
     }
 
-    // optional string hive_query_id = 4;
     public static final int HIVE_QUERY_ID_FIELD_NUMBER = 4;
-    private java.lang.Object hiveQueryId_;
+    @SuppressWarnings("serial")
+    private volatile java.lang.Object hiveQueryId_ = "";
     /**
      * <code>optional string hive_query_id = 4;</code>
+     * @return Whether the hiveQueryId field is set.
      */
+    @java.lang.Override
     public boolean hasHiveQueryId() {
-      return ((bitField0_ & 0x00000008) == 0x00000008);
+      return ((bitField0_ & 0x00000008) != 0);
     }
     /**
      * <code>optional string hive_query_id = 4;</code>
+     * @return The hiveQueryId.
      */
+    @java.lang.Override
     public java.lang.String getHiveQueryId() {
       java.lang.Object ref = hiveQueryId_;
       if (ref instanceof java.lang.String) {
@@ -3773,7 +4041,9 @@ public final class LlapDaemonProtocolProtos {
     }
     /**
      * <code>optional string hive_query_id = 4;</code>
+     * @return The bytes for hiveQueryId.
      */
+    @java.lang.Override
     public com.google.protobuf.ByteString
         getHiveQueryIdBytes() {
       java.lang.Object ref = hiveQueryId_;
@@ -3788,26 +4058,30 @@ public final class LlapDaemonProtocolProtos {
       }
     }
 
-    // optional string dag_name = 5;
     public static final int DAG_NAME_FIELD_NUMBER = 5;
-    private java.lang.Object dagName_;
+    @SuppressWarnings("serial")
+    private volatile java.lang.Object dagName_ = "";
     /**
-     * <code>optional string dag_name = 5;</code>
-     *
      * <pre>
      * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
      * </pre>
+     *
+     * <code>optional string dag_name = 5;</code>
+     * @return Whether the dagName field is set.
      */
+    @java.lang.Override
     public boolean hasDagName() {
-      return ((bitField0_ & 0x00000010) == 0x00000010);
+      return ((bitField0_ & 0x00000010) != 0);
     }
     /**
-     * <code>optional string dag_name = 5;</code>
-     *
      * <pre>
      * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
      * </pre>
+     *
+     * <code>optional string dag_name = 5;</code>
+     * @return The dagName.
      */
+    @java.lang.Override
     public java.lang.String getDagName() {
       java.lang.Object ref = dagName_;
       if (ref instanceof java.lang.String) {
@@ -3823,12 +4097,14 @@ public final class LlapDaemonProtocolProtos {
       }
     }
     /**
-     * <code>optional string dag_name = 5;</code>
-     *
      * <pre>
      * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
      * </pre>
+     *
+     * <code>optional string dag_name = 5;</code>
+     * @return The bytes for dagName.
      */
+    @java.lang.Override
     public com.google.protobuf.ByteString
         getDagNameBytes() {
       java.lang.Object ref = dagName_;
@@ -3843,18 +4119,22 @@ public final class LlapDaemonProtocolProtos {
       }
     }
 
-    // optional string vertex_name = 6;
     public static final int VERTEX_NAME_FIELD_NUMBER = 6;
-    private java.lang.Object vertexName_;
+    @SuppressWarnings("serial")
+    private volatile java.lang.Object vertexName_ = "";
     /**
      * <code>optional string vertex_name = 6;</code>
+     * @return Whether the vertexName field is set.
      */
+    @java.lang.Override
     public boolean hasVertexName() {
-      return ((bitField0_ & 0x00000020) == 0x00000020);
+      return ((bitField0_ & 0x00000020) != 0);
     }
     /**
      * <code>optional string vertex_name = 6;</code>
+     * @return The vertexName.
      */
+    @java.lang.Override
     public java.lang.String getVertexName() {
       java.lang.Object ref = vertexName_;
       if (ref instanceof java.lang.String) {
@@ -3871,7 +4151,9 @@ public final class LlapDaemonProtocolProtos {
     }
     /**
      * <code>optional string vertex_name = 6;</code>
+     * @return The bytes for vertexName.
      */
+    @java.lang.Override
     public com.google.protobuf.ByteString
         getVertexNameBytes() {
       java.lang.Object ref = vertexName_;
@@ -3886,42 +4168,49 @@ public final class LlapDaemonProtocolProtos {
       }
     }
 
-    // optional int32 vertex_index = 7;
     public static final int VERTEX_INDEX_FIELD_NUMBER = 7;
-    private int vertexIndex_;
+    private int vertexIndex_ = 0;
     /**
      * <code>optional int32 vertex_index = 7;</code>
+     * @return Whether the vertexIndex field is set.
      */
+    @java.lang.Override
     public boolean hasVertexIndex() {
-      return ((bitField0_ & 0x00000040) == 0x00000040);
+      return ((bitField0_ & 0x00000040) != 0);
     }
     /**
      * <code>optional int32 vertex_index = 7;</code>
+     * @return The vertexIndex.
      */
+    @java.lang.Override
     public int getVertexIndex() {
       return vertexIndex_;
     }
 
-    // optional string token_identifier = 8;
     public static final int TOKEN_IDENTIFIER_FIELD_NUMBER = 8;
-    private java.lang.Object tokenIdentifier_;
+    @SuppressWarnings("serial")
+    private volatile java.lang.Object tokenIdentifier_ = "";
     /**
-     * <code>optional string token_identifier = 8;</code>
-     *
      * <pre>
      * The core vertex stuff 
      * </pre>
+     *
+     * <code>optional string token_identifier = 8;</code>
+     * @return Whether the tokenIdentifier field is set.
      */
+    @java.lang.Override
     public boolean hasTokenIdentifier() {
-      return ((bitField0_ & 0x00000080) == 0x00000080);
+      return ((bitField0_ & 0x00000080) != 0);
     }
     /**
-     * <code>optional string token_identifier = 8;</code>
-     *
      * <pre>
      * The core vertex stuff 
      * </pre>
+     *
+     * <code>optional string token_identifier = 8;</code>
+     * @return The tokenIdentifier.
      */
+    @java.lang.Override
     public java.lang.String getTokenIdentifier() {
       java.lang.Object ref = tokenIdentifier_;
       if (ref instanceof java.lang.String) {
@@ -3937,12 +4226,14 @@ public final class LlapDaemonProtocolProtos {
       }
     }
     /**
-     * <code>optional string token_identifier = 8;</code>
-     *
      * <pre>
      * The core vertex stuff 
      * </pre>
+     *
+     * <code>optional string token_identifier = 8;</code>
+     * @return The bytes for tokenIdentifier.
      */
+    @java.lang.Override
     public com.google.protobuf.ByteString
         getTokenIdentifierBytes() {
       java.lang.Object ref = tokenIdentifier_;
@@ -3957,40 +4248,46 @@ public final class LlapDaemonProtocolProtos {
       }
     }
 
-    // optional .EntityDescriptorProto processor_descriptor = 9;
     public static final int PROCESSOR_DESCRIPTOR_FIELD_NUMBER = 9;
     private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto processorDescriptor_;
     /**
      * <code>optional .EntityDescriptorProto processor_descriptor = 9;</code>
+     * @return Whether the processorDescriptor field is set.
      */
+    @java.lang.Override
     public boolean hasProcessorDescriptor() {
-      return ((bitField0_ & 0x00000100) == 0x00000100);
+      return ((bitField0_ & 0x00000100) != 0);
     }
     /**
      * <code>optional .EntityDescriptorProto processor_descriptor = 9;</code>
+     * @return The processorDescriptor.
      */
+    @java.lang.Override
     public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getProcessorDescriptor() {
-      return processorDescriptor_;
+      return processorDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : processorDescriptor_;
     }
     /**
      * <code>optional .EntityDescriptorProto processor_descriptor = 9;</code>
      */
+    @java.lang.Override
     public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getProcessorDescriptorOrBuilder() {
-      return processorDescriptor_;
+      return processorDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : processorDescriptor_;
     }
 
-    // repeated .IOSpecProto input_specs = 10;
     public static final int INPUT_SPECS_FIELD_NUMBER = 10;
+    @SuppressWarnings("serial")
     private java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> inputSpecs_;
     /**
      * <code>repeated .IOSpecProto input_specs = 10;</code>
      */
+    @java.lang.Override
     public java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> getInputSpecsList() {
       return inputSpecs_;
     }
     /**
      * <code>repeated .IOSpecProto input_specs = 10;</code>
      */
+    @java.lang.Override
     public java.util.List<? extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder> 
         getInputSpecsOrBuilderList() {
       return inputSpecs_;
@@ -3998,35 +4295,40 @@ public final class LlapDaemonProtocolProtos {
     /**
      * <code>repeated .IOSpecProto input_specs = 10;</code>
      */
+    @java.lang.Override
     public int getInputSpecsCount() {
       return inputSpecs_.size();
     }
     /**
      * <code>repeated .IOSpecProto input_specs = 10;</code>
      */
+    @java.lang.Override
     public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getInputSpecs(int index) {
       return inputSpecs_.get(index);
     }
     /**
      * <code>repeated .IOSpecProto input_specs = 10;</code>
      */
+    @java.lang.Override
     public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder getInputSpecsOrBuilder(
         int index) {
       return inputSpecs_.get(index);
     }
 
-    // repeated .IOSpecProto output_specs = 11;
     public static final int OUTPUT_SPECS_FIELD_NUMBER = 11;
+    @SuppressWarnings("serial")
     private java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> outputSpecs_;
     /**
      * <code>repeated .IOSpecProto output_specs = 11;</code>
      */
+    @java.lang.Override
     public java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> getOutputSpecsList() {
       return outputSpecs_;
     }
     /**
      * <code>repeated .IOSpecProto output_specs = 11;</code>
      */
+    @java.lang.Override
     public java.util.List<? extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder> 
         getOutputSpecsOrBuilderList() {
       return outputSpecs_;
@@ -4034,35 +4336,40 @@ public final class LlapDaemonProtocolProtos {
     /**
      * <code>repeated .IOSpecProto output_specs = 11;</code>
      */
+    @java.lang.Override
     public int getOutputSpecsCount() {
       return outputSpecs_.size();
     }
     /**
      * <code>repeated .IOSpecProto output_specs = 11;</code>
      */
+    @java.lang.Override
     public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getOutputSpecs(int index) {
       return outputSpecs_.get(index);
     }
     /**
      * <code>repeated .IOSpecProto output_specs = 11;</code>
      */
+    @java.lang.Override
     public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder getOutputSpecsOrBuilder(
         int index) {
       return outputSpecs_.get(index);
     }
 
-    // repeated .GroupInputSpecProto grouped_input_specs = 12;
     public static final int GROUPED_INPUT_SPECS_FIELD_NUMBER = 12;
+    @SuppressWarnings("serial")
     private java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto> groupedInputSpecs_;
     /**
      * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code>
      */
+    @java.lang.Override
     public java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto> getGroupedInputSpecsList() {
       return groupedInputSpecs_;
     }
     /**
      * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code>
      */
+    @java.lang.Override
     public java.util.List<? extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder> 
         getGroupedInputSpecsOrBuilderList() {
       return groupedInputSpecs_;
@@ -4070,117 +4377,112 @@ public final class LlapDaemonProtocolProtos {
     /**
      * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code>
      */
+    @java.lang.Override
     public int getGroupedInputSpecsCount() {
       return groupedInputSpecs_.size();
     }
     /**
      * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code>
      */
+    @java.lang.Override
     public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto getGroupedInputSpecs(int index) {
       return groupedInputSpecs_.get(index);
     }
     /**
      * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code>
      */
+    @java.lang.Override
     public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder getGroupedInputSpecsOrBuilder(
         int index) {
       return groupedInputSpecs_.get(index);
     }
 
-    // optional int32 vertex_parallelism = 13;
     public static final int VERTEX_PARALLELISM_FIELD_NUMBER = 13;
-    private int vertexParallelism_;
+    private int vertexParallelism_ = 0;
     /**
-     * <code>optional int32 vertex_parallelism = 13;</code>
-     *
      * <pre>
      * An internal field required for Tez.
      * </pre>
+     *
+     * <code>optional int32 vertex_parallelism = 13;</code>
+     * @return Whether the vertexParallelism field is set.
      */
+    @java.lang.Override
     public boolean hasVertexParallelism() {
-      return ((bitField0_ & 0x00000200) == 0x00000200);
+      return ((bitField0_ & 0x00000200) != 0);
     }
     /**
-     * <code>optional int32 vertex_parallelism = 13;</code>
-     *
      * <pre>
      * An internal field required for Tez.
      * </pre>
+     *
+     * <code>optional int32 vertex_parallelism = 13;</code>
+     * @return The vertexParallelism.
      */
+    @java.lang.Override
     public int getVertexParallelism() {
       return vertexParallelism_;
     }
 
-    // optional bool is_external_submission = 14 [default = false];
     public static final int IS_EXTERNAL_SUBMISSION_FIELD_NUMBER = 14;
-    private boolean isExternalSubmission_;
+    private boolean isExternalSubmission_ = false;
     /**
      * <code>optional bool is_external_submission = 14 [default = false];</code>
+     * @return Whether the isExternalSubmission field is set.
      */
+    @java.lang.Override
     public boolean hasIsExternalSubmission() {
-      return ((bitField0_ & 0x00000400) == 0x00000400);
+      return ((bitField0_ & 0x00000400) != 0);
     }
     /**
      * <code>optional bool is_external_submission = 14 [default = false];</code>
+     * @return The isExternalSubmission.
      */
+    @java.lang.Override
     public boolean getIsExternalSubmission() {
       return isExternalSubmission_;
     }
 
-    private void initFields() {
-      user_ = "";
-      signatureKeyId_ = 0L;
-      queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance();
-      hiveQueryId_ = "";
-      dagName_ = "";
-      vertexName_ = "";
-      vertexIndex_ = 0;
-      tokenIdentifier_ = "";
-      processorDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
-      inputSpecs_ = java.util.Collections.emptyList();
-      outputSpecs_ = java.util.Collections.emptyList();
-      groupedInputSpecs_ = java.util.Collections.emptyList();
-      vertexParallelism_ = 0;
-      isExternalSubmission_ = false;
-    }
     private byte memoizedIsInitialized = -1;
+    @java.lang.Override
     public final boolean isInitialized() {
       byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
+      if (isInitialized == 1) return true;
+      if (isInitialized == 0) return false;
 
       memoizedIsInitialized = 1;
       return true;
     }
 
+    @java.lang.Override
     public void writeTo(com.google.protobuf.CodedOutputStream output)
                         throws java.io.IOException {
-      getSerializedSize();
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeBytes(1, getUserBytes());
+      if (((bitField0_ & 0x00000001) != 0)) {
+        com.google.protobuf.GeneratedMessageV3.writeString(output, 1, user_);
       }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
+      if (((bitField0_ & 0x00000002) != 0)) {
         output.writeInt64(2, signatureKeyId_);
       }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
-        output.writeMessage(3, queryIdentifier_);
+      if (((bitField0_ & 0x00000004) != 0)) {
+        output.writeMessage(3, getQueryIdentifier());
       }
-      if (((bitField0_ & 0x00000008) == 0x00000008)) {
-        output.writeBytes(4, getHiveQueryIdBytes());
+      if (((bitField0_ & 0x00000008) != 0)) {
+        com.google.protobuf.GeneratedMessageV3.writeString(output, 4, hiveQueryId_);
       }
-      if (((bitField0_ & 0x00000010) == 0x00000010)) {
-        output.writeBytes(5, getDagNameBytes());
+      if (((bitField0_ & 0x00000010) != 0)) {
+        com.google.protobuf.GeneratedMessageV3.writeString(output, 5, dagName_);
       }
-      if (((bitField0_ & 0x00000020) == 0x00000020)) {
-        output.writeBytes(6, getVertexNameBytes());
+      if (((bitField0_ & 0x00000020) != 0)) {
+        com.google.protobuf.GeneratedMessageV3.writeString(output, 6, vertexName_);
       }
-      if (((bitField0_ & 0x00000040) == 0x00000040)) {
+      if (((bitField0_ & 0x00000040) != 0)) {
         output.writeInt32(7, vertexIndex_);
       }
-      if (((bitField0_ & 0x00000080) == 0x00000080)) {
-        output.writeBytes(8, getTokenIdentifierBytes());
+      if (((bitField0_ & 0x00000080) != 0)) {
+        com.google.protobuf.GeneratedMessageV3.writeString(output, 8, tokenIdentifier_);
       }
-      if (((bitField0_ & 0x00000100) == 0x00000100)) {
-        output.writeMessage(9, processorDescriptor_);
+      if (((bitField0_ & 0x00000100) != 0)) {
+        output.writeMessage(9, getProcessorDescriptor());
       }
       for (int i = 0; i < inputSpecs_.size(); i++) {
         output.writeMessage(10, inputSpecs_.get(i));
@@ -4191,56 +4493,51 @@ public final class LlapDaemonProtocolProtos {
       for (int i = 0; i < groupedInputSpecs_.size(); i++) {
         output.writeMessage(12, groupedInputSpecs_.get(i));
       }
-      if (((bitField0_ & 0x00000200) == 0x00000200)) {
+      if (((bitField0_ & 0x00000200) != 0)) {
         output.writeInt32(13, vertexParallelism_);
       }
-      if (((bitField0_ & 0x00000400) == 0x00000400)) {
+      if (((bitField0_ & 0x00000400) != 0)) {
         output.writeBool(14, isExternalSubmission_);
       }
       getUnknownFields().writeTo(output);
     }
 
-    private int memoizedSerializedSize = -1;
+    @java.lang.Override
     public int getSerializedSize() {
-      int size = memoizedSerializedSize;
+      int size = memoizedSize;
       if (size != -1) return size;
 
       size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(1, getUserBytes());
+      if (((bitField0_ & 0x00000001) != 0)) {
+        size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, user_);
       }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
+      if (((bitField0_ & 0x00000002) != 0)) {
         size += com.google.protobuf.CodedOutputStream
           .computeInt64Size(2, signatureKeyId_);
       }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
+      if (((bitField0_ & 0x00000004) != 0)) {
         size += com.google.protobuf.CodedOutputStream
-          .computeMessageSize(3, queryIdentifier_);
+          .computeMessageSize(3, getQueryIdentifier());
       }
-      if (((bitField0_ & 0x00000008) == 0x00000008)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(4, getHiveQueryIdBytes());
+      if (((bitField0_ & 0x00000008) != 0)) {
+        size += com.google.protobuf.GeneratedMessageV3.computeStringSize(4, hiveQueryId_);
       }
-      if (((bitField0_ & 0x00000010) == 0x00000010)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(5, getDagNameBytes());
+      if (((bitField0_ & 0x00000010) != 0)) {
+        size += com.google.protobuf.GeneratedMessageV3.computeStringSize(5, dagName_);
       }
-      if (((bitField0_ & 0x00000020) == 0x00000020)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(6, getVertexNameBytes());
+      if (((bitField0_ & 0x00000020) != 0)) {
+        size += com.google.protobuf.GeneratedMessageV3.computeStringSize(6, vertexName_);
       }
-      if (((bitField0_ & 0x00000040) == 0x00000040)) {
+      if (((bitField0_ & 0x00000040) != 0)) {
         size += com.google.protobuf.CodedOutputStream
           .computeInt32Size(7, vertexIndex_);
       }
-      if (((bitField0_ & 0x00000080) == 0x00000080)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(8, getTokenIdentifierBytes());
+      if (((bitField0_ & 0x00000080) != 0)) {
+        size += com.google.protobuf.GeneratedMessageV3.computeStringSize(8, tokenIdentifier_);
       }
-      if (((bitField0_ & 0x00000100) == 0x00000100)) {
+      if (((bitField0_ & 0x00000100) != 0)) {
         size += com.google.protobuf.CodedOutputStream
-          .computeMessageSize(9, processorDescriptor_);
+          .computeMessageSize(9, getProcessorDescriptor());
       }
       for (int i = 0; i < inputSpecs_.size(); i++) {
         size += com.google.protobuf.CodedOutputStream
@@ -4254,26 +4551,19 @@ public final class LlapDaemonProtocolProtos {
         size += com.google.protobuf.CodedOutputStream
           .computeMessageSize(12, groupedInputSpecs_.get(i));
       }
-      if (((bitField0_ & 0x00000200) == 0x00000200)) {
+      if (((bitField0_ & 0x00000200) != 0)) {
         size += com.google.protobuf.CodedOutputStream
           .computeInt32Size(13, vertexParallelism_);
       }
-      if (((bitField0_ & 0x00000400) == 0x00000400)) {
+      if (((bitField0_ & 0x00000400) != 0)) {
         size += com.google.protobuf.CodedOutputStream
           .computeBoolSize(14, isExternalSubmission_);
       }
       size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
+      memoizedSize = size;
       return size;
     }
 
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
     @java.lang.Override
     public boolean equals(final java.lang.Object obj) {
       if (obj == this) {
@@ -4284,88 +4574,86 @@ public final class LlapDaemonProtocolProtos {
       }
       org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec) obj;
 
-      boolean result = true;
-      result = result && (hasUser() == other.hasUser());
+      if (hasUser() != other.hasUser()) return false;
       if (hasUser()) {
-        result = result && getUser()
-            .equals(other.getUser());
+        if (!getUser()
+            .equals(other.getUser())) return false;
       }
-      result = result && (hasSignatureKeyId() == other.hasSignatureKeyId());
+      if (hasSignatureKeyId() != other.hasSignatureKeyId()) return false;
       if (hasSignatureKeyId()) {
-        result = result && (getSignatureKeyId()
-            == other.getSignatureKeyId());
+        if (getSignatureKeyId()
+            != other.getSignatureKeyId()) return false;
       }
-      result = result && (hasQueryIdentifier() == other.hasQueryIdentifier());
+      if (hasQueryIdentifier() != other.hasQueryIdentifier()) return false;
       if (hasQueryIdentifier()) {
-        result = result && getQueryIdentifier()
-            .equals(other.getQueryIdentifier());
+        if (!getQueryIdentifier()
+            .equals(other.getQueryIdentifier())) return false;
       }
-      result = result && (hasHiveQueryId() == other.hasHiveQueryId());
+      if (hasHiveQueryId() != other.hasHiveQueryId()) return false;
       if (hasHiveQueryId()) {
-        result = result && getHiveQueryId()
-            .equals(other.getHiveQueryId());
+        if (!getHiveQueryId()
+            .equals(other.getHiveQueryId())) return false;
       }
-      result = result && (hasDagName() == other.hasDagName());
+      if (hasDagName() != other.hasDagName()) return false;
       if (hasDagName()) {
-        result = result && getDagName()
-            .equals(other.getDagName());
+        if (!getDagName()
+            .equals(other.getDagName())) return false;
       }
-      result = result && (hasVertexName() == other.hasVertexName());
+      if (hasVertexName() != other.hasVertexName()) return false;
       if (hasVertexName()) {
-        result = result && getVertexName()
-            .equals(other.getVertexName());
+        if (!getVertexName()
+            .equals(other.getVertexName())) return false;
       }
-      result = result && (hasVertexIndex() == other.hasVertexIndex());
+      if (hasVertexIndex() != other.hasVertexIndex()) return false;
       if (hasVertexIndex()) {
-        result = result && (getVertexIndex()
-            == other.getVertexIndex());
+        if (getVertexIndex()
+            != other.getVertexIndex()) return false;
       }
-      result = result && (hasTokenIdentifier() == other.hasTokenIdentifier());
+      if (hasTokenIdentifier() != other.hasTokenIdentifier()) return false;
       if (hasTokenIdentifier()) {
-        result = result && getTokenIdentifier()
-            .equals(other.getTokenIdentifier());
+        if (!getTokenIdentifier()
+            .equals(other.getTokenIdentifier())) return false;
       }
-      result = result && (hasProcessorDescriptor() == other.hasProcessorDescriptor());
+      if (hasProcessorDescriptor() != other.hasProcessorDescriptor()) return false;
       if (hasProcessorDescriptor()) {
-        result = result && getProcessorDescriptor()
-            .equals(other.getProcessorDescriptor());
-      }
-      result = result && getInputSpecsList()
-          .equals(other.getInputSpecsList());
-      result = result && getOutputSpecsList()
-          .equals(other.getOutputSpecsList());
-      result = result && getGroupedInputSpecsList()
-          .equals(other.getGroupedInputSpecsList());
-      result = result && (hasVertexParallelism() == other.hasVertexParallelism());
+        if (!getProcessorDescriptor()
+            .equals(other.getProcessorDescriptor())) return false;
+      }
+      if (!getInputSpecsList()
+          .equals(other.getInputSpecsList())) return false;
+      if (!getOutputSpecsList()
+          .equals(other.getOutputSpecsList())) return false;
+      if (!getGroupedInputSpecsList()
+          .equals(other.getGroupedInputSpecsList())) return false;
+      if (hasVertexParallelism() != other.hasVertexParallelism()) return false;
       if (hasVertexParallelism()) {
-        result = result && (getVertexParallelism()
-            == other.getVertexParallelism());
+        if (getVertexParallelism()
+            != other.getVertexParallelism()) return false;
       }
-      result = result && (hasIsExternalSubmission() == other.hasIsExternalSubmission());
+      if (hasIsExternalSubmission() != other.hasIsExternalSubmission()) return false;
       if (hasIsExternalSubmission()) {
-        result = result && (getIsExternalSubmission()
-            == other.getIsExternalSubmission());
+        if (getIsExternalSubmission()
+            != other.getIsExternalSubmission()) return false;
       }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
+      if (!getUnknownFields().equals(other.getUnknownFields())) return false;
+      return true;
     }
 
-    private int memoizedHashCode = 0;
     @java.lang.Override
     public int hashCode() {
       if (memoizedHashCode != 0) {
         return memoizedHashCode;
       }
       int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
+      hash = (19 * hash) + getDescriptor().hashCode();
       if (hasUser()) {
         hash = (37 * hash) + USER_FIELD_NUMBER;
         hash = (53 * hash) + getUser().hashCode();
       }
       if (hasSignatureKeyId()) {
         hash = (37 * hash) + SIGNATUREKEYID_FIELD_NUMBER;
-        hash = (53 * hash) + hashLong(getSignatureKeyId());
+        hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+            getSignatureKeyId());
       }
       if (hasQueryIdentifier()) {
         hash = (37 * hash) + QUERY_IDENTIFIER_FIELD_NUMBER;
@@ -4413,13 +4701,25 @@ public final class LlapDaemonProtocolProtos {
       }
       if (hasIsExternalSubmission()) {
         hash = (37 * hash) + IS_EXTERNAL_SUBMISSION_FIELD_NUMBER;
-        hash = (53 * hash) + hashBoolean(getIsExternalSubmission());
+        hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean(
+            getIsExternalSubmission());
       }
       hash = (29 * hash) + getUnknownFields().hashCode();
       memoizedHashCode = hash;
       return hash;
     }
 
+    public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom(
+        java.nio.ByteBuffer data)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      return PARSER.parseFrom(data);
+    }
+    public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom(
+        java.nio.ByteBuffer data,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      return PARSER.parseFrom(data, extensionRegistry);
+    }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom(
         com.google.protobuf.ByteString data)
         throws com.google.protobuf.InvalidProtocolBufferException {
@@ -4443,65 +4743,82 @@ public final class LlapDaemonProtocolProtos {
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return PARSER.parseFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input);
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input, extensionRegistry);
     }
+
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseDelimitedFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseDelimitedWithIOException(PARSER, input);
     }
+
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseDelimitedFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom(
         com.google.protobuf.CodedInputStream input)
         throws java.io.IOException {
-      return PARSER.parseFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input);
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom(
         com.google.protobuf.CodedInputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input, extensionRegistry);
     }
 
-    public static Builder newBuilder() { return Builder.create(); }
+    @java.lang.Override
     public Builder newBuilderForType() { return newBuilder(); }
+    public static Builder newBuilder() {
+      return DEFAULT_INSTANCE.toBuilder();
+    }
     public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec prototype) {
-      return newBuilder().mergeFrom(prototype);
+      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+    }
+    @java.lang.Override
+    public Builder toBuilder() {
+      return this == DEFAULT_INSTANCE
+          ? new Builder() : new Builder().mergeFrom(this);
     }
-    public Builder toBuilder() { return newBuilder(this); }
 
     @java.lang.Override
     protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+        com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
       Builder builder = new Builder(parent);
       return builder;
     }
     /**
-     * Protobuf type {@code SignableVertexSpec}
-     *
      * <pre>
      * The part of SubmitWork that can be signed 
      * </pre>
+     *
+     * Protobuf type {@code SignableVertexSpec}
      */
     public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder {
+        com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+        // @@protoc_insertion_point(builder_implements:SignableVertexSpec)
+        org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder {
       public static final com.google.protobuf.Descriptors.Descriptor
           getDescriptor() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SignableVertexSpec_descriptor;
       }
 
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+      @java.lang.Override
+      protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
           internalGetFieldAccessorTable() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SignableVertexSpec_fieldAccessorTable
             .ensureFieldAccessorsInitialized(
@@ -4514,12 +4831,13 @@ public final class LlapDaemonProtocolProtos {
       }
 
       private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+          com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
         super(parent);
         maybeForceBuilderInitialization();
       }
       private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+        if (com.google.protobuf.GeneratedMessageV3
+                .alwaysUseFieldBuilders) {
           getQueryIdentifierFieldBuilder();
           getProcessorDescriptorFieldBuilder();
           getInputSpecsFieldBuilder();
@@ -4527,76 +4845,65 @@ public final class LlapDaemonProtocolProtos {
           getGroupedInputSpecsFieldBuilder();
         }
       }
-      private static Builder create() {
-        return new Builder();
-      }
-
+      @java.lang.Override
       public Builder clear() {
         super.clear();
+        bitField0_ = 0;
         user_ = "";
-        bitField0_ = (bitField0_ & ~0x00000001);
         signatureKeyId_ = 0L;
-        bitField0_ = (bitField0_ & ~0x00000002);
-        if (queryIdentifierBuilder_ == null) {
-          queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance();
-        } else {
-          queryIdentifierBuilder_.clear();
+        queryIdentifier_ = null;
+        if (queryIdentifierBuilder_ != null) {
+          queryIdentifierBuilder_.dispose();
+          queryIdentifierBuilder_ = null;
         }
-        bitField0_ = (bitField0_ & ~0x00000004);
         hiveQueryId_ = "";
-        bitField0_ = (bitField0_ & ~0x00000008);
         dagName_ = "";
-        bitField0_ = (bitField0_ & ~0x00000010);
         vertexName_ = "";
-        bitField0_ = (bitField0_ & ~0x00000020);
         vertexIndex_ = 0;
-        bitField0_ = (bitField0_ & ~0x00000040);
         tokenIdentifier_ = "";
-        bitField0_ = (bitField0_ & ~0x00000080);
-        if (processorDescriptorBuilder_ == null) {
-          processorDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
-        } else {
-          processorDescriptorBuilder_.clear();
+        processorDescriptor_ = null;
+        if (processorDescriptorBuilder_ != null) {
+          processorDescriptorBuilder_.dispose();
+          processorDescriptorBuilder_ = null;
         }
-        bitField0_ = (bitField0_ & ~0x00000100);
         if (inputSpecsBuilder_ == null) {
           inputSpecs_ = java.util.Collections.emptyList();
-          bitField0_ = (bitField0_ & ~0x00000200);
         } else {
+          inputSpecs_ = null;
           inputSpecsBuilder_.clear();
         }
+        bitField0_ = (bitField0_ & ~0x00000200);
         if (outputSpecsBuilder_ == null) {
           outputSpecs_ = java.util.Collections.emptyList();
-          bitField0_ = (bitField0_ & ~0x00000400);
         } else {
+          outputSpecs_ = null;
           outputSpecsBuilder_.clear();
         }
+        bitField0_ = (bitField0_ & ~0x00000400);
         if (groupedInputSpecsBuilder_ == null) {
           groupedInputSpecs_ = java.util.Collections.emptyList();
-          bitField0_ = (bitField0_ & ~0x00000800);
         } else {
+          groupedInputSpecs_ = null;
           groupedInputSpecsBuilder_.clear();
         }
+        bitField0_ = (bitField0_ & ~0x00000800);
         vertexParallelism_ = 0;
-        bitField0_ = (bitField0_ & ~0x00001000);
         isExternalSubmission_ = false;
-        bitField0_ = (bitField0_ & ~0x00002000);
         return this;
       }
 
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
+      @java.lang.Override
       public com.google.protobuf.Descriptors.Descriptor
           getDescriptorForType() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SignableVertexSpec_descriptor;
       }
 
+      @java.lang.Override
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec getDefaultInstanceForType() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance();
       }
 
+      @java.lang.Override
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec build() {
         org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec result = buildPartial();
         if (!result.isInitialized()) {
@@ -4605,56 +4912,18 @@ public final class LlapDaemonProtocolProtos {
         return result;
       }
 
+      @java.lang.Override
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec buildPartial() {
         org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        result.user_ = user_;
-        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
-          to_bitField0_ |= 0x00000002;
-        }
-        result.signatureKeyId_ = signatureKeyId_;
-        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
-          to_bitField0_ |= 0x00000004;
-        }
-        if (queryIdentifierBuilder_ == null) {
-          result.queryIdentifier_ = queryIdentifier_;
-        } else {
-          result.queryIdentifier_ = queryIdentifierBuilder_.build();
-        }
-        if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
-          to_bitField0_ |= 0x00000008;
-        }
-        result.hiveQueryId_ = hiveQueryId_;
-        if (((from_bitField0_ & 0x00000010) == 0x00000010)) {
-          to_bitField0_ |= 0x00000010;
-        }
-        result.dagName_ = dagName_;
-        if (((from_bitField0_ & 0x00000020) == 0x00000020)) {
-          to_bitField0_ |= 0x00000020;
-        }
-        result.vertexName_ = vertexName_;
-        if (((from_bitField0_ & 0x00000040) == 0x00000040)) {
-          to_bitField0_ |= 0x00000040;
-        }
-        result.vertexIndex_ = vertexIndex_;
-        if (((from_bitField0_ & 0x00000080) == 0x00000080)) {
-          to_bitField0_ |= 0x00000080;
-        }
-        result.tokenIdentifier_ = tokenIdentifier_;
-        if (((from_bitField0_ & 0x00000100) == 0x00000100)) {
-          to_bitField0_ |= 0x00000100;
-        }
-        if (processorDescriptorBuilder_ == null) {
-          result.processorDescriptor_ = processorDescriptor_;
-        } else {
-          result.processorDescriptor_ = processorDescriptorBuilder_.build();
-        }
+        buildPartialRepeatedFields(result);
+        if (bitField0_ != 0) { buildPartial0(result); }
+        onBuilt();
+        return result;
+      }
+
+      private void buildPartialRepeatedFields(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec result) {
         if (inputSpecsBuilder_ == null) {
-          if (((bitField0_ & 0x00000200) == 0x00000200)) {
+          if (((bitField0_ & 0x00000200) != 0)) {
             inputSpecs_ = java.util.Collections.unmodifiableList(inputSpecs_);
             bitField0_ = (bitField0_ & ~0x00000200);
           }
@@ -4663,7 +4932,7 @@ public final class LlapDaemonProtocolProtos {
           result.inputSpecs_ = inputSpecsBuilder_.build();
         }
         if (outputSpecsBuilder_ == null) {
-          if (((bitField0_ & 0x00000400) == 0x00000400)) {
+          if (((bitField0_ & 0x00000400) != 0)) {
             outputSpecs_ = java.util.Collections.unmodifiableList(outputSpecs_);
             bitField0_ = (bitField0_ & ~0x00000400);
           }
@@ -4672,7 +4941,7 @@ public final class LlapDaemonProtocolProtos {
           result.outputSpecs_ = outputSpecsBuilder_.build();
         }
         if (groupedInputSpecsBuilder_ == null) {
-          if (((bitField0_ & 0x00000800) == 0x00000800)) {
+          if (((bitField0_ & 0x00000800) != 0)) {
             groupedInputSpecs_ = java.util.Collections.unmodifiableList(groupedInputSpecs_);
             bitField0_ = (bitField0_ & ~0x00000800);
           }
@@ -4680,19 +4949,95 @@ public final class LlapDaemonProtocolProtos {
         } else {
           result.groupedInputSpecs_ = groupedInputSpecsBuilder_.build();
         }
-        if (((from_bitField0_ & 0x00001000) == 0x00001000)) {
+      }
+
+      private void buildPartial0(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec result) {
+        int from_bitField0_ = bitField0_;
+        int to_bitField0_ = 0;
+        if (((from_bitField0_ & 0x00000001) != 0)) {
+          result.user_ = user_;
+          to_bitField0_ |= 0x00000001;
+        }
+        if (((from_bitField0_ & 0x00000002) != 0)) {
+          result.signatureKeyId_ = signatureKeyId_;
+          to_bitField0_ |= 0x00000002;
+        }
+        if (((from_bitField0_ & 0x00000004) != 0)) {
+          result.queryIdentifier_ = queryIdentifierBuilder_ == null
+              ? queryIdentifier_
+              : queryIdentifierBuilder_.build();
+          to_bitField0_ |= 0x00000004;
+        }
+        if (((from_bitField0_ & 0x00000008) != 0)) {
+          result.hiveQueryId_ = hiveQueryId_;
+          to_bitField0_ |= 0x00000008;
+        }
+        if (((from_bitField0_ & 0x00000010) != 0)) {
+          result.dagName_ = dagName_;
+          to_bitField0_ |= 0x00000010;
+        }
+        if (((from_bitField0_ & 0x00000020) != 0)) {
+          result.vertexName_ = vertexName_;
+          to_bitField0_ |= 0x00000020;
+        }
+        if (((from_bitField0_ & 0x00000040) != 0)) {
+          result.vertexIndex_ = vertexIndex_;
+          to_bitField0_ |= 0x00000040;
+        }
+        if (((from_bitField0_ & 0x00000080) != 0)) {
+          result.tokenIdentifier_ = tokenIdentifier_;
+          to_bitField0_ |= 0x00000080;
+        }
+        if (((from_bitField0_ & 0x00000100) != 0)) {
+          result.processorDescriptor_ = processorDescriptorBuilder_ == null
+              ? processorDescriptor_
+              : processorDescriptorBuilder_.build();
+          to_bitField0_ |= 0x00000100;
+        }
+        if (((from_bitField0_ & 0x00001000) != 0)) {
+          result.vertexParallelism_ = vertexParallelism_;
           to_bitField0_ |= 0x00000200;
         }
-        result.vertexParallelism_ = vertexParallelism_;
-        if (((from_bitField0_ & 0x00002000) == 0x00002000)) {
+        if (((from_bitField0_ & 0x00002000) != 0)) {
+          result.isExternalSubmission_ = isExternalSubmission_;
           to_bitField0_ |= 0x00000400;
         }
-        result.isExternalSubmission_ = isExternalSubmission_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
+        result.bitField0_ |= to_bitField0_;
       }
 
+      @java.lang.Override
+      public Builder clone() {
+        return super.clone();
+      }
+      @java.lang.Override
+      public Builder setField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          java.lang.Object value) {
+        return super.setField(field, value);
+      }
+      @java.lang.Override
+      public Builder clearField(
+          com.google.protobuf.Descriptors.FieldDescriptor field) {
+        return super.clearField(field);
+      }
+      @java.lang.Override
+      public Builder clearOneof(
+          com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+        return super.clearOneof(oneof);
+      }
+      @java.lang.Override
+      public Builder setRepeatedField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          int index, java.lang.Object value) {
+        return super.setRepeatedField(field, index, value);
+      }
+      @java.lang.Override
+      public Builder addRepeatedField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          java.lang.Object value) {
+        return super.addRepeatedField(field, value);
+      }
+      @java.lang.Override
       public Builder mergeFrom(com.google.protobuf.Message other) {
         if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec) {
           return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec)other);
@@ -4705,8 +5050,8 @@ public final class LlapDaemonProtocolProtos {
       public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec other) {
         if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance()) return this;
         if (other.hasUser()) {
-          bitField0_ |= 0x00000001;
           user_ = other.user_;
+          bitField0_ |= 0x00000001;
           onChanged();
         }
         if (other.hasSignatureKeyId()) {
@@ -4716,26 +5061,26 @@ public final class LlapDaemonProtocolProtos {
           mergeQueryIdentifier(other.getQueryIdentifier());
         }
         if (other.hasHiveQueryId()) {
-          bitField0_ |= 0x00000008;
           hiveQueryId_ = other.hiveQueryId_;
+          bitField0_ |= 0x00000008;
           onChanged();
         }
         if (other.hasDagName()) {
-          bitField0_ |= 0x00000010;
           dagName_ = other.dagName_;
+          bitField0_ |= 0x00000010;
           onChanged();
         }
         if (other.hasVertexName()) {
-          bitField0_ |= 0x00000020;
           vertexName_ = other.vertexName_;
+          bitField0_ |= 0x00000020;
           onChanged();
         }
         if (other.hasVertexIndex()) {
           setVertexIndex(other.getVertexIndex());
         }
         if (other.hasTokenIdentifier()) {
-          bitField0_ |= 0x00000080;
           tokenIdentifier_ = other.tokenIdentifier_;
+          bitField0_ |= 0x00000080;
           onChanged();
         }
         if (other.hasProcessorDescriptor()) {
@@ -4760,7 +5105,7 @@ public final class LlapDaemonProtocolProtos {
               inputSpecs_ = other.inputSpecs_;
               bitField0_ = (bitField0_ & ~0x00000200);
               inputSpecsBuilder_ = 
-                com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ?
+                com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
                    getInputSpecsFieldBuilder() : null;
             } else {
               inputSpecsBuilder_.addAllMessages(other.inputSpecs_);
@@ -4786,7 +5131,7 @@ public final class LlapDaemonProtocolProtos {
               outputSpecs_ = other.outputSpecs_;
               bitField0_ = (bitField0_ & ~0x00000400);
               outputSpecsBuilder_ = 
-                com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ?
+                com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
                    getOutputSpecsFieldBuilder() : null;
             } else {
               outputSpecsBuilder_.addAllMessages(other.outputSpecs_);
@@ -4812,7 +5157,7 @@ public final class LlapDaemonProtocolProtos {
               groupedInputSpecs_ = other.groupedInputSpecs_;
               bitField0_ = (bitField0_ & ~0x00000800);
               groupedInputSpecsBuilder_ = 
-                com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ?
+                com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
                    getGroupedInputSpecsFieldBuilder() : null;
             } else {
               groupedInputSpecsBuilder_.addAllMessages(other.groupedInputSpecs_);
@@ -4826,49 +5171,167 @@ public final class LlapDaemonProtocolProtos {
           setIsExternalSubmission(other.getIsExternalSubmission());
         }
         this.mergeUnknownFields(other.getUnknownFields());
+        onChanged();
         return this;
       }
 
+      @java.lang.Override
       public final boolean isInitialized() {
         return true;
       }
 
+      @java.lang.Override
       public Builder mergeFrom(
           com.google.protobuf.CodedInputStream input,
           com.google.protobuf.ExtensionRegistryLite extensionRegistry)
           throws java.io.IOException {
-        org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parsedMessage = null;
+        if (extensionRegistry == null) {
+          throw new java.lang.NullPointerException();
+        }
         try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+          boolean done = false;
+          while (!done) {
+            int tag = input.readTag();
+            switch (tag) {
+              case 0:
+                done = true;
+                break;
+              case 10: {
+                user_ = input.readBytes();
+                bitField0_ |= 0x00000001;
+                break;
+              } // case 10
+              case 16: {
+                signatureKeyId_ = input.readInt64();
+                bitField0_ |= 0x00000002;
+                break;
+              } // case 16
+              case 26: {
+                input.readMessage(
+                    getQueryIdentifierFieldBuilder().getBuilder(),
+                    extensionRegistry);
+                bitField0_ |= 0x00000004;
+                break;
+              } // case 26
+              case 34: {
+                hiveQueryId_ = input.readBytes();
+                bitField0_ |= 0x00000008;
+                break;
+              } // case 34
+              case 42: {
+                dagName_ = input.readBytes();
+                bitField0_ |= 0x00000010;
+                break;
+              } // case 42
+              case 50: {
+                vertexName_ = input.readBytes();
+                bitField0_ |= 0x00000020;
+                break;
+              } // case 50
+              case 56: {
+                vertexIndex_ = input.readInt32();
+                bitField0_ |= 0x00000040;
+                break;
+              } // case 56
+              case 66: {
+                tokenIdentifier_ = input.readBytes();
+                bitField0_ |= 0x00000080;
+                break;
+              } // case 66
+              case 74: {
+                input.readMessage(
+                    getProcessorDescriptorFieldBuilder().getBuilder(),
+                    extensionRegistry);
+                bitField0_ |= 0x00000100;
+                break;
+              } // case 74
+              case 82: {
+                org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto m =
+                    input.readMessage(
+                        org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.PARSER,
+                        extensionRegistry);
+                if (inputSpecsBuilder_ == null) {
+                  ensureInputSpecsIsMutable();
+                  inputSpecs_.add(m);
+                } else {
+                  inputSpecsBuilder_.addMessage(m);
+                }
+                break;
+              } // case 82
+              case 90: {
+                org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto m =
+                    input.readMessage(
+                        org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.PARSER,
+                        extensionRegistry);
+                if (outputSpecsBuilder_ == null) {
+                  ensureOutputSpecsIsMutable();
+                  outputSpecs_.add(m);
+                } else {
+                  outputSpecsBuilder_.addMessage(m);
+                }
+                break;
+              } // case 90
+              case 98: {
+                org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto m =
+                    input.readMessage(
+                        org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.PARSER,
+                        extensionRegistry);
+                if (groupedInputSpecsBuilder_ == null) {
+                  ensureGroupedInputSpecsIsMutable();
+                  groupedInputSpecs_.add(m);
+                } else {
+                  groupedInputSpecsBuilder_.addMessage(m);
+                }
+                break;
+              } // case 98
+              case 104: {
+                vertexParallelism_ = input.readInt32();
+                bitField0_ |= 0x00001000;
+                break;
+              } // case 104
+              case 112: {
+                isExternalSubmission_ = input.readBool();
+                bitField0_ |= 0x00002000;
+                break;
+              } // case 112
+              default: {
+                if (!super.parseUnknownField(input, extensionRegistry, tag)) {
+                  done = true; // was an endgroup tag
+                }
+                break;
+              } // default:
+            } // switch (tag)
+          } // while (!done)
         } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec) e.getUnfinishedMessage();
-          throw e;
+          throw e.unwrapIOException();
         } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
+          onChanged();
+        } // finally
         return this;
       }
       private int bitField0_;
 
-      // optional string user = 1;
       private java.lang.Object user_ = "";
       /**
        * <code>optional string user = 1;</code>
+       * @return Whether the user field is set.
        */
       public boolean hasUser() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
+        return ((bitField0_ & 0x00000001) != 0);
       }
       /**
        * <code>optional string user = 1;</code>
+       * @return The user.
        */
       public java.lang.String getUser() {
         java.lang.Object ref = user_;
         if (!(ref instanceof java.lang.String)) {
-          java.lang.String s = ((com.google.protobuf.ByteString) ref)
-              .toStringUtf8();
-          user_ = s;
+          com.google.protobuf.ByteString bs =
+              (com.google.protobuf.ByteString) ref;
+          java.lang.String s = bs.toStringUtf8();
+          if (bs.isValidUtf8()) {
+            user_ = s;
+          }
           return s;
         } else {
           return (java.lang.String) ref;
@@ -4876,6 +5339,7 @@ public final class LlapDaemonProtocolProtos {
       }
       /**
        * <code>optional string user = 1;</code>
+       * @return The bytes for user.
        */
       public com.google.protobuf.ByteString
           getUserBytes() {
@@ -4892,65 +5356,73 @@ public final class LlapDaemonProtocolProtos {
       }
       /**
        * <code>optional string user = 1;</code>
+       * @param value The user to set.
+       * @return This builder for chaining.
        */
       public Builder setUser(
           java.lang.String value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000001;
+        if (value == null) { throw new NullPointerException(); }
         user_ = value;
+        bitField0_ |= 0x00000001;
         onChanged();
         return this;
       }
       /**
        * <code>optional string user = 1;</code>
+       * @return This builder for chaining.
        */
       public Builder clearUser() {
-        bitField0_ = (bitField0_ & ~0x00000001);
         user_ = getDefaultInstance().getUser();
+        bitField0_ = (bitField0_ & ~0x00000001);
         onChanged();
         return this;
       }
       /**
        * <code>optional string user = 1;</code>
+       * @param value The bytes for user to set.
+       * @return This builder for chaining.
        */
       public Builder setUserBytes(
           com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000001;
+        if (value == null) { throw new NullPointerException(); }
         user_ = value;
+        bitField0_ |= 0x00000001;
         onChanged();
         return this;
       }
 
-      // optional int64 signatureKeyId = 2;
       private long signatureKeyId_ ;
       /**
        * <code>optional int64 signatureKeyId = 2;</code>
+       * @return Whether the signatureKeyId field is set.
        */
+      @java.lang.Override
       public boolean hasSignatureKeyId() {
-        return ((bitField0_ & 0x00000002) == 0x00000002);
+        return ((bitField0_ & 0x00000002) != 0);
       }
       /**
        * <code>optional int64 signatureKeyId = 2;</code>
+       * @return The signatureKeyId.
        */
+      @java.lang.Override
       public long getSignatureKeyId() {
         return signatureKeyId_;
       }
       /**
        * <code>optional int64 signatureKeyId = 2;</code>
+       * @param value The signatureKeyId to set.
+       * @return This builder for chaining.
        */
       public Builder setSignatureKeyId(long value) {
-        bitField0_ |= 0x00000002;
+
         signatureKeyId_ = value;
+        bitField0_ |= 0x00000002;
         onChanged();
         return this;
       }
       /**
        * <code>optional int64 signatureKeyId = 2;</code>
+       * @return This builder for chaining.
        */
       public Builder clearSignatureKeyId() {
         bitField0_ = (bitField0_ & ~0x00000002);
@@ -4959,22 +5431,23 @@ public final class LlapDaemonProtocolProtos {
         return this;
       }
 
-      // optional .QueryIdentifierProto query_identifier = 3;
-      private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance();
-      private com.google.protobuf.SingleFieldBuilder<
+      private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_;
+      private com.google.protobuf.SingleFieldBuilderV3<
           org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> queryIdentifierBuilder_;
       /**
        * <code>optional .QueryIdentifierProto query_identifier = 3;</code>
+       * @return Whether the queryIdentifier field is set.
        */
       public boolean hasQueryIdentifier() {
-        return ((bitField0_ & 0x00000004) == 0x00000004);
+        return ((bitField0_ & 0x00000004) != 0);
       }
       /**
        * <code>optional .QueryIdentifierProto query_identifier = 3;</code>
+       * @return The queryIdentifier.
        */
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() {
         if (queryIdentifierBuilder_ == null) {
-          return queryIdentifier_;
+          return queryIdentifier_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_;
         } else {
           return queryIdentifierBuilder_.getMessage();
         }
@@ -4988,11 +5461,11 @@ public final class LlapDaemonProtocolProtos {
             throw new NullPointerException();
           }
           queryIdentifier_ = value;
-          onChanged();
         } else {
           queryIdentifierBuilder_.setMessage(value);
         }
         bitField0_ |= 0x00000004;
+        onChanged();
         return this;
       }
       /**
@@ -5002,11 +5475,11 @@ public final class LlapDaemonProtocolProtos {
           org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder builderForValue) {
         if (queryIdentifierBuilder_ == null) {
           queryIdentifier_ = builderForValue.build();
-          onChanged();
         } else {
           queryIdentifierBuilder_.setMessage(builderForValue.build());
         }
         bitField0_ |= 0x00000004;
+        onChanged();
         return this;
       }
       /**
@@ -5014,31 +5487,33 @@ public final class LlapDaemonProtocolProtos {
        */
       public Builder mergeQueryIdentifier(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto value) {
         if (queryIdentifierBuilder_ == null) {
-          if (((bitField0_ & 0x00000004) == 0x00000004) &&
-              queryIdentifier_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance()) {
-            queryIdentifier_ =
-              org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.newBuilder(queryIdentifier_).mergeFrom(value).buildPartial();
+          if (((bitField0_ & 0x00000004) != 0) &&
+            queryIdentifier_ != null &&
+            queryIdentifier_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance()) {
+            getQueryIdentifierBuilder().mergeFrom(value);
           } else {
             queryIdentifier_ = value;
           }
-          onChanged();
         } else {
           queryIdentifierBuilder_.mergeFrom(value);
         }
-        bitField0_ |= 0x00000004;
+        if (queryIdentifier_ != null) {
+          bitField0_ |= 0x00000004;
+          onChanged();
+        }
         return this;
       }
       /**
        * <code>optional .QueryIdentifierProto query_identifier = 3;</code>
        */
       public Builder clearQueryIdentifier() {
-        if (queryIdentifierBuilder_ == null) {
-          queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance();
-          onChanged();
-        } else {
-          queryIdentifierBuilder_.clear();
-        }
         bitField0_ = (bitField0_ & ~0x00000004);
+        queryIdentifier_ = null;
+        if (queryIdentifierBuilder_ != null) {
+          queryIdentifierBuilder_.dispose();
+          queryIdentifierBuilder_ = null;
+        }
+        onChanged();
         return this;
       }
       /**
@@ -5056,19 +5531,20 @@ public final class LlapDaemonProtocolProtos {
         if (queryIdentifierBuilder_ != null) {
           return queryIdentifierBuilder_.getMessageOrBuilder();
         } else {
-          return queryIdentifier_;
+          return queryIdentifier_ == null ?
+              org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_;
         }
       }
       /**
        * <code>optional .QueryIdentifierProto query_identifier = 3;</code>
        */
-      private com.google.protobuf.SingleFieldBuilder<
+      private com.google.protobuf.SingleFieldBuilderV3<
           org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> 
           getQueryIdentifierFieldBuilder() {
         if (queryIdentifierBuilder_ == null) {
-          queryIdentifierBuilder_ = new com.google.protobuf.SingleFieldBuilder<
+          queryIdentifierBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
               org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder>(
-                  queryIdentifier_,
+                  getQueryIdentifier(),
                   getParentForChildren(),
                   isClean());
           queryIdentifier_ = null;
@@ -5076,23 +5552,27 @@ public final class LlapDaemonProtocolProtos {
         return queryIdentifierBuilder_;
       }
 
-      // optional string hive_query_id = 4;
       private java.lang.Object hiveQueryId_ = "";
       /**
        * <code>optional string hive_query_id = 4;</code>
+       * @return Whether the hiveQueryId field is set.
        */
       public boolean hasHiveQueryId() {
-        return ((bitField0_ & 0x00000008) == 0x00000008);
+        return ((bitField0_ & 0x00000008) != 0);
       }
       /**
        * <code>optional string hive_query_id = 4;</code>
+       * @return The hiveQueryId.
        */
       public java.lang.String getHiveQueryId() {
         java.lang.Object ref = hiveQueryId_;
         if (!(ref instanceof java.lang.String)) {
-          java.lang.String s = ((com.google.protobuf.ByteString) ref)
-              .toStringUtf8();
-          hiveQueryId_ = s;
+          com.google.protobuf.ByteString bs =
+              (com.google.protobuf.ByteString) ref;
+          java.lang.String s = bs.toStringUtf8();
+          if (bs.isValidUtf8()) {
+            hiveQueryId_ = s;
+          }
           return s;
         } else {
           return (java.lang.String) ref;
@@ -5100,6 +5580,7 @@ public final class LlapDaemonProtocolProtos {
       }
       /**
        * <code>optional string hive_query_id = 4;</code>
+       * @return The bytes for hiveQueryId.
        */
       public com.google.protobuf.ByteString
           getHiveQueryIdBytes() {
@@ -5116,76 +5597,82 @@ public final class LlapDaemonProtocolProtos {
       }
       /**
        * <code>optional string hive_query_id = 4;</code>
+       * @param value The hiveQueryId to set.
+       * @return This builder for chaining.
        */
       public Builder setHiveQueryId(
           java.lang.String value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000008;
+        if (value == null) { throw new NullPointerException(); }
         hiveQueryId_ = value;
+        bitField0_ |= 0x00000008;
         onChanged();
         return this;
       }
       /**
        * <code>optional string hive_query_id = 4;</code>
+       * @return This builder for chaining.
        */
       public Builder clearHiveQueryId() {
-        bitField0_ = (bitField0_ & ~0x00000008);
         hiveQueryId_ = getDefaultInstance().getHiveQueryId();
+        bitField0_ = (bitField0_ & ~0x00000008);
         onChanged();
         return this;
       }
       /**
        * <code>optional string hive_query_id = 4;</code>
+       * @param value The bytes for hiveQueryId to set.
+       * @return This builder for chaining.
        */
       public Builder setHiveQueryIdBytes(
           com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000008;
+        if (value == null) { throw new NullPointerException(); }
         hiveQueryId_ = value;
+        bitField0_ |= 0x00000008;
         onChanged();
         return this;
       }
 
-      // optional string dag_name = 5;
       private java.lang.Object dagName_ = "";
       /**
-       * <code>optional string dag_name = 5;</code>
-       *
        * <pre>
        * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
        * </pre>
+       *
+       * <code>optional string dag_name = 5;</code>
+       * @return Whether the dagName field is set.
        */
       public boolean hasDagName() {
-        return ((bitField0_ & 0x00000010) == 0x00000010);
+        return ((bitField0_ & 0x00000010) != 0);
       }
       /**
-       * <code>optional string dag_name = 5;</code>
-       *
        * <pre>
        * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
        * </pre>
+       *
+       * <code>optional string dag_name = 5;</code>
+       * @return The dagName.
        */
       public java.lang.String getDagName() {
         java.lang.Object ref = dagName_;
         if (!(ref instanceof java.lang.String)) {
-          java.lang.String s = ((com.google.protobuf.ByteString) ref)
-              .toStringUtf8();
-          dagName_ = s;
+          com.google.protobuf.ByteString bs =
+              (com.google.protobuf.ByteString) ref;
+          java.lang.String s = bs.toStringUtf8();
+          if (bs.isValidUtf8()) {
+            dagName_ = s;
+          }
           return s;
         } else {
           return (java.lang.String) ref;
         }
       }
       /**
-       * <code>optional string dag_name = 5;</code>
-       *
        * <pre>
        * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
        * </pre>
+       *
+       * <code>optional string dag_name = 5;</code>
+       * @return The bytes for dagName.
        */
       public com.google.protobuf.ByteString
           getDagNameBytes() {
@@ -5201,70 +5688,75 @@ public final class LlapDaemonProtocolProtos {
         }
       }
       /**
-       * <code>optional string dag_name = 5;</code>
-       *
        * <pre>
        * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
        * </pre>
+       *
+       * <code>optional string dag_name = 5;</code>
+       * @param value The dagName to set.
+       * @return This builder for chaining.
        */
       public Builder setDagName(
           java.lang.String value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000010;
+        if (value == null) { throw new NullPointerException(); }
         dagName_ = value;
+        bitField0_ |= 0x00000010;
         onChanged();
         return this;
       }
       /**
-       * <code>optional string dag_name = 5;</code>
-       *
        * <pre>
        * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
        * </pre>
+       *
+       * <code>optional string dag_name = 5;</code>
+       * @return This builder for chaining.
        */
       public Builder clearDagName() {
-        bitField0_ = (bitField0_ & ~0x00000010);
         dagName_ = getDefaultInstance().getDagName();
+        bitField0_ = (bitField0_ & ~0x00000010);
         onChanged();
         return this;
       }
       /**
-       * <code>optional string dag_name = 5;</code>
-       *
        * <pre>
        * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
        * </pre>
+       *
+       * <code>optional string dag_name = 5;</code>
+       * @param value The bytes for dagName to set.
+       * @return This builder for chaining.
        */
       public Builder setDagNameBytes(
           com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000010;
+        if (value == null) { throw new NullPointerException(); }
         dagName_ = value;
+        bitField0_ |= 0x00000010;
         onChanged();
         return this;
       }
 
-      // optional string vertex_name = 6;
       private java.lang.Object vertexName_ = "";
       /**
        * <code>optional string vertex_name = 6;</code>
+       * @return Whether the vertexName field is set.
        */
       public boolean hasVertexName() {
-        return ((bitField0_ & 0x00000020) == 0x00000020);
+        return ((bitField0_ & 0x00000020) != 0);
       }
       /**
        * <code>optional string vertex_name = 6;</code>
+       * @return The vertexName.
        */
       public java.lang.String getVertexName() {
         java.lang.Object ref = vertexName_;
         if (!(ref instanceof java.lang.String)) {
-          java.lang.String s = ((com.google.protobuf.ByteString) ref)
-              .toStringUtf8();
-          vertexName_ = s;
+          com.google.protobuf.ByteString bs =
+              (com.google.protobuf.ByteString) ref;
+          java.lang.String s = bs.toStringUtf8();
+          if (bs.isValidUtf8()) {
+            vertexName_ = s;
+          }
           return s;
         } else {
           return (java.lang.String) ref;
@@ -5272,6 +5764,7 @@ public final class LlapDaemonProtocolProtos {
       }
       /**
        * <code>optional string vertex_name = 6;</code>
+       * @return The bytes for vertexName.
        */
       public com.google.protobuf.ByteString
           getVertexNameBytes() {
@@ -5288,65 +5781,73 @@ public final class LlapDaemonProtocolProtos {
       }
       /**
        * <code>optional string vertex_name = 6;</code>
+       * @param value The vertexName to set.
+       * @return This builder for chaining.
        */
       public Builder setVertexName(
           java.lang.String value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000020;
+        if (value == null) { throw new NullPointerException(); }
         vertexName_ = value;
+        bitField0_ |= 0x00000020;
         onChanged();
         return this;
       }
       /**
        * <code>optional string vertex_name = 6;</code>
+       * @return This builder for chaining.
        */
       public Builder clearVertexName() {
-        bitField0_ = (bitField0_ & ~0x00000020);
         vertexName_ = getDefaultInstance().getVertexName();
+        bitField0_ = (bitField0_ & ~0x00000020);
         onChanged();
         return this;
       }
       /**
        * <code>optional string vertex_name = 6;</code>
+       * @param value The bytes for vertexName to set.
+       * @return This builder for chaining.
        */
       public Builder setVertexNameBytes(
           com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000020;
+        if (value == null) { throw new NullPointerException(); }
         vertexName_ = value;
+        bitField0_ |= 0x00000020;
         onChanged();
         return this;
       }
 
-      // optional int32 vertex_index = 7;
       private int vertexIndex_ ;
       /**
        * <code>optional int32 vertex_index = 7;</code>
+       * @return Whether the vertexIndex field is set.
        */
+      @java.lang.Override
       public boolean hasVertexIndex() {
-        return ((bitField0_ & 0x00000040) == 0x00000040);
+        return ((bitField0_ & 0x00000040) != 0);
       }
       /**
        * <code>optional int32 vertex_index = 7;</code>
+       * @return The vertexIndex.
        */
+      @java.lang.Override
       public int getVertexIndex() {
         return vertexIndex_;
       }
       /**
        * <code>optional int32 vertex_index = 7;</code>
+       * @param value The vertexIndex to set.
+       * @return This builder for chaining.
        */
       public Builder setVertexIndex(int value) {
-        bitField0_ |= 0x00000040;
+
         vertexIndex_ = value;
+        bitField0_ |= 0x00000040;
         onChanged();
         return this;
       }
       /**
        * <code>optional int32 vertex_index = 7;</code>
+       * @return This builder for chaining.
        */
       public Builder clearVertexIndex() {
         bitField0_ = (bitField0_ & ~0x00000040);
@@ -5355,42 +5856,47 @@ public final class LlapDaemonProtocolProtos {
         return this;
       }
 
-      // optional string token_identifier = 8;
       private java.lang.Object tokenIdentifier_ = "";
       /**
-       * <code>optional string token_identifier = 8;</code>
-       *
        * <pre>
        * The core vertex stuff 
        * </pre>
+       *
+       * <code>optional string token_identifier = 8;</code>
+       * @return Whether the tokenIdentifier field is set.
        */
       public boolean hasTokenIdentifier() {
-        return ((bitField0_ & 0x00000080) == 0x00000080);
+        return ((bitField0_ & 0x00000080) != 0);
       }
       /**
-       * <code>optional string token_identifier = 8;</code>
-       *
        * <pre>
        * The core vertex stuff 
        * </pre>
+       *
+       * <code>optional string token_identifier = 8;</code>
+       * @return The tokenIdentifier.
        */
       public java.lang.String getTokenIdentifier() {
         java.lang.Object ref = tokenIdentifier_;
         if (!(ref instanceof java.lang.String)) {
-          java.lang.String s = ((com.google.protobuf.ByteString) ref)
-              .toStringUtf8();
-          tokenIdentifier_ = s;
+          com.google.protobuf.ByteString bs =
+              (com.google.protobuf.ByteString) ref;
+          java.lang.String s = bs.toStringUtf8();
+          if (bs.isValidUtf8()) {
+            tokenIdentifier_ = s;
+          }
           return s;
         } else {
           return (java.lang.String) ref;
         }
       }
       /**
-       * <code>optional string token_identifier = 8;</code>
-       *
        * <pre>
        * The core vertex stuff 
        * </pre>
+       *
+       * <code>optional string token_identifier = 8;</code>
+       * @return The bytes for tokenIdentifier.
        */
       public com.google.protobuf.ByteString
           getTokenIdentifierBytes() {
@@ -5406,69 +5912,71 @@ public final class LlapDaemonProtocolProtos {
         }
       }
       /**
-       * <code>optional string token_identifier = 8;</code>
-       *
        * <pre>
        * The core vertex stuff 
        * </pre>
+       *
+       * <code>optional string token_identifier = 8;</code>
+       * @param value The tokenIdentifier to set.
+       * @return This builder for chaining.
        */
       public Builder setTokenIdentifier(
           java.lang.String value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000080;
+        if (value == null) { throw new NullPointerException(); }
         tokenIdentifier_ = value;
+        bitField0_ |= 0x00000080;
         onChanged();
         return this;
       }
       /**
-       * <code>optional string token_identifier = 8;</code>
-       *
        * <pre>
        * The core vertex stuff 
        * </pre>
+       *
+       * <code>optional string token_identifier = 8;</code>
+       * @return This builder for chaining.
        */
       public Builder clearTokenIdentifier() {
-        bitField0_ = (bitField0_ & ~0x00000080);
         tokenIdentifier_ = getDefaultInstance().getTokenIdentifier();
+        bitField0_ = (bitField0_ & ~0x00000080);
         onChanged();
         return this;
       }
       /**
-       * <code>optional string token_identifier = 8;</code>
-       *
        * <pre>
        * The core vertex stuff 
        * </pre>
+       *
+       * <code>optional string token_identifier = 8;</code>
+       * @param value The bytes for tokenIdentifier to set.
+       * @return This builder for chaining.
        */
       public Builder setTokenIdentifierBytes(
           com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000080;
+        if (value == null) { throw new NullPointerException(); }
         tokenIdentifier_ = value;
+        bitField0_ |= 0x00000080;
         onChanged();
         return this;
       }
 
-      // optional .EntityDescriptorProto processor_descriptor = 9;
-      private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto processorDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
-      private com.google.protobuf.SingleFieldBuilder<
+      private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto processorDescriptor_;
+      private com.google.protobuf.SingleFieldBuilderV3<
           org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> processorDescriptorBuilder_;
       /**
        * <code>optional .EntityDescriptorProto processor_descriptor = 9;</code>
+       * @return Whether the processorDescriptor field is set.
        */
       public boolean hasProcessorDescriptor() {
-        return ((bitField0_ & 0x00000100) == 0x00000100);
+        return ((bitField0_ & 0x00000100) != 0);
       }
       /**
        * <code>optional .EntityDescriptorProto processor_descriptor = 9;</code>
+       * @return The processorDescriptor.
        */
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getProcessorDescriptor() {
         if (processorDescriptorBuilder_ == null) {
-          return processorDescriptor_;
+          return processorDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : processorDescriptor_;
         } else {
           return processorDescriptorBuilder_.getMessage();
         }
@@ -5482,11 +5990,11 @@ public final class LlapDaemonProtocolProtos {
             throw new NullPointerException();
           }
           processorDescriptor_ = value;
-          onChanged();
         } else {
           processorDescriptorBuilder_.setMessage(value);
         }
         bitField0_ |= 0x00000100;
+        onChanged();
         return this;
       }
       /**
@@ -5496,11 +6004,11 @@ public final class LlapDaemonProtocolProtos {
           org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder builderForValue) {
         if (processorDescriptorBuilder_ == null) {
           processorDescriptor_ = builderForValue.build();
-          onChanged();
         } else {
           processorDescriptorBuilder_.setMessage(builderForValue.build());
         }
         bitField0_ |= 0x00000100;
+        onChanged();
         return this;
       }
       /**
@@ -5508,31 +6016,33 @@ public final class LlapDaemonProtocolProtos {
        */
       public Builder mergeProcessorDescriptor(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto value) {
         if (processorDescriptorBuilder_ == null) {
-          if (((bitField0_ & 0x00000100) == 0x00000100) &&
-              processorDescriptor_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) {
-            processorDescriptor_ =
-              org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.newBuilder(processorDescriptor_).mergeFrom(value).buildPartial();
+          if (((bitField0_ & 0x00000100) != 0) &&
+            processorDescriptor_ != null &&
+            processorDescriptor_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) {
+            getProcessorDescriptorBuilder().mergeFrom(value);
           } else {
             processorDescriptor_ = value;
           }
-          onChanged();
         } else {
           processorDescriptorBuilder_.mergeFrom(value);
         }
-        bitField0_ |= 0x00000100;
+        if (processorDescriptor_ != null) {
+          bitField0_ |= 0x00000100;
+          onChanged();
+        }
         return this;
       }
       /**
        * <code>optional .EntityDescriptorProto processor_descriptor = 9;</code>
        */
       public Builder clearProcessorDescriptor() {
-        if (processorDescriptorBuilder_ == null) {
-          processorDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance();
-          onChanged();
-        } else {
-          processorDescriptorBuilder_.clear();
-        }
         bitField0_ = (bitField0_ & ~0x00000100);
+        processorDescriptor_ = null;
+        if (processorDescriptorBuilder_ != null) {
+          processorDescriptorBuilder_.dispose();
+          processorDescriptorBuilder_ = null;
+        }
+        onChanged();
         return this;
       }
       /**
@@ -5550,19 +6060,20 @@ public final class LlapDaemonProtocolProtos {
         if (processorDescriptorBuilder_ != null) {
           return processorDescriptorBuilder_.getMessageOrBuilder();
         } else {
-          return processorDescriptor_;
+          return processorDescriptor_ == null ?
+              org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : processorDescriptor_;
         }
       }
       /**
        * <code>optional .EntityDescriptorProto processor_descriptor = 9;</code>
        */
-      private com.google.protobuf.SingleFieldBuilder<
+      private com.google.protobuf.SingleFieldBuilderV3<
           org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> 
           getProcessorDescriptorFieldBuilder() {
         if (processorDescriptorBuilder_ == null) {
-          processorDescriptorBuilder_ = new com.google.protobuf.SingleFieldBuilder<
+          processorDescriptorBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
               org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder>(
-                  processorDescriptor_,
+                  getProcessorDescriptor(),
                   getParentForChildren(),
                   isClean());
           processorDescriptor_ = null;
@@ -5570,17 +6081,16 @@ public final class LlapDaemonProtocolProtos {
         return processorDescriptorBuilder_;
       }
 
-      // repeated .IOSpecProto input_specs = 10;
       private java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> inputSpecs_ =
         java.util.Collections.emptyList();
       private void ensureInputSpecsIsMutable() {
-        if (!((bitField0_ & 0x00000200) == 0x00000200)) {
+        if (!((bitField0_ & 0x00000200) != 0)) {
           inputSpecs_ = new java.util.ArrayList<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto>(inputSpecs_);
           bitField0_ |= 0x00000200;
          }
       }
 
-      private com.google.protobuf.RepeatedFieldBuilder<
+      private com.google.protobuf.RepeatedFieldBuilderV3<
           org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder> inputSpecsBuilder_;
 
       /**
@@ -5712,7 +6222,8 @@ public final class LlapDaemonProtocolProtos {
           java.lang.Iterable<? extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> values) {
         if (inputSpecsBuilder_ == null) {
           ensureInputSpecsIsMutable();
-          super.addAll(values, inputSpecs_);
+          com.google.protobuf.AbstractMessageLite.Builder.addAll(
+              values, inputSpecs_);
           onChanged();
         } else {
           inputSpecsBuilder_.addAllMessages(values);
@@ -5795,14 +6306,14 @@ public final class LlapDaemonProtocolProtos {
            getInputSpecsBuilderList() {
         return getInputSpecsFieldBuilder().getBuilderList();
       }
-      private com.google.protobuf.RepeatedFieldBuilder<
+      private com.google.protobuf.RepeatedFieldBuilderV3<
           org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder> 
           getInputSpecsFieldBuilder() {
         if (inputSpecsBuilder_ == null) {
-          inputSpecsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
+          inputSpecsBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3<
               org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder>(
                   inputSpecs_,
-                  ((bitField0_ & 0x00000200) == 0x00000200),
+                  ((bitField0_ & 0x00000200) != 0),
                   getParentForChildren(),
                   isClean());
           inputSpecs_ = null;
@@ -5810,17 +6321,16 @@ public final class LlapDaemonProtocolProtos {
         return inputSpecsBuilder_;
       }
 
-      // repeated .IOSpecProto output_specs = 11;
       private java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> outputSpecs_ =
         java.util.Collections.emptyList();
       private void ensureOutputSpecsIsMutable() {
-        if (!((bitField0_ & 0x00000400) == 0x00000400)) {
+        if (!((bitField0_ & 0x00000400) != 0)) {
           outputSpecs_ = new java.util.ArrayList<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto>(outputSpecs_);
           bitField0_ |= 0x00000400;
          }
       }
 
-      private com.google.protobuf.RepeatedFieldBuilder<
+      private com.google.protobuf.RepeatedFieldBuilderV3<
           org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder> outputSpecsBuilder_;
 
       /**
@@ -5952,7 +6462,8 @@ public final class LlapDaemonProtocolProtos {
           java.lang.Iterable<? extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> values) {
         if (outputSpecsBuilder_ == null) {
           ensureOutputSpecsIsMutable();
-          super.addAll(values, outputSpecs_);
+          com.google.protobuf.AbstractMessageLite.Builder.addAll(
+              values, outputSpecs_);
           onChanged();
         } else {
           outputSpecsBuilder_.addAllMessages(values);
@@ -6035,14 +6546,14 @@ public final class LlapDaemonProtocolProtos {
            getOutputSpecsBuilderList() {
         return getOutputSpecsFieldBuilder().getBuilderList();
       }
-      private com.google.protobuf.RepeatedFieldBuilder<
+      private com.google.protobuf.RepeatedFieldBuilderV3<
           org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder> 
           getOutputSpecsFieldBuilder() {
         if (outputSpecsBuilder_ == null) {
-          outputSpecsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
+          outputSpecsBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3<
               org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder>(
                   outputSpecs_,
-                  ((bitField0_ & 0x00000400) == 0x00000400),
+                  ((bitField0_ & 0x00000400) != 0),
                   getParentForChildren(),
                   isClean());
           outputSpecs_ = null;
@@ -6050,17 +6561,16 @@ public final class LlapDaemonProtocolProtos {
         return outputSpecsBuilder_;
       }
 
-      // repeated .GroupInputSpecProto grouped_input_specs = 12;
       private java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto> groupedInputSpecs_ =
         java.util.Collections.emptyList();
       private void ensureGroupedInputSpecsIsMutable() {
-        if (!((bitField0_ & 0x00000800) == 0x00000800)) {
+        if (!((bitField0_ & 0x00000800) != 0)) {
           groupedInputSpecs_ = new java.util.ArrayList<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto>(groupedInputSpecs_);
           bitField0_ |= 0x00000800;
          }
       }
 
-      private com.google.protobuf.RepeatedFieldBuilder<
+      private com.google.protobuf.RepeatedFieldBuilderV3<
           org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder> groupedInputSpecsBuilder_;
 
       /**
@@ -6192,7 +6702,8 @@ public final class LlapDaemonProtocolProtos {
           java.lang.Iterable<? extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto> values) {
         if (groupedInputSpecsBuilder_ == null) {
           ensureGroupedInputSpecsIsMutable();
-          super.addAll(values, groupedInputSpecs_);
+          com.google.protobuf.AbstractMessageLite.Builder.addAll(
+              values, groupedInputSpecs_);
           onChanged();
         } else {
           groupedInputSpecsBuilder_.addAllMessages(values);
@@ -6275,14 +6786,14 @@ public final class LlapDaemonProtocolProtos {
            getGroupedInputSpecsBuilderList() {
         return getGroupedInputSpecsFieldBuilder().getBuilderList();
       }
-      private com.google.protobuf.RepeatedFieldBuilder<
+      private com.google.protobuf.RepeatedFieldBuilderV3<
           org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder> 
           getGroupedInputSpecsFieldBuilder() {
         if (groupedInputSpecsBuilder_ == null) {
-          groupedInputSpecsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
+          groupedInputSpecsBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3<
               org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder>(
                   groupedInputSpecs_,
-                  ((bitField0_ & 0x00000800) == 0x00000800),
+                  ((bitField0_ & 0x00000800) != 0),
                   getParentForChildren(),
                   isClean());
           groupedInputSpecs_ = null;
@@ -6290,47 +6801,54 @@ public final class LlapDaemonProtocolProtos {
         return groupedInputSpecsBuilder_;
       }
 
-      // optional int32 vertex_parallelism = 13;
       private int vertexParallelism_ ;
       /**
-       * <code>optional int32 vertex_parallelism = 13;</code>
-       *
        * <pre>
        * An internal field required for Tez.
        * </pre>
+       *
+       * <code>optional int32 vertex_parallelism = 13;</code>
+       * @return Whether the vertexParallelism field is set.
        */
+      @java.lang.Override
       public boolean hasVertexParallelism() {
-        return ((bitField0_ & 0x00001000) == 0x00001000);
+        return ((bitField0_ & 0x00001000) != 0);
       }
       /**
-       * <code>optional int32 vertex_parallelism = 13;</code>
-       *
        * <pre>
        * An internal field required for Tez.
        * </pre>
+       *
+       * <code>optional int32 vertex_parallelism = 13;</code>
+       * @return The vertexParallelism.
        */
+      @java.lang.Override
       public int getVertexParallelism() {
         return vertexParallelism_;
       }
       /**
-       * <code>optional int32 vertex_parallelism = 13;</code>
-       *
        * <pre>
        * An internal field required for Tez.
        * </pre>
+       *
+       * <code>optional int32 vertex_parallelism = 13;</code>
+       * @param value The vertexParallelism to set.
+       * @return This builder for chaining.
        */
       public Builder setVertexParallelism(int value) {
-        bitField0_ |= 0x00001000;
+
         vertexParallelism_ = value;
+        bitField0_ |= 0x00001000;
         onChanged();
         return this;
       }
       /**
-       * <code>optional int32 vertex_parallelism = 13;</code>
-       *
        * <pre>
        * An internal field required for Tez.
        * </pre>
+       *
+       * <code>optional int32 vertex_parallelism = 13;</code>
+       * @return This builder for chaining.
        */
       public Builder clearVertexParallelism() {
         bitField0_ = (bitField0_ & ~0x00001000);
@@ -6339,31 +6857,38 @@ public final class LlapDaemonProtocolProtos {
         return this;
       }
 
-      // optional bool is_external_submission = 14 [default = false];
       private boolean isExternalSubmission_ ;
       /**
        * <code>optional bool is_external_submission = 14 [default = false];</code>
+       * @return Whether the isExternalSubmission field is set.
        */
+      @java.lang.Override
       public boolean hasIsExternalSubmission() {
-        return ((bitField0_ & 0x00002000) == 0x00002000);
+        return ((bitField0_ & 0x00002000) != 0);
       }
       /**
        * <code>optional bool is_external_submission = 14 [default = false];</code>
+       * @return The isExternalSubmission.
        */
+      @java.lang.Override
       public boolean getIsExternalSubmission() {
         return isExternalSubmission_;
       }
       /**
        * <code>optional bool is_external_submission = 14 [default = false];</code>
+       * @param value The isExternalSubmission to set.
+       * @return This builder for chaining.
        */
       public Builder setIsExternalSubmission(boolean value) {
-        bitField0_ |= 0x00002000;
+
         isExternalSubmission_ = value;
+        bitField0_ |= 0x00002000;
         onChanged();
         return this;
       }
       /**
        * <code>optional bool is_external_submission = 14 [default = false];</code>
+       * @return This builder for chaining.
        */
       public Builder clearIsExternalSubmission() {
         bitField0_ = (bitField0_ & ~0x00002000);
@@ -6371,28 +6896,82 @@ public final class LlapDaemonProtocolProtos {
         onChanged();
         return this;
       }
+      @java.lang.Override
+      public final Builder setUnknownFields(
+          final com.google.protobuf.UnknownFieldSet unknownFields) {
+        return super.setUnknownFields(unknownFields);
+      }
+
+      @java.lang.Override
+      public final Builder mergeUnknownFields(
+          final com.google.protobuf.UnknownFieldSet unknownFields) {
+        return super.mergeUnknownFields(unknownFields);
+      }
+
 
       // @@protoc_insertion_point(builder_scope:SignableVertexSpec)
     }
 
+    // @@protoc_insertion_point(class_scope:SignableVertexSpec)
+    private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec DEFAULT_INSTANCE;
     static {
-      defaultInstance = new SignableVertexSpec(true);
-      defaultInstance.initFields();
+      DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec();
+    }
+
+    public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec getDefaultInstance() {
+      return DEFAULT_INSTANCE;
+    }
+
+    @java.lang.Deprecated public static final com.google.protobuf.Parser<SignableVertexSpec>
+        PARSER = new com.google.protobuf.AbstractParser<SignableVertexSpec>() {
+      @java.lang.Override
+      public SignableVertexSpec parsePartialFrom(
+          com.google.protobuf.CodedInputStream input,
+          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+          throws com.google.protobuf.InvalidProtocolBufferException {
+        Builder builder = newBuilder();
+        try {
+          builder.mergeFrom(input, extensionRegistry);
+        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+          throw e.setUnfinishedMessage(builder.buildPartial());
+        } catch (com.google.protobuf.UninitializedMessageException e) {
+          throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
+        } catch (java.io.IOException e) {
+          throw new com.google.protobuf.InvalidProtocolBufferException(e)
+              .setUnfinishedMessage(builder.buildPartial());
+        }
+        return builder.buildPartial();
+      }
+    };
+
+    public static com.google.protobuf.Parser<SignableVertexSpec> parser() {
+      return PARSER;
+    }
+
+    @java.lang.Override
+    public com.google.protobuf.Parser<SignableVertexSpec> getParserForType() {
+      return PARSER;
+    }
+
+    @java.lang.Override
+    public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec getDefaultInstanceForType() {
+      return DEFAULT_INSTANCE;
     }
 
-    // @@protoc_insertion_point(class_scope:SignableVertexSpec)
   }
 
-  public interface VertexOrBinaryOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
+  public interface VertexOrBinaryOrBuilder extends
+      // @@protoc_insertion_point(interface_extends:VertexOrBinary)
+      com.google.protobuf.MessageOrBuilder {
 
-    // optional .SignableVertexSpec vertex = 1;
     /**
      * <code>optional .SignableVertexSpec vertex = 1;</code>
+     * @return Whether the vertex field is set.
      */
     boolean hasVertex();
     /**
      * <code>optional .SignableVertexSpec vertex = 1;</code>
+     * @return The vertex.
      */
     org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec getVertex();
     /**
@@ -6400,234 +6979,161 @@ public final class LlapDaemonProtocolProtos {
      */
     org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder getVertexOrBuilder();
 
-    // optional bytes vertexBinary = 2;
     /**
-     * <code>optional bytes vertexBinary = 2;</code>
-     *
      * <pre>
      * SignableVertexSpec
      * </pre>
+     *
+     * <code>optional bytes vertexBinary = 2;</code>
+     * @return Whether the vertexBinary field is set.
      */
     boolean hasVertexBinary();
     /**
-     * <code>optional bytes vertexBinary = 2;</code>
-     *
      * <pre>
      * SignableVertexSpec
      * </pre>
+     *
+     * <code>optional bytes vertexBinary = 2;</code>
+     * @return The vertexBinary.
      */
     com.google.protobuf.ByteString getVertexBinary();
   }
   /**
-   * Protobuf type {@code VertexOrBinary}
-   *
    * <pre>
    * Union
    * </pre>
+   *
+   * Protobuf type {@code VertexOrBinary}
    */
   public static final class VertexOrBinary extends
-      com.google.protobuf.GeneratedMessage
-      implements VertexOrBinaryOrBuilder {
+      com.google.protobuf.GeneratedMessageV3 implements
+      // @@protoc_insertion_point(message_implements:VertexOrBinary)
+      VertexOrBinaryOrBuilder {
+  private static final long serialVersionUID = 0L;
     // Use VertexOrBinary.newBuilder() to construct.
-    private VertexOrBinary(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+    private VertexOrBinary(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
       super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private VertexOrBinary(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final VertexOrBinary defaultInstance;
-    public static VertexOrBinary getDefaultInstance() {
-      return defaultInstance;
     }
-
-    public VertexOrBinary getDefaultInstanceForType() {
-      return defaultInstance;
+    private VertexOrBinary() {
+      vertexBinary_ = com.google.protobuf.ByteString.EMPTY;
     }
 
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
     @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private VertexOrBinary(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder subBuilder = null;
-              if (((bitField0_ & 0x00000001) == 0x00000001)) {
-                subBuilder = vertex_.toBuilder();
-              }
-              vertex_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.PARSER, extensionRegistry);
-              if (subBuilder != null) {
-                subBuilder.mergeFrom(vertex_);
-                vertex_ = subBuilder.buildPartial();
-              }
-              bitField0_ |= 0x00000001;
-              break;
-            }
-            case 18: {
-              bitField0_ |= 0x00000002;
-              vertexBinary_ = input.readBytes();
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
+    @SuppressWarnings({"unused"})
+    protected java.lang.Object newInstance(
+        UnusedPrivateParameter unused) {
+      return new VertexOrBinary();
     }
+
     public static final com.google.protobuf.Descriptors.Descriptor
         getDescriptor() {
       return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexOrBinary_descriptor;
     }
 
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+    @java.lang.Override
+    protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
         internalGetFieldAccessorTable() {
       return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexOrBinary_fieldAccessorTable
           .ensureFieldAccessorsInitialized(
               org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.Builder.class);
     }
 
-    public static com.google.protobuf.Parser<VertexOrBinary> PARSER =
-        new com.google.protobuf.AbstractParser<VertexOrBinary>() {
-      public VertexOrBinary parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new VertexOrBinary(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<VertexOrBinary> getParserForType() {
-      return PARSER;
-    }
-
     private int bitField0_;
-    // optional .SignableVertexSpec vertex = 1;
     public static final int VERTEX_FIELD_NUMBER = 1;
     private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec vertex_;
     /**
      * <code>optional .SignableVertexSpec vertex = 1;</code>
+     * @return Whether the vertex field is set.
      */
+    @java.lang.Override
     public boolean hasVertex() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
+      return ((bitField0_ & 0x00000001) != 0);
     }
     /**
      * <code>optional .SignableVertexSpec vertex = 1;</code>
+     * @return The vertex.
      */
+    @java.lang.Override
     public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec getVertex() {
-      return vertex_;
+      return vertex_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance() : vertex_;
     }
     /**
      * <code>optional .SignableVertexSpec vertex = 1;</code>
      */
+    @java.lang.Override
     public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder getVertexOrBuilder() {
-      return vertex_;
+      return vertex_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance() : vertex_;
     }
 
-    // optional bytes vertexBinary = 2;
     public static final int VERTEXBINARY_FIELD_NUMBER = 2;
-    private com.google.protobuf.ByteString vertexBinary_;
+    private com.google.protobuf.ByteString vertexBinary_ = com.google.protobuf.ByteString.EMPTY;
     /**
-     * <code>optional bytes vertexBinary = 2;</code>
-     *
      * <pre>
      * SignableVertexSpec
      * </pre>
+     *
+     * <code>optional bytes vertexBinary = 2;</code>
+     * @return Whether the vertexBinary field is set.
      */
+    @java.lang.Override
     public boolean hasVertexBinary() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
+      return ((bitField0_ & 0x00000002) != 0);
     }
     /**
-     * <code>optional bytes vertexBinary = 2;</code>
-     *
      * <pre>
      * SignableVertexSpec
      * </pre>
+     *
+     * <code>optional bytes vertexBinary = 2;</code>
+     * @return The vertexBinary.
      */
+    @java.lang.Override
     public com.google.protobuf.ByteString getVertexBinary() {
       return vertexBinary_;
     }
 
-    private void initFields() {
-      vertex_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance();
-      vertexBinary_ = com.google.protobuf.ByteString.EMPTY;
-    }
     private byte memoizedIsInitialized = -1;
+    @java.lang.Override
     public final boolean isInitialized() {
       byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
+      if (isInitialized == 1) return true;
+      if (isInitialized == 0) return false;
 
       memoizedIsInitialized = 1;
       return true;
     }
 
+    @java.lang.Override
     public void writeTo(com.google.protobuf.CodedOutputStream output)
                         throws java.io.IOException {
-      getSerializedSize();
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeMessage(1, vertex_);
+      if (((bitField0_ & 0x00000001) != 0)) {
+        output.writeMessage(1, getVertex());
       }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
+      if (((bitField0_ & 0x00000002) != 0)) {
         output.writeBytes(2, vertexBinary_);
       }
       getUnknownFields().writeTo(output);
     }
 
-    private int memoizedSerializedSize = -1;
+    @java.lang.Override
     public int getSerializedSize() {
-      int size = memoizedSerializedSize;
+      int size = memoizedSize;
       if (size != -1) return size;
 
       size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
+      if (((bitField0_ & 0x00000001) != 0)) {
         size += com.google.protobuf.CodedOutputStream
-          .computeMessageSize(1, vertex_);
+          .computeMessageSize(1, getVertex());
       }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
+      if (((bitField0_ & 0x00000002) != 0)) {
         size += com.google.protobuf.CodedOutputStream
           .computeBytesSize(2, vertexBinary_);
       }
       size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
+      memoizedSize = size;
       return size;
     }
 
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
     @java.lang.Override
     public boolean equals(final java.lang.Object obj) {
       if (obj == this) {
@@ -6638,30 +7144,27 @@ public final class LlapDaemonProtocolProtos {
       }
       org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary) obj;
 
-      boolean result = true;
-      result = result && (hasVertex() == other.hasVertex());
+      if (hasVertex() != other.hasVertex()) return false;
       if (hasVertex()) {
-        result = result && getVertex()
-            .equals(other.getVertex());
+        if (!getVertex()
+            .equals(other.getVertex())) return false;
       }
-      result = result && (hasVertexBinary() == other.hasVertexBinary());
+      if (hasVertexBinary() != other.hasVertexBinary()) return false;
       if (hasVertexBinary()) {
-        result = result && getVertexBinary()
-            .equals(other.getVertexBinary());
+        if (!getVertexBinary()
+            .equals(other.getVertexBinary())) return false;
       }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
+      if (!getUnknownFields().equals(other.getUnknownFields())) return false;
+      return true;
     }
 
-    private int memoizedHashCode = 0;
     @java.lang.Override
     public int hashCode() {
       if (memoizedHashCode != 0) {
         return memoizedHashCode;
       }
       int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
+      hash = (19 * hash) + getDescriptor().hashCode();
       if (hasVertex()) {
         hash = (37 * hash) + VERTEX_FIELD_NUMBER;
         hash = (53 * hash) + getVertex().hashCode();
@@ -6675,6 +7178,17 @@ public final class LlapDaemonProtocolProtos {
       return hash;
     }
 
+    public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom(
+        java.nio.ByteBuffer data)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      return PARSER.parseFrom(data);
+    }
+    public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom(
+        java.nio.ByteBuffer data,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      return PARSER.parseFrom(data, extensionRegistry);
+    }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom(
         com.google.protobuf.ByteString data)
         throws com.google.protobuf.InvalidProtocolBufferException {
@@ -6698,65 +7212,82 @@ public final class LlapDaemonProtocolProtos {
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return PARSER.parseFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input);
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input, extensionRegistry);
     }
+
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseDelimitedFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseDelimitedWithIOException(PARSER, input);
     }
+
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseDelimitedFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom(
         com.google.protobuf.CodedInputStream input)
         throws java.io.IOException {
-      return PARSER.parseFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input);
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom(
         com.google.protobuf.CodedInputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input, extensionRegistry);
     }
 
-    public static Builder newBuilder() { return Builder.create(); }
+    @java.lang.Override
     public Builder newBuilderForType() { return newBuilder(); }
+    public static Builder newBuilder() {
+      return DEFAULT_INSTANCE.toBuilder();
+    }
     public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary prototype) {
-      return newBuilder().mergeFrom(prototype);
+      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+    }
+    @java.lang.Override
+    public Builder toBuilder() {
+      return this == DEFAULT_INSTANCE
+          ? new Builder() : new Builder().mergeFrom(this);
     }
-    public Builder toBuilder() { return newBuilder(this); }
 
     @java.lang.Override
     protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+        com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
       Builder builder = new Builder(parent);
       return builder;
     }
     /**
-     * Protobuf type {@code VertexOrBinary}
-     *
      * <pre>
      * Union
      * </pre>
+     *
+     * Protobuf type {@code VertexOrBinary}
      */
     public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinaryOrBuilder {
+        com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+        // @@protoc_insertion_point(builder_implements:VertexOrBinary)
+        org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinaryOrBuilder {
       public static final com.google.protobuf.Descriptors.Descriptor
           getDescriptor() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexOrBinary_descriptor;
       }
 
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+      @java.lang.Override
+      protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
           internalGetFieldAccessorTable() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexOrBinary_fieldAccessorTable
             .ensureFieldAccessorsInitialized(
@@ -6769,45 +7300,41 @@ public final class LlapDaemonProtocolProtos {
       }
 
       private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+          com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
         super(parent);
         maybeForceBuilderInitialization();
       }
       private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+        if (com.google.protobuf.GeneratedMessageV3
+                .alwaysUseFieldBuilders) {
           getVertexFieldBuilder();
         }
       }
-      private static Builder create() {
-        return new Builder();
-      }
-
+      @java.lang.Override
       public Builder clear() {
         super.clear();
-        if (vertexBuilder_ == null) {
-          vertex_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance();
-        } else {
-          vertexBuilder_.clear();
+        bitField0_ = 0;
+        vertex_ = null;
+        if (vertexBuilder_ != null) {
+          vertexBuilder_.dispose();
+          vertexBuilder_ = null;
         }
-        bitField0_ = (bitField0_ & ~0x00000001);
         vertexBinary_ = com.google.protobuf.ByteString.EMPTY;
-        bitField0_ = (bitField0_ & ~0x00000002);
         return this;
       }
 
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
+      @java.lang.Override
       public com.google.protobuf.Descriptors.Descriptor
           getDescriptorForType() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexOrBinary_descriptor;
       }
 
+      @java.lang.Override
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary getDefaultInstanceForType() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.getDefaultInstance();
       }
 
+      @java.lang.Override
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary build() {
         org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary result = buildPartial();
         if (!result.isInitialized()) {
@@ -6816,27 +7343,63 @@ public final class LlapDaemonProtocolProtos {
         return result;
       }
 
+      @java.lang.Override
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary buildPartial() {
         org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary(this);
+        if (bitField0_ != 0) { buildPartial0(result); }
+        onBuilt();
+        return result;
+      }
+
+      private void buildPartial0(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary result) {
         int from_bitField0_ = bitField0_;
         int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
+        if (((from_bitField0_ & 0x00000001) != 0)) {
+          result.vertex_ = vertexBuilder_ == null
+              ? vertex_
+              : vertexBuilder_.build();
           to_bitField0_ |= 0x00000001;
         }
-        if (vertexBuilder_ == null) {
-          result.vertex_ = vertex_;
-        } else {
-          result.vertex_ = vertexBuilder_.build();
-        }
-        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
+        if (((from_bitField0_ & 0x00000002) != 0)) {
+          result.vertexBinary_ = vertexBinary_;
           to_bitField0_ |= 0x00000002;
         }
-        result.vertexBinary_ = vertexBinary_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
+        result.bitField0_ |= to_bitField0_;
       }
 
+      @java.lang.Override
+      public Builder clone() {
+        return super.clone();
+      }
+      @java.lang.Override
+      public Builder setField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          java.lang.Object value) {
+        return super.setField(field, value);
+      }
+      @java.lang.Override
+      public Builder clearField(
+          com.google.protobuf.Descriptors.FieldDescriptor field) {
+        return super.clearField(field);
+      }
+      @java.lang.Override
+      public Builder clearOneof(
+          com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+        return super.clearOneof(oneof);
+      }
+      @java.lang.Override
+      public Builder setRepeatedField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          int index, java.lang.Object value) {
+        return super.setRepeatedField(field, index, value);
+      }
+      @java.lang.Override
+      public Builder addRepeatedField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          java.lang.Object value) {
+        return super.addRepeatedField(field, value);
+      }
+      @java.lang.Override
       public Builder mergeFrom(com.google.protobuf.Message other) {
         if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary) {
           return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary)other);
@@ -6855,48 +7418,77 @@ public final class LlapDaemonProtocolProtos {
           setVertexBinary(other.getVertexBinary());
         }
         this.mergeUnknownFields(other.getUnknownFields());
+        onChanged();
         return this;
       }
 
+      @java.lang.Override
       public final boolean isInitialized() {
         return true;
       }
 
+      @java.lang.Override
       public Builder mergeFrom(
           com.google.protobuf.CodedInputStream input,
           com.google.protobuf.ExtensionRegistryLite extensionRegistry)
           throws java.io.IOException {
-        org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parsedMessage = null;
+        if (extensionRegistry == null) {
+          throw new java.lang.NullPointerException();
+        }
         try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+          boolean done = false;
+          while (!done) {
+            int tag = input.readTag();
+            switch (tag) {
+              case 0:
+                done = true;
+                break;
+              case 10: {
+                input.readMessage(
+                    getVertexFieldBuilder().getBuilder(),
+                    extensionRegistry);
+                bitField0_ |= 0x00000001;
+                break;
+              } // case 10
+              case 18: {
+                vertexBinary_ = input.readBytes();
+                bitField0_ |= 0x00000002;
+                break;
+              } // case 18
+              default: {
+                if (!super.parseUnknownField(input, extensionRegistry, tag)) {
+                  done = true; // was an endgroup tag
+                }
+                break;
+              } // default:
+            } // switch (tag)
+          } // while (!done)
         } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary) e.getUnfinishedMessage();
-          throw e;
+          throw e.unwrapIOException();
         } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
+          onChanged();
+        } // finally
         return this;
       }
       private int bitField0_;
 
-      // optional .SignableVertexSpec vertex = 1;
-      private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec vertex_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance();
-      private com.google.protobuf.SingleFieldBuilder<
+      private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec vertex_;
+      private com.google.protobuf.SingleFieldBuilderV3<
           org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder> vertexBuilder_;
       /**
        * <code>optional .SignableVertexSpec vertex = 1;</code>
+       * @return Whether the vertex field is set.
        */
       public boolean hasVertex() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
+        return ((bitField0_ & 0x00000001) != 0);
       }
       /**
        * <code>optional .SignableVertexSpec vertex = 1;</code>
+       * @return The vertex.
        */
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec getVertex() {
         if (vertexBuilder_ == null) {
-          return vertex_;
+          return vertex_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance() : vertex_;
         } else {
           return vertexBuilder_.getMessage();
         }
@@ -6910,11 +7502,11 @@ public final class LlapDaemonProtocolProtos {
             throw new NullPointerException();
           }
           vertex_ = value;
-          onChanged();
         } else {
           vertexBuilder_.setMessage(value);
         }
         bitField0_ |= 0x00000001;
+        onChanged();
         return this;
       }
       /**
@@ -6924,11 +7516,11 @@ public final class LlapDaemonProtocolProtos {
           org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder builderForValue) {
         if (vertexBuilder_ == null) {
           vertex_ = builderForValue.build();
-          onChanged();
         } else {
           vertexBuilder_.setMessage(builderForValue.build());
         }
         bitField0_ |= 0x00000001;
+        onChanged();
         return this;
       }
       /**
@@ -6936,31 +7528,33 @@ public final class LlapDaemonProtocolProtos {
        */
       public Builder mergeVertex(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec value) {
         if (vertexBuilder_ == null) {
-          if (((bitField0_ & 0x00000001) == 0x00000001) &&
-              vertex_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance()) {
-            vertex_ =
-              org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.newBuilder(vertex_).mergeFrom(value).buildPartial();
+          if (((bitField0_ & 0x00000001) != 0) &&
+            vertex_ != null &&
+            vertex_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance()) {
+            getVertexBuilder().mergeFrom(value);
           } else {
             vertex_ = value;
           }
-          onChanged();
         } else {
           vertexBuilder_.mergeFrom(value);
         }
-        bitField0_ |= 0x00000001;
+        if (vertex_ != null) {
+          bitField0_ |= 0x00000001;
+          onChanged();
+        }
         return this;
       }
       /**
        * <code>optional .SignableVertexSpec vertex = 1;</code>
        */
       public Builder clearVertex() {
-        if (vertexBuilder_ == null) {
-          vertex_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance();
-          onChanged();
-        } else {
-          vertexBuilder_.clear();
-        }
         bitField0_ = (bitField0_ & ~0x00000001);
+        vertex_ = null;
+        if (vertexBuilder_ != null) {
+          vertexBuilder_.dispose();
+          vertexBuilder_ = null;
+        }
+        onChanged();
         return this;
       }
       /**
@@ -6978,19 +7572,20 @@ public final class LlapDaemonProtocolProtos {
         if (vertexBuilder_ != null) {
           return vertexBuilder_.getMessageOrBuilder();
         } else {
-          return vertex_;
+          return vertex_ == null ?
+              org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance() : vertex_;
         }
       }
       /**
        * <code>optional .SignableVertexSpec vertex = 1;</code>
        */
-      private com.google.protobuf.SingleFieldBuilder<
+      private com.google.protobuf.SingleFieldBuilderV3<
           org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder> 
           getVertexFieldBuilder() {
         if (vertexBuilder_ == null) {
-          vertexBuilder_ = new com.google.protobuf.SingleFieldBuilder<
+          vertexBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
               org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder>(
-                  vertex_,
+                  getVertex(),
                   getParentForChildren(),
                   isClean());
           vertex_ = null;
@@ -6998,50 +7593,54 @@ public final class LlapDaemonProtocolProtos {
         return vertexBuilder_;
       }
 
-      // optional bytes vertexBinary = 2;
       private com.google.protobuf.ByteString vertexBinary_ = com.google.protobuf.ByteString.EMPTY;
       /**
-       * <code>optional bytes vertexBinary = 2;</code>
-       *
        * <pre>
        * SignableVertexSpec
        * </pre>
+       *
+       * <code>optional bytes vertexBinary = 2;</code>
+       * @return Whether the vertexBinary field is set.
        */
+      @java.lang.Override
       public boolean hasVertexBinary() {
-        return ((bitField0_ & 0x00000002) == 0x00000002);
+        return ((bitField0_ & 0x00000002) != 0);
       }
       /**
-       * <code>optional bytes vertexBinary = 2;</code>
-       *
        * <pre>
        * SignableVertexSpec
        * </pre>
+       *
+       * <code>optional bytes vertexBinary = 2;</code>
+       * @return The vertexBinary.
        */
+      @java.lang.Override
       public com.google.protobuf.ByteString getVertexBinary() {
         return vertexBinary_;
       }
       /**
-       * <code>optional bytes vertexBinary = 2;</code>
-       *
        * <pre>
        * SignableVertexSpec
        * </pre>
+       *
+       * <code>optional bytes vertexBinary = 2;</code>
+       * @param value The vertexBinary to set.
+       * @return This builder for chaining.
        */
       public Builder setVertexBinary(com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000002;
+        if (value == null) { throw new NullPointerException(); }
         vertexBinary_ = value;
+        bitField0_ |= 0x00000002;
         onChanged();
         return this;
       }
       /**
-       * <code>optional bytes vertexBinary = 2;</code>
-       *
        * <pre>
        * SignableVertexSpec
        * </pre>
+       *
+       * <code>optional bytes vertexBinary = 2;</code>
+       * @return This builder for chaining.
        */
       public Builder clearVertexBinary() {
         bitField0_ = (bitField0_ & ~0x00000002);
@@ -7049,78 +7648,137 @@ public final class LlapDaemonProtocolProtos {
         onChanged();
         return this;
       }
+      @java.lang.Override
+      public final Builder setUnknownFields(
+          final com.google.protobuf.UnknownFieldSet unknownFields) {
+        return super.setUnknownFields(unknownFields);
+      }
+
+      @java.lang.Override
+      public final Builder mergeUnknownFields(
+          final com.google.protobuf.UnknownFieldSet unknownFields) {
+        return super.mergeUnknownFields(unknownFields);
+      }
+
 
       // @@protoc_insertion_point(builder_scope:VertexOrBinary)
     }
 
+    // @@protoc_insertion_point(class_scope:VertexOrBinary)
+    private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary DEFAULT_INSTANCE;
     static {
-      defaultInstance = new VertexOrBinary(true);
-      defaultInstance.initFields();
+      DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary();
     }
 
-    // @@protoc_insertion_point(class_scope:VertexOrBinary)
-  }
+    public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary getDefaultInstance() {
+      return DEFAULT_INSTANCE;
+    }
 
-  public interface FragmentRuntimeInfoOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
+    @java.lang.Deprecated public static final com.google.protobuf.Parser<VertexOrBinary>
+        PARSER = new com.google.protobuf.AbstractParser<VertexOrBinary>() {
+      @java.lang.Override
+      public VertexOrBinary parsePartialFrom(
+          com.google.protobuf.CodedInputStream input,
+          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+          throws com.google.protobuf.InvalidProtocolBufferException {
+        Builder builder = newBuilder();
+        try {
+          builder.mergeFrom(input, extensionRegistry);
+        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+          throw e.setUnfinishedMessage(builder.buildPartial());
+        } catch (com.google.protobuf.UninitializedMessageException e) {
+          throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
+        } catch (java.io.IOException e) {
+          throw new com.google.protobuf.InvalidProtocolBufferException(e)
+              .setUnfinishedMessage(builder.buildPartial());
+        }
+        return builder.buildPartial();
+      }
+    };
 
-    // optional int32 num_self_and_upstream_tasks = 1;
-    /**
-     * <code>optional int32 num_self_and_upstream_tasks = 1;</code>
-     */
-    boolean hasNumSelfAndUpstreamTasks();
-    /**
+    public static com.google.protobuf.Parser<VertexOrBinary> parser() {
+      return PARSER;
+    }
+
+    @java.lang.Override
+    public com.google.protobuf.Parser<VertexOrBinary> getParserForType() {
+      return PARSER;
+    }
+
+    @java.lang.Override
+    public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary getDefaultInstanceForType() {
+      return DEFAULT_INSTANCE;
+    }
+
+  }
+
+  public interface FragmentRuntimeInfoOrBuilder extends
+      // @@protoc_insertion_point(interface_extends:FragmentRuntimeInfo)
+      com.google.protobuf.MessageOrBuilder {
+
+    /**
+     * <code>optional int32 num_self_and_upstream_tasks = 1;</code>
+     * @return Whether the numSelfAndUpstreamTasks field is set.
+     */
+    boolean hasNumSelfAndUpstreamTasks();
+    /**
      * <code>optional int32 num_self_and_upstream_tasks = 1;</code>
+     * @return The numSelfAndUpstreamTasks.
      */
     int getNumSelfAndUpstreamTasks();
 
-    // optional int32 num_self_and_upstream_completed_tasks = 2;
     /**
      * <code>optional int32 num_self_and_upstream_completed_tasks = 2;</code>
+     * @return Whether the numSelfAndUpstreamCompletedTasks field is set.
      */
     boolean hasNumSelfAndUpstreamCompletedTasks();
     /**
      * <code>optional int32 num_self_and_upstream_completed_tasks = 2;</code>
+     * @return The numSelfAndUpstreamCompletedTasks.
      */
     int getNumSelfAndUpstreamCompletedTasks();
 
-    // optional int32 within_dag_priority = 3;
     /**
      * <code>optional int32 within_dag_priority = 3;</code>
+     * @return Whether the withinDagPriority field is set.
      */
     boolean hasWithinDagPriority();
     /**
      * <code>optional int32 within_dag_priority = 3;</code>
+     * @return The withinDagPriority.
      */
     int getWithinDagPriority();
 
-    // optional int64 dag_start_time = 4;
     /**
      * <code>optional int64 dag_start_time = 4;</code>
+     * @return Whether the dagStartTime field is set.
      */
     boolean hasDagStartTime();
     /**
      * <code>optional int64 dag_start_time = 4;</code>
+     * @return The dagStartTime.
      */
     long getDagStartTime();
 
-    // optional int64 first_attempt_start_time = 5;
     /**
      * <code>optional int64 first_attempt_start_time = 5;</code>
+     * @return Whether the firstAttemptStartTime field is set.
      */
     boolean hasFirstAttemptStartTime();
     /**
      * <code>optional int64 first_attempt_start_time = 5;</code>
+     * @return The firstAttemptStartTime.
      */
     long getFirstAttemptStartTime();
 
-    // optional int64 current_attempt_start_time = 6;
     /**
      * <code>optional int64 current_attempt_start_time = 6;</code>
+     * @return Whether the currentAttemptStartTime field is set.
      */
     boolean hasCurrentAttemptStartTime();
     /**
      * <code>optional int64 current_attempt_start_time = 6;</code>
+     * @return The currentAttemptStartTime.
      */
     long getCurrentAttemptStartTime();
   }
@@ -7128,302 +7786,222 @@ public final class LlapDaemonProtocolProtos {
    * Protobuf type {@code FragmentRuntimeInfo}
    */
   public static final class FragmentRuntimeInfo extends
-      com.google.protobuf.GeneratedMessage
-      implements FragmentRuntimeInfoOrBuilder {
+      com.google.protobuf.GeneratedMessageV3 implements
+      // @@protoc_insertion_point(message_implements:FragmentRuntimeInfo)
+      FragmentRuntimeInfoOrBuilder {
+  private static final long serialVersionUID = 0L;
     // Use FragmentRuntimeInfo.newBuilder() to construct.
-    private FragmentRuntimeInfo(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+    private FragmentRuntimeInfo(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
       super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private FragmentRuntimeInfo(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final FragmentRuntimeInfo defaultInstance;
-    public static FragmentRuntimeInfo getDefaultInstance() {
-      return defaultInstance;
     }
-
-    public FragmentRuntimeInfo getDefaultInstanceForType() {
-      return defaultInstance;
+    private FragmentRuntimeInfo() {
     }
 
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
     @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private FragmentRuntimeInfo(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 8: {
-              bitField0_ |= 0x00000001;
-              numSelfAndUpstreamTasks_ = input.readInt32();
-              break;
-            }
-            case 16: {
-              bitField0_ |= 0x00000002;
-              numSelfAndUpstreamCompletedTasks_ = input.readInt32();
-              break;
-            }
-            case 24: {
-              bitField0_ |= 0x00000004;
-              withinDagPriority_ = input.readInt32();
-              break;
-            }
-            case 32: {
-              bitField0_ |= 0x00000008;
-              dagStartTime_ = input.readInt64();
-              break;
-            }
-            case 40: {
-              bitField0_ |= 0x00000010;
-              firstAttemptStartTime_ = input.readInt64();
-              break;
-            }
-            case 48: {
-              bitField0_ |= 0x00000020;
-              currentAttemptStartTime_ = input.readInt64();
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
+    @SuppressWarnings({"unused"})
+    protected java.lang.Object newInstance(
+        UnusedPrivateParameter unused) {
+      return new FragmentRuntimeInfo();
     }
+
     public static final com.google.protobuf.Descriptors.Descriptor
         getDescriptor() {
       return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentRuntimeInfo_descriptor;
     }
 
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+    @java.lang.Override
+    protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
         internalGetFieldAccessorTable() {
       return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentRuntimeInfo_fieldAccessorTable
           .ensureFieldAccessorsInitialized(
               org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.Builder.class);
     }
 
-    public static com.google.protobuf.Parser<FragmentRuntimeInfo> PARSER =
-        new com.google.protobuf.AbstractParser<FragmentRuntimeInfo>() {
-      public FragmentRuntimeInfo parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new FragmentRuntimeInfo(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<FragmentRuntimeInfo> getParserForType() {
-      return PARSER;
-    }
-
     private int bitField0_;
-    // optional int32 num_self_and_upstream_tasks = 1;
     public static final int NUM_SELF_AND_UPSTREAM_TASKS_FIELD_NUMBER = 1;
-    private int numSelfAndUpstreamTasks_;
+    private int numSelfAndUpstreamTasks_ = 0;
     /**
      * <code>optional int32 num_self_and_upstream_tasks = 1;</code>
+     * @return Whether the numSelfAndUpstreamTasks field is set.
      */
+    @java.lang.Override
     public boolean hasNumSelfAndUpstreamTasks() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
+      return ((bitField0_ & 0x00000001) != 0);
     }
     /**
      * <code>optional int32 num_self_and_upstream_tasks = 1;</code>
+     * @return The numSelfAndUpstreamTasks.
      */
+    @java.lang.Override
     public int getNumSelfAndUpstreamTasks() {
       return numSelfAndUpstreamTasks_;
     }
 
-    // optional int32 num_self_and_upstream_completed_tasks = 2;
     public static final int NUM_SELF_AND_UPSTREAM_COMPLETED_TASKS_FIELD_NUMBER = 2;
-    private int numSelfAndUpstreamCompletedTasks_;
+    private int numSelfAndUpstreamCompletedTasks_ = 0;
     /**
      * <code>optional int32 num_self_and_upstream_completed_tasks = 2;</code>
+     * @return Whether the numSelfAndUpstreamCompletedTasks field is set.
      */
+    @java.lang.Override
     public boolean hasNumSelfAndUpstreamCompletedTasks() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
+      return ((bitField0_ & 0x00000002) != 0);
     }
     /**
      * <code>optional int32 num_self_and_upstream_completed_tasks = 2;</code>
+     * @return The numSelfAndUpstreamCompletedTasks.
      */
+    @java.lang.Override
     public int getNumSelfAndUpstreamCompletedTasks() {
       return numSelfAndUpstreamCompletedTasks_;
     }
 
-    // optional int32 within_dag_priority = 3;
     public static final int WITHIN_DAG_PRIORITY_FIELD_NUMBER = 3;
-    private int withinDagPriority_;
+    private int withinDagPriority_ = 0;
     /**
      * <code>optional int32 within_dag_priority = 3;</code>
+     * @return Whether the withinDagPriority field is set.
      */
+    @java.lang.Override
     public boolean hasWithinDagPriority() {
-      return ((bitField0_ & 0x00000004) == 0x00000004);
+      return ((bitField0_ & 0x00000004) != 0);
     }
     /**
      * <code>optional int32 within_dag_priority = 3;</code>
+     * @return The withinDagPriority.
      */
+    @java.lang.Override
     public int getWithinDagPriority() {
       return withinDagPriority_;
     }
 
-    // optional int64 dag_start_time = 4;
     public static final int DAG_START_TIME_FIELD_NUMBER = 4;
-    private long dagStartTime_;
+    private long dagStartTime_ = 0L;
     /**
      * <code>optional int64 dag_start_time = 4;</code>
+     * @return Whether the dagStartTime field is set.
      */
+    @java.lang.Override
     public boolean hasDagStartTime() {
-      return ((bitField0_ & 0x00000008) == 0x00000008);
+      return ((bitField0_ & 0x00000008) != 0);
     }
     /**
      * <code>optional int64 dag_start_time = 4;</code>
+     * @return The dagStartTime.
      */
+    @java.lang.Override
     public long getDagStartTime() {
       return dagStartTime_;
     }
 
-    // optional int64 first_attempt_start_time = 5;
     public static final int FIRST_ATTEMPT_START_TIME_FIELD_NUMBER = 5;
-    private long firstAttemptStartTime_;
+    private long firstAttemptStartTime_ = 0L;
     /**
      * <code>optional int64 first_attempt_start_time = 5;</code>
+     * @return Whether the firstAttemptStartTime field is set.
      */
+    @java.lang.Override
     public boolean hasFirstAttemptStartTime() {
-      return ((bitField0_ & 0x00000010) == 0x00000010);
+      return ((bitField0_ & 0x00000010) != 0);
     }
     /**
      * <code>optional int64 first_attempt_start_time = 5;</code>
+     * @return The firstAttemptStartTime.
      */
+    @java.lang.Override
     public long getFirstAttemptStartTime() {
       return firstAttemptStartTime_;
     }
 
-    // optional int64 current_attempt_start_time = 6;
     public static final int CURRENT_ATTEMPT_START_TIME_FIELD_NUMBER = 6;
-    private long currentAttemptStartTime_;
+    private long currentAttemptStartTime_ = 0L;
     /**
      * <code>optional int64 current_attempt_start_time = 6;</code>
+     * @return Whether the currentAttemptStartTime field is set.
      */
+    @java.lang.Override
     public boolean hasCurrentAttemptStartTime() {
-      return ((bitField0_ & 0x00000020) == 0x00000020);
+      return ((bitField0_ & 0x00000020) != 0);
     }
     /**
      * <code>optional int64 current_attempt_start_time = 6;</code>
+     * @return The currentAttemptStartTime.
      */
+    @java.lang.Override
     public long getCurrentAttemptStartTime() {
       return currentAttemptStartTime_;
     }
 
-    private void initFields() {
-      numSelfAndUpstreamTasks_ = 0;
-      numSelfAndUpstreamCompletedTasks_ = 0;
-      withinDagPriority_ = 0;
-      dagStartTime_ = 0L;
-      firstAttemptStartTime_ = 0L;
-      currentAttemptStartTime_ = 0L;
-    }
     private byte memoizedIsInitialized = -1;
+    @java.lang.Override
     public final boolean isInitialized() {
       byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
+      if (isInitialized == 1) return true;
+      if (isInitialized == 0) return false;
 
       memoizedIsInitialized = 1;
       return true;
     }
 
+    @java.lang.Override
     public void writeTo(com.google.protobuf.CodedOutputStream output)
                         throws java.io.IOException {
-      getSerializedSize();
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
+      if (((bitField0_ & 0x00000001) != 0)) {
         output.writeInt32(1, numSelfAndUpstreamTasks_);
       }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
+      if (((bitField0_ & 0x00000002) != 0)) {
         output.writeInt32(2, numSelfAndUpstreamCompletedTasks_);
       }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
+      if (((bitField0_ & 0x00000004) != 0)) {
         output.writeInt32(3, withinDagPriority_);
       }
-      if (((bitField0_ & 0x00000008) == 0x00000008)) {
+      if (((bitField0_ & 0x00000008) != 0)) {
         output.writeInt64(4, dagStartTime_);
       }
-      if (((bitField0_ & 0x00000010) == 0x00000010)) {
+      if (((bitField0_ & 0x00000010) != 0)) {
         output.writeInt64(5, firstAttemptStartTime_);
       }
-      if (((bitField0_ & 0x00000020) == 0x00000020)) {
+      if (((bitField0_ & 0x00000020) != 0)) {
         output.writeInt64(6, currentAttemptStartTime_);
       }
       getUnknownFields().writeTo(output);
     }
 
-    private int memoizedSerializedSize = -1;
+    @java.lang.Override
     public int getSerializedSize() {
-      int size = memoizedSerializedSize;
+      int size = memoizedSize;
       if (size != -1) return size;
 
       size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
+      if (((bitField0_ & 0x00000001) != 0)) {
         size += com.google.protobuf.CodedOutputStream
           .computeInt32Size(1, numSelfAndUpstreamTasks_);
       }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
+      if (((bitField0_ & 0x00000002) != 0)) {
         size += com.google.protobuf.CodedOutputStream
           .computeInt32Size(2, numSelfAndUpstreamCompletedTasks_);
       }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
+      if (((bitField0_ & 0x00000004) != 0)) {
         size += com.google.protobuf.CodedOutputStream
           .computeInt32Size(3, withinDagPriority_);
       }
-      if (((bitField0_ & 0x00000008) == 0x00000008)) {
+      if (((bitField0_ & 0x00000008) != 0)) {
         size += com.google.protobuf.CodedOutputStream
           .computeInt64Size(4, dagStartTime_);
       }
-      if (((bitField0_ & 0x00000010) == 0x00000010)) {
+      if (((bitField0_ & 0x00000010) != 0)) {
         size += com.google.protobuf.CodedOutputStream
           .computeInt64Size(5, firstAttemptStartTime_);
       }
-      if (((bitField0_ & 0x00000020) == 0x00000020)) {
+      if (((bitField0_ & 0x00000020) != 0)) {
         size += com.google.protobuf.CodedOutputStream
           .computeInt64Size(6, currentAttemptStartTime_);
       }
       size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
+      memoizedSize = size;
       return size;
     }
 
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
     @java.lang.Override
     public boolean equals(final java.lang.Object obj) {
       if (obj == this) {
@@ -7434,50 +8012,47 @@ public final class LlapDaemonProtocolProtos {
       }
       org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo) obj;
 
-      boolean result = true;
-      result = result && (hasNumSelfAndUpstreamTasks() == other.hasNumSelfAndUpstreamTasks());
+      if (hasNumSelfAndUpstreamTasks() != other.hasNumSelfAndUpstreamTasks()) return false;
       if (hasNumSelfAndUpstreamTasks()) {
-        result = result && (getNumSelfAndUpstreamTasks()
-            == other.getNumSelfAndUpstreamTasks());
+        if (getNumSelfAndUpstreamTasks()
+            != other.getNumSelfAndUpstreamTasks()) return false;
       }
-      result = result && (hasNumSelfAndUpstreamCompletedTasks() == other.hasNumSelfAndUpstreamCompletedTasks());
+      if (hasNumSelfAndUpstreamCompletedTasks() != other.hasNumSelfAndUpstreamCompletedTasks()) return false;
       if (hasNumSelfAndUpstreamCompletedTasks()) {
-        result = result && (getNumSelfAndUpstreamCompletedTasks()
-            == other.getNumSelfAndUpstreamCompletedTasks());
+        if (getNumSelfAndUpstreamCompletedTasks()
+            != other.getNumSelfAndUpstreamCompletedTasks()) return false;
       }
-      result = result && (hasWithinDagPriority() == other.hasWithinDagPriority());
+      if (hasWithinDagPriority() != other.hasWithinDagPriority()) return false;
       if (hasWithinDagPriority()) {
-        result = result && (getWithinDagPriority()
-            == other.getWithinDagPriority());
+        if (getWithinDagPriority()
+            != other.getWithinDagPriority()) return false;
       }
-      result = result && (hasDagStartTime() == other.hasDagStartTime());
+      if (hasDagStartTime() != other.hasDagStartTime()) return false;
       if (hasDagStartTime()) {
-        result = result && (getDagStartTime()
-            == other.getDagStartTime());
+        if (getDagStartTime()
+            != other.getDagStartTime()) return false;
       }
-      result = result && (hasFirstAttemptStartTime() == other.hasFirstAttemptStartTime());
+      if (hasFirstAttemptStartTime() != other.hasFirstAttemptStartTime()) return false;
       if (hasFirstAttemptStartTime()) {
-        result = result && (getFirstAttemptStartTime()
-            == other.getFirstAttemptStartTime());
+        if (getFirstAttemptStartTime()
+            != other.getFirstAttemptStartTime()) return false;
       }
-      result = result && (hasCurrentAttemptStartTime() == other.hasCurrentAttemptStartTime());
+      if (hasCurrentAttemptStartTime() != other.hasCurrentAttemptStartTime()) return false;
       if (hasCurrentAttemptStartTime()) {
-        result = result && (getCurrentAttemptStartTime()
-            == other.getCurrentAttemptStartTime());
+        if (getCurrentAttemptStartTime()
+            != other.getCurrentAttemptStartTime()) return false;
       }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
+      if (!getUnknownFields().equals(other.getUnknownFields())) return false;
+      return true;
     }
 
-    private int memoizedHashCode = 0;
     @java.lang.Override
     public int hashCode() {
       if (memoizedHashCode != 0) {
         return memoizedHashCode;
       }
       int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
+      hash = (19 * hash) + getDescriptor().hashCode();
       if (hasNumSelfAndUpstreamTasks()) {
         hash = (37 * hash) + NUM_SELF_AND_UPSTREAM_TASKS_FIELD_NUMBER;
         hash = (53 * hash) + getNumSelfAndUpstreamTasks();
@@ -7492,21 +8067,35 @@ public final class LlapDaemonProtocolProtos {
       }
       if (hasDagStartTime()) {
         hash = (37 * hash) + DAG_START_TIME_FIELD_NUMBER;
-        hash = (53 * hash) + hashLong(getDagStartTime());
+        hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+            getDagStartTime());
       }
       if (hasFirstAttemptStartTime()) {
         hash = (37 * hash) + FIRST_ATTEMPT_START_TIME_FIELD_NUMBER;
-        hash = (53 * hash) + hashLong(getFirstAttemptStartTime());
+        hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+            getFirstAttemptStartTime());
       }
       if (hasCurrentAttemptStartTime()) {
         hash = (37 * hash) + CURRENT_ATTEMPT_START_TIME_FIELD_NUMBER;
-        hash = (53 * hash) + hashLong(getCurrentAttemptStartTime());
+        hash = (53 * hash) + com.google.protobuf.Internal.hashLong(
+            getCurrentAttemptStartTime());
       }
       hash = (29 * hash) + getUnknownFields().hashCode();
       memoizedHashCode = hash;
       return hash;
     }
 
+    public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom(
+        java.nio.ByteBuffer data)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      return PARSER.parseFrom(data);
+    }
+    public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom(
+        java.nio.ByteBuffer data,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      return PARSER.parseFrom(data, extensionRegistry);
+    }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom(
         com.google.protobuf.ByteString data)
         throws com.google.protobuf.InvalidProtocolBufferException {
@@ -7530,46 +8119,61 @@ public final class LlapDaemonProtocolProtos {
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return PARSER.parseFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input);
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input, extensionRegistry);
     }
+
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseDelimitedFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseDelimitedWithIOException(PARSER, input);
     }
+
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseDelimitedFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom(
         com.google.protobuf.CodedInputStream input)
         throws java.io.IOException {
-      return PARSER.parseFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input);
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom(
         com.google.protobuf.CodedInputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input, extensionRegistry);
     }
 
-    public static Builder newBuilder() { return Builder.create(); }
+    @java.lang.Override
     public Builder newBuilderForType() { return newBuilder(); }
+    public static Builder newBuilder() {
+      return DEFAULT_INSTANCE.toBuilder();
+    }
     public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo prototype) {
-      return newBuilder().mergeFrom(prototype);
+      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+    }
+    @java.lang.Override
+    public Builder toBuilder() {
+      return this == DEFAULT_INSTANCE
+          ? new Builder() : new Builder().mergeFrom(this);
     }
-    public Builder toBuilder() { return newBuilder(this); }
 
     @java.lang.Override
     protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+        com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
       Builder builder = new Builder(parent);
       return builder;
     }
@@ -7577,14 +8181,16 @@ public final class LlapDaemonProtocolProtos {
      * Protobuf type {@code FragmentRuntimeInfo}
      */
     public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfoOrBuilder {
+        com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+        // @@protoc_insertion_point(builder_implements:FragmentRuntimeInfo)
+        org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfoOrBuilder {
       public static final com.google.protobuf.Descriptors.Descriptor
           getDescriptor() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentRuntimeInfo_descriptor;
       }
 
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+      @java.lang.Override
+      protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
           internalGetFieldAccessorTable() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentRuntimeInfo_fieldAccessorTable
             .ensureFieldAccessorsInitialized(
@@ -7593,52 +8199,39 @@ public final class LlapDaemonProtocolProtos {
 
       // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.newBuilder()
       private Builder() {
-        maybeForceBuilderInitialization();
+
       }
 
       private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+          com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
         super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-        }
-      }
-      private static Builder create() {
-        return new Builder();
-      }
 
+      }
+      @java.lang.Override
       public Builder clear() {
         super.clear();
+        bitField0_ = 0;
         numSelfAndUpstreamTasks_ = 0;
-        bitField0_ = (bitField0_ & ~0x00000001);
         numSelfAndUpstreamCompletedTasks_ = 0;
-        bitField0_ = (bitField0_ & ~0x00000002);
         withinDagPriority_ = 0;
-        bitField0_ = (bitField0_ & ~0x00000004);
         dagStartTime_ = 0L;
-        bitField0_ = (bitField0_ & ~0x00000008);
         firstAttemptStartTime_ = 0L;
-        bitField0_ = (bitField0_ & ~0x00000010);
         currentAttemptStartTime_ = 0L;
-        bitField0_ = (bitField0_ & ~0x00000020);
         return this;
       }
 
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
+      @java.lang.Override
       public com.google.protobuf.Descriptors.Descriptor
           getDescriptorForType() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentRuntimeInfo_descriptor;
       }
 
+      @java.lang.Override
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo getDefaultInstanceForType() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.getDefaultInstance();
       }
 
+      @java.lang.Override
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo build() {
         org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo result = buildPartial();
         if (!result.isInitialized()) {
@@ -7647,39 +8240,77 @@ public final class LlapDaemonProtocolProtos {
         return result;
       }
 
+      @java.lang.Override
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo buildPartial() {
         org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo(this);
+        if (bitField0_ != 0) { buildPartial0(result); }
+        onBuilt();
+        return result;
+      }
+
+      private void buildPartial0(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo result) {
         int from_bitField0_ = bitField0_;
         int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
+        if (((from_bitField0_ & 0x00000001) != 0)) {
+          result.numSelfAndUpstreamTasks_ = numSelfAndUpstreamTasks_;
           to_bitField0_ |= 0x00000001;
         }
-        result.numSelfAndUpstreamTasks_ = numSelfAndUpstreamTasks_;
-        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
+        if (((from_bitField0_ & 0x00000002) != 0)) {
+          result.numSelfAndUpstreamCompletedTasks_ = numSelfAndUpstreamCompletedTasks_;
           to_bitField0_ |= 0x00000002;
         }
-        result.numSelfAndUpstreamCompletedTasks_ = numSelfAndUpstreamCompletedTasks_;
-        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
+        if (((from_bitField0_ & 0x00000004) != 0)) {
+          result.withinDagPriority_ = withinDagPriority_;
           to_bitField0_ |= 0x00000004;
         }
-        result.withinDagPriority_ = withinDagPriority_;
-        if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
+        if (((from_bitField0_ & 0x00000008) != 0)) {
+          result.dagStartTime_ = dagStartTime_;
           to_bitField0_ |= 0x00000008;
         }
-        result.dagStartTime_ = dagStartTime_;
-        if (((from_bitField0_ & 0x00000010) == 0x00000010)) {
+        if (((from_bitField0_ & 0x00000010) != 0)) {
+          result.firstAttemptStartTime_ = firstAttemptStartTime_;
           to_bitField0_ |= 0x00000010;
         }
-        result.firstAttemptStartTime_ = firstAttemptStartTime_;
-        if (((from_bitField0_ & 0x00000020) == 0x00000020)) {
+        if (((from_bitField0_ & 0x00000020) != 0)) {
+          result.currentAttemptStartTime_ = currentAttemptStartTime_;
           to_bitField0_ |= 0x00000020;
         }
-        result.currentAttemptStartTime_ = currentAttemptStartTime_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
+        result.bitField0_ |= to_bitField0_;
       }
 
+      @java.lang.Override
+      public Builder clone() {
+        return super.clone();
+      }
+      @java.lang.Override
+      public Builder setField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          java.lang.Object value) {
+        return super.setField(field, value);
+      }
+      @java.lang.Override
+      public Builder clearField(
+          com.google.protobuf.Descriptors.FieldDescriptor field) {
+        return super.clearField(field);
+      }
+      @java.lang.Override
+      public Builder clearOneof(
+          com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+        return super.clearOneof(oneof);
+      }
+      @java.lang.Override
+      public Builder setRepeatedField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          int index, java.lang.Object value) {
+        return super.setRepeatedField(field, index, value);
+      }
+      @java.lang.Override
+      public Builder addRepeatedField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          java.lang.Object value) {
+        return super.addRepeatedField(field, value);
+      }
+      @java.lang.Override
       public Builder mergeFrom(com.google.protobuf.Message other) {
         if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo) {
           return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo)other);
@@ -7710,57 +8341,110 @@ public final class LlapDaemonProtocolProtos {
           setCurrentAttemptStartTime(other.getCurrentAttemptStartTime());
         }
         this.mergeUnknownFields(other.getUnknownFields());
+        onChanged();
         return this;
       }
 
+      @java.lang.Override
       public final boolean isInitialized() {
         return true;
       }
 
+      @java.lang.Override
       public Builder mergeFrom(
           com.google.protobuf.CodedInputStream input,
           com.google.protobuf.ExtensionRegistryLite extensionRegistry)
           throws java.io.IOException {
-        org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parsedMessage = null;
+        if (extensionRegistry == null) {
+          throw new java.lang.NullPointerException();
+        }
         try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+          boolean done = false;
+          while (!done) {
+            int tag = input.readTag();
+            switch (tag) {
+              case 0:
+                done = true;
+                break;
+              case 8: {
+                numSelfAndUpstreamTasks_ = input.readInt32();
+                bitField0_ |= 0x00000001;
+                break;
+              } // case 8
+              case 16: {
+                numSelfAndUpstreamCompletedTasks_ = input.readInt32();
+                bitField0_ |= 0x00000002;
+                break;
+              } // case 16
+              case 24: {
+                withinDagPriority_ = input.readInt32();
+                bitField0_ |= 0x00000004;
+                break;
+              } // case 24
+              case 32: {
+                dagStartTime_ = input.readInt64();
+                bitField0_ |= 0x00000008;
+                break;
+              } // case 32
+              case 40: {
+                firstAttemptStartTime_ = input.readInt64();
+                bitField0_ |= 0x00000010;
+                break;
+              } // case 40
+              case 48: {
+                currentAttemptStartTime_ = input.readInt64();
+                bitField0_ |= 0x00000020;
+                break;
+              } // case 48
+              default: {
+                if (!super.parseUnknownField(input, extensionRegistry, tag)) {
+                  done = true; // was an endgroup tag
+                }
+                break;
+              } // default:
+            } // switch (tag)
+          } // while (!done)
         } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo) e.getUnfinishedMessage();
-          throw e;
+          throw e.unwrapIOException();
         } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
+          onChanged();
+        } // finally
         return this;
       }
       private int bitField0_;
 
-      // optional int32 num_self_and_upstream_tasks = 1;
       private int numSelfAndUpstreamTasks_ ;
       /**
        * <code>optional int32 num_self_and_upstream_tasks = 1;</code>
+       * @return Whether the numSelfAndUpstreamTasks field is set.
        */
+      @java.lang.Override
       public boolean hasNumSelfAndUpstreamTasks() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
+        return ((bitField0_ & 0x00000001) != 0);
       }
       /**
        * <code>optional int32 num_self_and_upstream_tasks = 1;</code>
+       * @return The numSelfAndUpstreamTasks.
        */
+      @java.lang.Override
       public int getNumSelfAndUpstreamTasks() {
         return numSelfAndUpstreamTasks_;
       }
       /**
        * <code>optional int32 num_self_and_upstream_tasks = 1;</code>
+       * @param value The numSelfAndUpstreamTasks to set.
+       * @return This builder for chaining.
        */
       public Builder setNumSelfAndUpstreamTasks(int value) {
-        bitField0_ |= 0x00000001;
+
         numSelfAndUpstreamTasks_ = value;
+        bitField0_ |= 0x00000001;
         onChanged();
         return this;
       }
       /**
        * <code>optional int32 num_self_and_upstream_tasks = 1;</code>
+       * @return This builder for chaining.
        */
       public Builder clearNumSelfAndUpstreamTasks() {
         bitField0_ = (bitField0_ & ~0x00000001);
@@ -7769,31 +8453,38 @@ public final class LlapDaemonProtocolProtos {
         return this;
       }
 
-      // optional int32 num_self_and_upstream_completed_tasks = 2;
       private int numSelfAndUpstreamCompletedTasks_ ;
       /**
        * <code>optional int32 num_self_and_upstream_completed_tasks = 2;</code>
+       * @return Whether the numSelfAndUpstreamCompletedTasks field is set.
        */
+      @java.lang.Override
       public boolean hasNumSelfAndUpstreamCompletedTasks() {
-        return ((bitField0_ & 0x00000002) == 0x00000002);
+        return ((bitField0_ & 0x00000002) != 0);
       }
       /**
        * <code>optional int32 num_self_and_upstream_completed_tasks = 2;</code>
+       * @return The numSelfAndUpstreamCompletedTasks.
        */
+      @java.lang.Override
       public int getNumSelfAndUpstreamCompletedTasks() {
         return numSelfAndUpstreamCompletedTasks_;
       }
       /**
        * <code>optional int32 num_self_and_upstream_completed_tasks = 2;</code>
+       * @param value The numSelfAndUpstreamCompletedTasks to set.
+       * @return This builder for chaining.
        */
       public Builder setNumSelfAndUpstreamCompletedTasks(int value) {
-        bitField0_ |= 0x00000002;
+
         numSelfAndUpstreamCompletedTasks_ = value;
+        bitField0_ |= 0x00000002;
         onChanged();
         return this;
       }
       /**
        * <code>optional int32 num_self_and_upstream_completed_tasks = 2;</code>
+       * @return This builder for chaining.
        */
       public Builder clearNumSelfAndUpstreamCompletedTasks() {
         bitField0_ = (bitField0_ & ~0x00000002);
@@ -7802,31 +8493,38 @@ public final class LlapDaemonProtocolProtos {
         return this;
       }
 
-      // optional int32 within_dag_priority = 3;
       private int withinDagPriority_ ;
       /**
        * <code>optional int32 within_dag_priority = 3;</code>
+       * @return Whether the withinDagPriority field is set.
        */
+      @java.lang.Override
       public boolean hasWithinDagPriority() {
-        return ((bitField0_ & 0x00000004) == 0x00000004);
+        return ((bitField0_ & 0x00000004) != 0);
       }
       /**
        * <code>optional int32 within_dag_priority = 3;</code>
+       * @return The withinDagPriority.
        */
+      @java.lang.Override
       public int getWithinDagPriority() {
         return withinDagPriority_;
       }
       /**
        * <code>optional int32 within_dag_priority = 3;</code>
+       * @param value The withinDagPriority to set.
+       * @return This builder for chaining.
        */
       public Builder setWithinDagPriority(int value) {
-        bitField0_ |= 0x00000004;
+
         withinDagPriority_ = value;
+        bitField0_ |= 0x00000004;
         onChanged();
         return this;
       }
       /**
        * <code>optional int32 within_dag_priority = 3;</code>
+       * @return This builder for chaining.
        */
       public Builder clearWithinDagPriority() {
         bitField0_ = (bitField0_ & ~0x00000004);
@@ -7835,31 +8533,38 @@ public final class LlapDaemonProtocolProtos {
         return this;
       }
 
-      // optional int64 dag_start_time = 4;
       private long dagStartTime_ ;
       /**
        * <code>optional int64 dag_start_time = 4;</code>
+       * @return Whether the dagStartTime field is set.
        */
+      @java.lang.Override
       public boolean hasDagStartTime() {
-        return ((bitField0_ & 0x00000008) == 0x00000008);
+        return ((bitField0_ & 0x00000008) != 0);
       }
       /**
        * <code>optional int64 dag_start_time = 4;</code>
+       * @return The dagStartTime.
        */
+      @java.lang.Override
       public long getDagStartTime() {
         return dagStartTime_;
       }
       /**
        * <code>optional int64 dag_start_time = 4;</code>
+       * @param value The dagStartTime to set.
+       * @return This builder for chaining.
        */
       public Builder setDagStartTime(long value) {
-        bitField0_ |= 0x00000008;
+
         dagStartTime_ = value;
+        bitField0_ |= 0x00000008;
         onChanged();
         return this;
       }
       /**
        * <code>optional int64 dag_start_time = 4;</code>
+       * @return This builder for chaining.
        */
       public Builder clearDagStartTime() {
         bitField0_ = (bitField0_ & ~0x00000008);
@@ -7868,31 +8573,38 @@ public final class LlapDaemonProtocolProtos {
         return this;
       }
 
-      // optional int64 first_attempt_start_time = 5;
       private long firstAttemptStartTime_ ;
       /**
        * <code>optional int64 first_attempt_start_time = 5;</code>
+       * @return Whether the firstAttemptStartTime field is set.
        */
+      @java.lang.Override
       public boolean hasFirstAttemptStartTime() {
-        return ((bitField0_ & 0x00000010) == 0x00000010);
+        return ((bitField0_ & 0x00000010) != 0);
       }
       /**
        * <code>optional int64 first_attempt_start_time = 5;</code>
+       * @return The firstAttemptStartTime.
        */
+      @java.lang.Override
       public long getFirstAttemptStartTime() {
         return firstAttemptStartTime_;
       }
       /**
        * <code>optional int64 first_attempt_start_time = 5;</code>
+       * @param value The firstAttemptStartTime to set.
+       * @return This builder for chaining.
        */
       public Builder setFirstAttemptStartTime(long value) {
-        bitField0_ |= 0x00000010;
+
         firstAttemptStartTime_ = value;
+        bitField0_ |= 0x00000010;
         onChanged();
         return this;
       }
       /**
        * <code>optional int64 first_attempt_start_time = 5;</code>
+       * @return This builder for chaining.
        */
       public Builder clearFirstAttemptStartTime() {
         bitField0_ = (bitField0_ & ~0x00000010);
@@ -7901,31 +8613,38 @@ public final class LlapDaemonProtocolProtos {
         return this;
       }
 
-      // optional int64 current_attempt_start_time = 6;
       private long currentAttemptStartTime_ ;
       /**
        * <code>optional int64 current_attempt_start_time = 6;</code>
+       * @return Whether the currentAttemptStartTime field is set.
        */
+      @java.lang.Override
       public boolean hasCurrentAttemptStartTime() {
-        return ((bitField0_ & 0x00000020) == 0x00000020);
+        return ((bitField0_ & 0x00000020) != 0);
       }
       /**
        * <code>optional int64 current_attempt_start_time = 6;</code>
+       * @return The currentAttemptStartTime.
        */
+      @java.lang.Override
       public long getCurrentAttemptStartTime() {
         return currentAttemptStartTime_;
       }
       /**
        * <code>optional int64 current_attempt_start_time = 6;</code>
+       * @param value The currentAttemptStartTime to set.
+       * @return This builder for chaining.
        */
       public Builder setCurrentAttemptStartTime(long value) {
-        bitField0_ |= 0x00000020;
+
         currentAttemptStartTime_ = value;
+        bitField0_ |= 0x00000020;
         onChanged();
         return this;
       }
       /**
        * <code>optional int64 current_attempt_start_time = 6;</code>
+       * @return This builder for chaining.
        */
       public Builder clearCurrentAttemptStartTime() {
         bitField0_ = (bitField0_ & ~0x00000020);
@@ -7933,53 +8652,110 @@ public final class LlapDaemonProtocolProtos {
         onChanged();
         return this;
       }
+      @java.lang.Override
+      public final Builder setUnknownFields(
+          final com.google.protobuf.UnknownFieldSet unknownFields) {
+        return super.setUnknownFields(unknownFields);
+      }
+
+      @java.lang.Override
+      public final Builder mergeUnknownFields(
+          final com.google.protobuf.UnknownFieldSet unknownFields) {
+        return super.mergeUnknownFields(unknownFields);
+      }
+
 
       // @@protoc_insertion_point(builder_scope:FragmentRuntimeInfo)
     }
 
+    // @@protoc_insertion_point(class_scope:FragmentRuntimeInfo)
+    private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo DEFAULT_INSTANCE;
     static {
-      defaultInstance = new FragmentRuntimeInfo(true);
-      defaultInstance.initFields();
+      DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo();
+    }
+
+    public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo getDefaultInstance() {
+      return DEFAULT_INSTANCE;
+    }
+
+    @java.lang.Deprecated public static final com.google.protobuf.Parser<FragmentRuntimeInfo>
+        PARSER = new com.google.protobuf.AbstractParser<FragmentRuntimeInfo>() {
+      @java.lang.Override
+      public FragmentRuntimeInfo parsePartialFrom(
+          com.google.protobuf.CodedInputStream input,
+          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+          throws com.google.protobuf.InvalidProtocolBufferException {
+        Builder builder = newBuilder();
+        try {
+          builder.mergeFrom(input, extensionRegistry);
+        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+          throw e.setUnfinishedMessage(builder.buildPartial());
+        } catch (com.google.protobuf.UninitializedMessageException e) {
+          throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
+        } catch (java.io.IOException e) {
+          throw new com.google.protobuf.InvalidProtocolBufferException(e)
+              .setUnfinishedMessage(builder.buildPartial());
+        }
+        return builder.buildPartial();
+      }
+    };
+
+    public static com.google.protobuf.Parser<FragmentRuntimeInfo> parser() {
+      return PARSER;
+    }
+
+    @java.lang.Override
+    public com.google.protobuf.Parser<FragmentRuntimeInfo> getParserForType() {
+      return PARSER;
+    }
+
+    @java.lang.Override
+    public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo getDefaultInstanceForType() {
+      return DEFAULT_INSTANCE;
     }
 
-    // @@protoc_insertion_point(class_scope:FragmentRuntimeInfo)
   }
 
-  public interface QueryIdentifierProtoOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
+  public interface QueryIdentifierProtoOrBuilder extends
+      // @@protoc_insertion_point(interface_extends:QueryIdentifierProto)
+      com.google.protobuf.MessageOrBuilder {
 
-    // optional string application_id_string = 1;
     /**
      * <code>optional string application_id_string = 1;</code>
+     * @return Whether the applicationIdString field is set.
      */
     boolean hasApplicationIdString();
     /**
      * <code>optional string application_id_string = 1;</code>
+     * @return The applicationIdString.
      */
     java.lang.String getApplicationIdString();
     /**
      * <code>optional string application_id_string = 1;</code>
+     * @return The bytes for applicationIdString.
      */
     com.google.protobuf.ByteString
         getApplicationIdStringBytes();
 
-    // optional int32 dag_index = 2;
     /**
      * <code>optional int32 dag_index = 2;</code>
+     * @return Whether the dagIndex field is set.
      */
     boolean hasDagIndex();
     /**
      * <code>optional int32 dag_index = 2;</code>
+     * @return The dagIndex.
      */
     int getDagIndex();
 
-    // optional int32 app_attempt_number = 3;
     /**
      * <code>optional int32 app_attempt_number = 3;</code>
+     * @return Whether the appAttemptNumber field is set.
      */
     boolean hasAppAttemptNumber();
     /**
      * <code>optional int32 app_attempt_number = 3;</code>
+     * @return The appAttemptNumber.
      */
     int getAppAttemptNumber();
   }
@@ -7987,120 +8763,55 @@ public final class LlapDaemonProtocolProtos {
    * Protobuf type {@code QueryIdentifierProto}
    */
   public static final class QueryIdentifierProto extends
-      com.google.protobuf.GeneratedMessage
-      implements QueryIdentifierProtoOrBuilder {
+      com.google.protobuf.GeneratedMessageV3 implements
+      // @@protoc_insertion_point(message_implements:QueryIdentifierProto)
+      QueryIdentifierProtoOrBuilder {
+  private static final long serialVersionUID = 0L;
     // Use QueryIdentifierProto.newBuilder() to construct.
-    private QueryIdentifierProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+    private QueryIdentifierProto(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
       super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private QueryIdentifierProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final QueryIdentifierProto defaultInstance;
-    public static QueryIdentifierProto getDefaultInstance() {
-      return defaultInstance;
     }
-
-    public QueryIdentifierProto getDefaultInstanceForType() {
-      return defaultInstance;
+    private QueryIdentifierProto() {
+      applicationIdString_ = "";
     }
 
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
     @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private QueryIdentifierProto(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              bitField0_ |= 0x00000001;
-              applicationIdString_ = input.readBytes();
-              break;
-            }
-            case 16: {
-              bitField0_ |= 0x00000002;
-              dagIndex_ = input.readInt32();
-              break;
-            }
-            case 24: {
-              bitField0_ |= 0x00000004;
-              appAttemptNumber_ = input.readInt32();
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
+    @SuppressWarnings({"unused"})
+    protected java.lang.Object newInstance(
+        UnusedPrivateParameter unused) {
+      return new QueryIdentifierProto();
     }
+
     public static final com.google.protobuf.Descriptors.Descriptor
         getDescriptor() {
       return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_descriptor;
     }
 
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+    @java.lang.Override
+    protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
         internalGetFieldAccessorTable() {
       return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_fieldAccessorTable
           .ensureFieldAccessorsInitialized(
               org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder.class);
     }
 
-    public static com.google.protobuf.Parser<QueryIdentifierProto> PARSER =
-        new com.google.protobuf.AbstractParser<QueryIdentifierProto>() {
-      public QueryIdentifierProto parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new QueryIdentifierProto(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<QueryIdentifierProto> getParserForType() {
-      return PARSER;
-    }
-
     private int bitField0_;
-    // optional string application_id_string = 1;
     public static final int APPLICATION_ID_STRING_FIELD_NUMBER = 1;
-    private java.lang.Object applicationIdString_;
+    @SuppressWarnings("serial")
+    private volatile java.lang.Object applicationIdString_ = "";
     /**
      * <code>optional string application_id_string = 1;</code>
+     * @return Whether the applicationIdString field is set.
      */
+    @java.lang.Override
     public boolean hasApplicationIdString() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
+      return ((bitField0_ & 0x00000001) != 0);
     }
     /**
      * <code>optional string application_id_string = 1;</code>
+     * @return The applicationIdString.
      */
+    @java.lang.Override
     public java.lang.String getApplicationIdString() {
       java.lang.Object ref = applicationIdString_;
       if (ref instanceof java.lang.String) {
@@ -8117,7 +8828,9 @@ public final class LlapDaemonProtocolProtos {
     }
     /**
      * <code>optional string application_id_string = 1;</code>
+     * @return The bytes for applicationIdString.
      */
+    @java.lang.Override
     public com.google.protobuf.ByteString
         getApplicationIdStringBytes() {
       java.lang.Object ref = applicationIdString_;
@@ -8132,97 +8845,92 @@ public final class LlapDaemonProtocolProtos {
       }
     }
 
-    // optional int32 dag_index = 2;
     public static final int DAG_INDEX_FIELD_NUMBER = 2;
-    private int dagIndex_;
+    private int dagIndex_ = 0;
     /**
      * <code>optional int32 dag_index = 2;</code>
+     * @return Whether the dagIndex field is set.
      */
+    @java.lang.Override
     public boolean hasDagIndex() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
+      return ((bitField0_ & 0x00000002) != 0);
     }
     /**
      * <code>optional int32 dag_index = 2;</code>
+     * @return The dagIndex.
      */
+    @java.lang.Override
     public int getDagIndex() {
       return dagIndex_;
     }
 
-    // optional int32 app_attempt_number = 3;
     public static final int APP_ATTEMPT_NUMBER_FIELD_NUMBER = 3;
-    private int appAttemptNumber_;
+    private int appAttemptNumber_ = 0;
     /**
      * <code>optional int32 app_attempt_number = 3;</code>
+     * @return Whether the appAttemptNumber field is set.
      */
+    @java.lang.Override
     public boolean hasAppAttemptNumber() {
-      return ((bitField0_ & 0x00000004) == 0x00000004);
+      return ((bitField0_ & 0x00000004) != 0);
     }
     /**
      * <code>optional int32 app_attempt_number = 3;</code>
+     * @return The appAttemptNumber.
      */
+    @java.lang.Override
     public int getAppAttemptNumber() {
       return appAttemptNumber_;
     }
 
-    private void initFields() {
-      applicationIdString_ = "";
-      dagIndex_ = 0;
-      appAttemptNumber_ = 0;
-    }
     private byte memoizedIsInitialized = -1;
+    @java.lang.Override
     public final boolean isInitialized() {
       byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
+      if (isInitialized == 1) return true;
+      if (isInitialized == 0) return false;
 
       memoizedIsInitialized = 1;
       return true;
     }
 
+    @java.lang.Override
     public void writeTo(com.google.protobuf.CodedOutputStream output)
                         throws java.io.IOException {
-      getSerializedSize();
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeBytes(1, getApplicationIdStringBytes());
+      if (((bitField0_ & 0x00000001) != 0)) {
+        com.google.protobuf.GeneratedMessageV3.writeString(output, 1, applicationIdString_);
       }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
+      if (((bitField0_ & 0x00000002) != 0)) {
         output.writeInt32(2, dagIndex_);
       }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
+      if (((bitField0_ & 0x00000004) != 0)) {
         output.writeInt32(3, appAttemptNumber_);
       }
       getUnknownFields().writeTo(output);
     }
 
-    private int memoizedSerializedSize = -1;
+    @java.lang.Override
     public int getSerializedSize() {
-      int size = memoizedSerializedSize;
+      int size = memoizedSize;
       if (size != -1) return size;
 
       size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(1, getApplicationIdStringBytes());
+      if (((bitField0_ & 0x00000001) != 0)) {
+        size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, applicationIdString_);
       }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
+      if (((bitField0_ & 0x00000002) != 0)) {
         size += com.google.protobuf.CodedOutputStream
           .computeInt32Size(2, dagIndex_);
       }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
+      if (((bitField0_ & 0x00000004) != 0)) {
         size += com.google.protobuf.CodedOutputStream
           .computeInt32Size(3, appAttemptNumber_);
       }
       size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
+      memoizedSize = size;
       return size;
     }
 
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
     @java.lang.Override
     public boolean equals(final java.lang.Object obj) {
       if (obj == this) {
@@ -8233,35 +8941,32 @@ public final class LlapDaemonProtocolProtos {
       }
       org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto) obj;
 
-      boolean result = true;
-      result = result && (hasApplicationIdString() == other.hasApplicationIdString());
+      if (hasApplicationIdString() != other.hasApplicationIdString()) return false;
       if (hasApplicationIdString()) {
-        result = result && getApplicationIdString()
-            .equals(other.getApplicationIdString());
+        if (!getApplicationIdString()
+            .equals(other.getApplicationIdString())) return false;
       }
-      result = result && (hasDagIndex() == other.hasDagIndex());
+      if (hasDagIndex() != other.hasDagIndex()) return false;
       if (hasDagIndex()) {
-        result = result && (getDagIndex()
-            == other.getDagIndex());
+        if (getDagIndex()
+            != other.getDagIndex()) return false;
       }
-      result = result && (hasAppAttemptNumber() == other.hasAppAttemptNumber());
+      if (hasAppAttemptNumber() != other.hasAppAttemptNumber()) return false;
       if (hasAppAttemptNumber()) {
-        result = result && (getAppAttemptNumber()
-            == other.getAppAttemptNumber());
+        if (getAppAttemptNumber()
+            != other.getAppAttemptNumber()) return false;
       }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
+      if (!getUnknownFields().equals(other.getUnknownFields())) return false;
+      return true;
     }
 
-    private int memoizedHashCode = 0;
     @java.lang.Override
     public int hashCode() {
       if (memoizedHashCode != 0) {
         return memoizedHashCode;
       }
       int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
+      hash = (19 * hash) + getDescriptor().hashCode();
       if (hasApplicationIdString()) {
         hash = (37 * hash) + APPLICATION_ID_STRING_FIELD_NUMBER;
         hash = (53 * hash) + getApplicationIdString().hashCode();
@@ -8279,6 +8984,17 @@ public final class LlapDaemonProtocolProtos {
       return hash;
     }
 
+    public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(
+        java.nio.ByteBuffer data)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      return PARSER.parseFrom(data);
+    }
+    public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(
+        java.nio.ByteBuffer data,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      return PARSER.parseFrom(data, extensionRegistry);
+    }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(
         com.google.protobuf.ByteString data)
         throws com.google.protobuf.InvalidProtocolBufferException {
@@ -8302,46 +9018,61 @@ public final class LlapDaemonProtocolProtos {
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return PARSER.parseFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input);
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input, extensionRegistry);
     }
+
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseDelimitedFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseDelimitedWithIOException(PARSER, input);
     }
+
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseDelimitedFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(
         com.google.protobuf.CodedInputStream input)
         throws java.io.IOException {
-      return PARSER.parseFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input);
     }
     public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(
         com.google.protobuf.CodedInputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input, extensionRegistry);
     }
 
-    public static Builder newBuilder() { return Builder.create(); }
+    @java.lang.Override
     public Builder newBuilderForType() { return newBuilder(); }
+    public static Builder newBuilder() {
+      return DEFAULT_INSTANCE.toBuilder();
+    }
     public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto prototype) {
-      return newBuilder().mergeFrom(prototype);
+      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+    }
+    @java.lang.Override
+    public Builder toBuilder() {
+      return this == DEFAULT_INSTANCE
+          ? new Builder() : new Builder().mergeFrom(this);
     }
-    public Builder toBuilder() { return newBuilder(this); }
 
     @java.lang.Override
     protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+        com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
       Builder builder = new Builder(parent);
       return builder;
     }
@@ -8349,14 +9080,16 @@ public final class LlapDaemonProtocolProtos {
      * Protobuf type {@code QueryIdentifierProto}
      */
     public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder {
+        com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+        // @@protoc_insertion_point(builder_implements:QueryIdentifierProto)
+        org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder {
       public static final com.google.protobuf.Descriptors.Descriptor
           getDescriptor() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_descriptor;
       }
 
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+      @java.lang.Override
+      protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
           internalGetFieldAccessorTable() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_fieldAccessorTable
             .ensureFieldAccessorsInitialized(
@@ -8365,46 +9098,36 @@ public final class LlapDaemonProtocolProtos {
 
       // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.newBuilder()
       private Builder() {
-        maybeForceBuilderInitialization();
+
       }
 
       private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+          com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
         super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-        }
-      }
-      private static Builder create() {
-        return new Builder();
-      }
 
+      }
+      @java.lang.Override
       public Builder clear() {
         super.clear();
+        bitField0_ = 0;
         applicationIdString_ = "";
-        bitField0_ = (bitField0_ & ~0x00000001);
         dagIndex_ = 0;
-        bitField0_ = (bitField0_ & ~0x00000002);
         appAttemptNumber_ = 0;
-        bitField0_ = (bitField0_ & ~0x00000004);
         return this;
       }
 
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
+      @java.lang.Override
       public com.google.protobuf.Descriptors.Descriptor
           getDescriptorForType() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_descriptor;
       }
 
+      @java.lang.Override
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getDefaultInstanceForType() {
         return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance();
       }
 
+      @java.lang.Override
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto build() {
         org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto result = buildPartial();
         if (!result.isInitialized()) {
@@ -8413,27 +9136,65 @@ public final class LlapDaemonProtocolProtos {
         return result;
       }
 
+      @java.lang.Override
       public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto buildPartial() {
         org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto(this);
+        if (bitField0_ != 0) { buildPartial0(result); }
+        onBuilt();
... 38296 lines suppressed ...